0

When executing from within my Jupyter Notebook some cells containing some Spark commands (eg., some DataFrame.show() methods or some spark.sql select commands involving 6 million row DataFrames), I get the following sequence of message errors:

Py4JJavaError: An error occurred while calling xxxx.showString.
SparkException: Job aborted due to stage failure.
Caused by: org.apache.spark.SparkException: Python worker failed to connect back.
Caused by: java.net.SocketTimeoutException: Accept timed out.

How can I interpret them?
I work in a local 8g Spark session.

desertnaut
  • 57,590
  • 26
  • 140
  • 166

0 Answers0