Hi ,

I am running pyspark in dockerized jupyter environment , I am constantly
getting this error :

```

Py4JJavaError: An error occurred while calling
z:org.apache.spark.api.python.PythonRDD.runJob.
: org.apache.spark.SparkException: Job aborted due to stage failure:
Task 33 in stage 25.0 failed 1 times, most recent failure: Lost task
33.0 in stage 25.0 (TID 35067, localhost, executor driver)
: org.apache.hadoop.hdfs.BlockMissingException
: Could not obtain block:
BP-1742911633-10.225.201.50-1479296658503:blk_1233169822_159765693

```

Please can anyone help me with how to handle such exception in pyspark.

-- 
Best Regards
*Divay Jindal*

Reply via email to