Hey there!

I was running SQL queries via PySpark (using the wmfdata package
<https://github.com/neilpquinn/wmfdata/blob/master/wmfdata/hive.py>) on
SWAP when one of my queries failed with "java.lang.OutofMemoryError: Java
heap space".

After that, when I tried to call the spark.sql function again (via
wmfdata.hive.run), it failed with "java.lang.IllegalStateException: Cannot
call methods on a stopped SparkContext."

When I tried to create a new Spark context using
SparkSession.builder.getOrCreate (whether using wmfdata.spark.get_session
or directly), it returned a SparkContent object properly, but calling the
object's sql function still gave the "stopped SparkContext error".

Any idea what's going on? I assume restarting the notebook kernel would
take care of the problem, but it seems like there has to be a better way to
recover.

Thank you!
_______________________________________________
Analytics mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/analytics

Reply via email to