[ 
https://issues.apache.org/jira/browse/SPARK-4194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14193732#comment-14193732
 ] 

Josh Rosen commented on SPARK-4194:
-----------------------------------

I've marked this a blocker of SPARK-4180, an issue that tries to add exceptions 
when users try to create multiple active SparkContexts in the same JVM.  
PySpark already guards against this, but earlier versions of its error-checking 
code ran into issues where users would fail their initial attempt to create a 
SparkContext and then be unable to create new ones because we didn't clear the 
{{activeSparkContext}} variable after the constructor threw an exception.

To fix this, we need to wrap the constructor code in a {{try}} block. 

> Exceptions thrown during SparkContext or SparkEnv construction might lead to 
> resource leaks or corrupted global state
> ---------------------------------------------------------------------------------------------------------------------
>
>                 Key: SPARK-4194
>                 URL: https://issues.apache.org/jira/browse/SPARK-4194
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>            Reporter: Josh Rosen
>            Priority: Critical
>
> The SparkContext and SparkEnv constructors instantiate a bunch of objects 
> that may need to be cleaned up after they're no longer needed.  If an 
> exception is thrown during SparkContext or SparkEnv construction (e.g. due to 
> a bad configuration setting), then objects created earlier in the constructor 
> may not be properly cleaned up.
> This is unlikely to cause problems for batch jobs submitted through 
> {{spark-submit}}, since failure to construct SparkContext will probably cause 
> the JVM to exit, but it is a potentially serious issue in interactive 
> environments where a user might attempt to create SparkContext with some 
> configuration, fail due to an error, and re-attempt the creation with new 
> settings.  In this case, resources from the previous creation attempt might 
> not have been cleaned up and could lead to confusing errors (especially if 
> the old, leaked resources share global state with the new SparkContext).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to