[
https://issues.apache.org/jira/browse/SPARK-1550?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14075509#comment-14075509
]
Josh Rosen commented on SPARK-1550:
-----------------------------------
Actually, there's still a similar problem in Spark 1.01:
{code}
>>> sc = SparkContext('sjkdlsjfsl')
14/07/26 16:04:14 INFO SecurityManager: Changing view acls to: joshrosen
14/07/26 16:04:14 INFO SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions: Set(joshrosen)
14/07/26 16:04:14 INFO Slf4jLogger: Slf4jLogger started
14/07/26 16:04:14 INFO Remoting: Starting remoting
14/07/26 16:04:14 INFO Remoting: Remoting started; listening on addresses
:[akka.tcp://spark@localhost:57189]
14/07/26 16:04:14 INFO Remoting: Remoting now listens on addresses:
[akka.tcp://spark@localhost:57189]
14/07/26 16:04:14 INFO SparkEnv: Registering MapOutputTracker
14/07/26 16:04:14 INFO SparkEnv: Registering BlockManagerMaster
14/07/26 16:04:14 INFO DiskBlockManager: Created local directory at
/var/folders/d2/mt2f3tx14vgg3b38xq_3ydfw0000gp/T/spark-local-20140726160414-c9d9
14/07/26 16:04:14 INFO ConnectionManager: Bound socket to port 57190 with id =
ConnectionManagerId(localhost,57190)
14/07/26 16:04:14 INFO MemoryStore: MemoryStore started with capacity 297.0 MB
14/07/26 16:04:14 INFO BlockManagerMaster: Trying to register BlockManager
14/07/26 16:04:14 INFO BlockManagerMasterActor: Registering block manager
localhost:57190 with 297.0 MB RAM
14/07/26 16:04:14 INFO BlockManagerMaster: Registered BlockManager
14/07/26 16:04:14 INFO HttpFileServer: HTTP File server directory is
/var/folders/d2/mt2f3tx14vgg3b38xq_3ydfw0000gp/T/spark-a7453bad-5278-43fd-aa43-efb33228c660
14/07/26 16:04:14 INFO HttpServer: Starting HTTP Server
14/07/26 16:04:14 INFO SparkUI: Started SparkUI at http://localhost:4040
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/Users/joshrosen/Documents/spark/spark/python/pyspark/context.py", line
133, in __init__
self._jsc = self._initialize_context(self._conf._jconf)
File "/Users/joshrosen/Documents/spark/spark/python/pyspark/context.py", line
179, in _initialize_context
return self._jvm.JavaSparkContext(jconf)
File
"/Users/joshrosen/Documents/spark/spark/python/lib/py4j-0.8.1-src.zip/py4j/java_gateway.py",
line 669, in __call__
File
"/Users/joshrosen/Documents/spark/spark/python/lib/py4j-0.8.1-src.zip/py4j/protocol.py",
line 300, in get_return_value
py4j.protocol.Py4JJavaError: An error occurred while calling
None.org.apache.spark.api.java.JavaSparkContext.
: org.apache.spark.SparkException: Could not parse Master URL: 'sjkdlsjfsl'
at
org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:1594)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:309)
at
org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:53)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:525)
at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:234)
at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:379)
at py4j.Gateway.invoke(Gateway.java:214)
at
py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:79)
at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:68)
at py4j.GatewayConnection.run(GatewayConnection.java:207)
at java.lang.Thread.run(Thread.java:722)
>>> sc = SparkContext('local')
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/Users/joshrosen/Documents/spark/spark/python/pyspark/context.py", line
93, in __init__
SparkContext._ensure_initialized(self, gateway=gateway)
File "/Users/joshrosen/Documents/spark/spark/python/pyspark/context.py", line
206, in _ensure_initialized
callsite.function, callsite.file, callsite.linenum))
ValueError: Cannot run multiple SparkContexts at once; existing
SparkContext(app=pyspark-shell, master=sjkdlsjfsl) created by __init__ at
<stdin>:1
>>>
{code}
The right solution is to clear {{active_spark_context}} if the SparkContext
creation fails. I'll submit a pull request for this in a little bit.
> Successive creation of spark context fails in pyspark, if the previous
> initialization of spark context had failed.
> ------------------------------------------------------------------------------------------------------------------
>
> Key: SPARK-1550
> URL: https://issues.apache.org/jira/browse/SPARK-1550
> Project: Spark
> Issue Type: Improvement
> Components: PySpark
> Reporter: Prabin Banka
> Labels: pyspark, sparkcontext
>
> For example;-
> In PySpark, if we try to initialize spark context with insufficient
> arguments, >>>sc=SparkContext('local')
> it fails with an exception
> Exception: An application name must be set in your configuration
> This is all fine.
> However, any successive creation of spark context with correct arguments,
> also fails,
> >>>s1=SparkContext('local', 'test1')
> AttributeError: 'SparkContext' object has no attribute 'master'
--
This message was sent by Atlassian JIRA
(v6.2#6252)