Akash,
Are you able to run your code in pyspark shell with no issues? Best Regards, Vamshi T ________________________________ From: Hyukjin Kwon <gurwls...@gmail.com> Sent: Friday, June 15, 2018 10:18 AM To: Marcelo Vanzin Cc: aakash.spark....@gmail.com; user @spark Subject: Re: Issue upgrading to Spark 2.3.1 (Maintenance Release) I use PyCharm. Mind if I ask to elaborate what you did step by step? 2018년 6월 16일 (토) 오전 12:11, Marcelo Vanzin <van...@cloudera.com.invalid>님이 작성: I'm not familiar with PyCharm. But if you can run "pyspark" from the command line and not hit this, then this might be an issue with PyCharm or your environment - e.g. having an old version of the pyspark code around, or maybe PyCharm itself might need to be updated. On Thu, Jun 14, 2018 at 10:01 PM, Aakash Basu <aakash.spark....@gmail.com<mailto:aakash.spark....@gmail.com>> wrote: > Hi, > > Downloaded the latest Spark version because the of the fix for "ERROR > AsyncEventQueue:70 - Dropping event from queue appStatus." > > After setting environment variables and running the same code in PyCharm, > I'm getting this error, which I can't find a solution of. > > Exception in thread "main" java.util.NoSuchElementException: key not found: > _PYSPARK_DRIVER_CONN_INFO_PATH > at scala.collection.MapLike$class.default(MapLike.scala:228) > at scala.collection.AbstractMap.default(Map.scala:59) > at scala.collection.MapLike$class.apply(MapLike.scala:141) > at scala.collection.AbstractMap.apply(Map.scala:59) > at > org.apache.spark.api.python.PythonGatewayServer$.main(PythonGatewayServer.scala:64) > at > org.apache.spark.api.python.PythonGatewayServer.main(PythonGatewayServer.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) > at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:894) > at > org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:198) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:228) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > > Any help? > > Thanks, > Aakash. -- Marcelo --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org<mailto:user-unsubscr...@spark.apache.org>