[ 
https://issues.apache.org/jira/browse/SPARK-1403?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14581933#comment-14581933
 ] 

John Omernik commented on SPARK-1403:
-------------------------------------

Per Kannan: Seeing this in Spark 1.2.2, 1.3.0, an 1.3.1.  1.2.0 does not 
produce the error. The case I have is a NPE caused by what appears to be this 
bug. 

Tasks fail with the stack trace below.  Per a discussion on the relavent code 
in the MapR Shim, ] it is trying to get a root classloader to use for loading a 
bunch of classes. It uses the thread's context class loader (TCCL) and keeps 
going up the parent chain. The Null being sent causes the NPE.  The concern 
here is not so much (in this case) the MapR component not handling the NPE 
(Althought that should be addressed too) but that Spark is switching between 
versions.  Happy to provide any other details. 

Full Error on 1.3.1 on Mesos:
15/05/19 09:31:26 INFO MemoryStore: MemoryStore started with capacity 1060.3 MB 
java.lang.NullPointerException at 
com.mapr.fs.ShimLoader.getRootClassLoader(ShimLoader.java:96) at 
com.mapr.fs.ShimLoader.injectNativeLoader(ShimLoader.java:232) at 
com.mapr.fs.ShimLoader.load(ShimLoader.java:194) at 
org.apache.hadoop.conf.CoreDefaultProperties.(CoreDefaultProperties.java:60) at 
java.lang.Class.forName0(Native Method) at 
java.lang.Class.forName(Class.java:274) at 
org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.java:1847)
 at org.apache.hadoop.conf.Configuration.getProperties(Configuration.java:2062) 
at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2272) 
at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2224) 
at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2141) at 
org.apache.hadoop.conf.Configuration.set(Configuration.java:992) at 
org.apache.hadoop.conf.Configuration.set(Configuration.java:966) at 
org.apache.spark.deploy.SparkHadoopUtil.newConfiguration(SparkHadoopUtil.scala:98)
 at org.apache.spark.deploy.SparkHadoopUtil.(SparkHadoopUtil.scala:43) at 
org.apache.spark.deploy.SparkHadoopUtil$.(SparkHadoopUtil.scala:220) at 
org.apache.spark.deploy.SparkHadoopUtil$.(SparkHadoopUtil.scala) at 
org.apache.spark.util.Utils$.getSparkOrYarnConfig(Utils.scala:1959) at 
org.apache.spark.storage.BlockManager.(BlockManager.scala:104) at 
org.apache.spark.storage.BlockManager.(BlockManager.scala:179) at 
org.apache.spark.SparkEnv$.create(SparkEnv.scala:310) at 
org.apache.spark.SparkEnv$.createExecutorEnv(SparkEnv.scala:186) at 
org.apache.spark.executor.MesosExecutorBackend.registered(MesosExecutorBackend.scala:70)
 java.lang.RuntimeException: Failure loading MapRClient. at 
com.mapr.fs.ShimLoader.injectNativeLoader(ShimLoader.java:283) at 
com.mapr.fs.ShimLoader.load(ShimLoader.java:194) at 
org.apache.hadoop.conf.CoreDefaultProperties.(CoreDefaultProperties.java:60) at 
java.lang.Class.forName0(Native Method) at 
java.lang.Class.forName(Class.java:274) at 
org.apache.hadoop.conf.Configuration.getClassByNameOrNull(Configuration.java:1847)
 at org.apache.hadoop.conf.Configuration.getProperties(Configuration.java:2062) 
at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2272) 
at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2224) 
at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2141) at 
org.apache.hadoop.conf.Configuration.set(Configuration.java:992) at 
org.apache.hadoop.conf.Configuration.set(Configuration.java:966) at 
org.apache.spark.deploy.SparkHadoopUtil.newConfiguration(SparkHadoopUtil.scala:98)
 at org.apache.spark.deploy.SparkHadoopUtil.(SparkHadoopUtil.scala:43) at 
org.apache.spark.deploy.SparkHadoopUtil$.(SparkHadoopUtil.scala:220) at 
org.apache.spark.deploy.SparkHadoopUtil$.(SparkHadoopUtil.scala) at 
org.apache.spark.util.Utils$.getSparkOrYarnConfig(Utils.scala:1959) at 
org.apache.spark.storage.BlockManager.(BlockManager.scala:104) at 
org.apache.spark.storage.BlockManager.(BlockManager.scala:179) at 
org.apache.spark.SparkEnv$.create(SparkEnv.scala:310) at 
org.apache.spark.SparkEnv$.createExecutorEnv(SparkEnv.scala:186) at 
org.apache.spark.executor.MesosExecutorBackend.registered(MesosExecutorBackend.scala:70)

> Spark on Mesos does not set Thread's context class loader
> ---------------------------------------------------------
>
>                 Key: SPARK-1403
>                 URL: https://issues.apache.org/jira/browse/SPARK-1403
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 1.0.0, 1.3.0
>         Environment: ubuntu 12.04 on vagrant
>            Reporter: Bharath Bhushan
>            Priority: Blocker
>             Fix For: 1.0.0
>
>
> I can run spark 0.9.0 on mesos but not spark 1.0.0. This is because the spark 
> executor on mesos slave throws a  java.lang.ClassNotFoundException for 
> org.apache.spark.serializer.JavaSerializer.
> The lengthy discussion is here: 
> http://apache-spark-user-list.1001560.n3.nabble.com/java-lang-ClassNotFoundException-spark-on-mesos-td3510.html#a3513



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to