Hi. I try to run the Spark Pi on the cluster, some strange errors happen and
I do not know what cause the error. Although I have posted this error to the
user@spark, I think it may be not a simple configuration error and the
developers may know it well.
When I am using the hadoop2.6 and spark-1.5.1-bin-hadoop2.6 the error log
is below:
118 10/01/01 11:59:14 ERROR yarn.ApplicationMaster: User class threw
exception: java.lang.reflect.InvocationTargetException
119 java.lang.reflect.InvocationTargetException
Caused by: java.lang.IllegalArgumentException:
java.lang.UnsatisfiedLinkError:
/opt/hadoop/tmp/nm-local-dir/usercache/root/appcache/application_1444839345484_0006/container_1444839345484_0006_01_000001/tmp/snappy-1.0.4.1-8378427e-4d5c-42b1-ae49-c9600
c204bd7-libsnappyjava.so:
/opt/hadoop/tmp/nm-local-dir/usercache/root/appcache/application_1444839345484_0006/container_14448
39345484_0006_01_000001/tmp/*snappy-1.0.4.1-8378427e-4d5c-42b1-ae49-c9600c204bd7-libsnappyjava.so:
cannot open shared object file: No such file or directory*
When I am using the hadoop2.6 and spark-1.3.0-bin-hadoop2.4 the error
seems different and the log is below:
* 108 org.apache.spark.SparkException: Job aborted due to stage failure:
Task serialization failed: java.lang.reflect.InvocationTargetException*
109 sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method)
110
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
111
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
112 java.lang.reflect.Constructor.newInstance(Constructor.java:408)
113
org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:68)
114
org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:60)
115
org.apache.spark.broadcast.TorrentBroadcast.org$apache$spark$broadcast$TorrentBroadcast$$setConf(TorrentBroadcast.scala:73)
116
org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:79)
117
org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34)
118
org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29)
119
org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62)
120 org.apache.spark.SparkContext.broadcast(SparkContext.scala:1051)
121
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$submitMissingTasks(DAGScheduler.scala:839)
122
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$submitStage(DAGScheduler.scala:778)
123
org.apache.spark.scheduler.DAGScheduler.handleJobSubmitted(DAGScheduler.scala:762)
124
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1362)
125
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1354)
126 org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
What causes the error? the java compatibility or the hadoop compatibility?
Thank you for your help
--
View this message in context:
http://apache-spark-developers-list.1001551.n3.nabble.com/Strange-spark-problems-among-different-versions-tp14609.html
Sent from the Apache Spark Developers List mailing list archive at Nabble.com.
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]