Re: EOFException when reading from HDFS

2014-09-12 Thread kent
Can anyone help me with this?  I have been stuck on this for a few days and
don't know what to try anymore.



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/EOFException-when-reading-from-HDFS-tp13844p14115.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



EOFException when reading from HDFS

2014-09-12 Thread kents
(FileSplit.java:87) 
org.apache.hadoop.io.ObjectWritable.readObject(ObjectWritable.java:285) 
org.apache.hadoop.io.ObjectWritable.readFields(ObjectWritable.java:77) 
   
org.apache.spark.SerializableWritable.readObject(SerializableWritable.scala:42) 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
   
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) 
   
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 
java.lang.reflect.Method.invoke(Method.java:606) 
java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017) 
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893) 
   
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) 
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) 
   
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 
java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) 
org.apache.spark.scheduler.ResultTask.readExternal(ResultTask.scala:147) 
java.io.ObjectInputStream.readExternalData(ObjectInputStream.java:1837) 
   
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1796) 
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 
java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) 
   
org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:63)
 
   
org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:85)
 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:169) 
   
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) 
   
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) 
java.lang.Thread.run(Thread.java:745) 
Driver stacktrace: 
at
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1033)
 
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1017)
 
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1015)
 
at
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) 
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) 
at
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1015) 
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
 
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
 
at scala.Option.foreach(Option.scala:236) 
at
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:633)
 
at
org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1207)
at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498) 
at akka.actor.ActorCell.invoke(ActorCell.scala:456) 
at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237) 
at akka.dispatch.Mailbox.run(Mailbox.scala:219) 
at
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) 
at
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
 
at
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) 
at
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
 
14/09/09 16:56:45 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0,
whose tasks have all completed, from pool 
14/09/09 16:56:45 INFO scheduler.TaskSchedulerImpl: Stage 0 was cancelled 
14/09/09 16:56:45 INFO scheduler.TaskSetManager: Loss was due to
java.io.EOFException [duplicate 7] 
14/09/09 16:56:45 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0,
whose tasks have all completed, from pool



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/EOFException-when-reading-from-HDFS-tp14118.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



EOFException when reading from HDFS

2014-09-09 Thread kent
(ObjectInputStream.java:1798) 
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 
java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) 
   
org.apache.spark.scheduler.ResultTask.readExternal(ResultTask.scala:147) 
   
java.io.ObjectInputStream.readExternalData(ObjectInputStream.java:1837) 
   
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1796) 
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 
java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) 
   
org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:63)
 
   
org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:85)
 
   
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:169) 
   
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) 
   
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) 
java.lang.Thread.run(Thread.java:745) 
Driver stacktrace: 
at
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1033)
 
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1017)
 
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1015)
 
at
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) 
at
scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) 
at
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1015) 
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
 
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
 
at scala.Option.foreach(Option.scala:236) 
at
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:633)
 
at
org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1207)
 
at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498) 
at akka.actor.ActorCell.invoke(ActorCell.scala:456) 
at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237) 
at akka.dispatch.Mailbox.run(Mailbox.scala:219) 
at
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
 
at
scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) 
at
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
 
at
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) 
at
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
 
14/09/09 16:56:45 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0,
whose tasks have all completed, from pool 
14/09/09 16:56:45 INFO scheduler.TaskSchedulerImpl: Stage 0 was cancelled 
14/09/09 16:56:45 INFO scheduler.TaskSetManager: Loss was due to
java.io.EOFException [duplicate 7] 
14/09/09 16:56:45 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0,
whose tasks have all completed, from pool 



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/EOFException-when-reading-from-HDFS-tp13844.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org