Hi everyone, I'm having problems with loading files. Either with java code or spark-shell, I got same errors when I try to load a text file.
I added hadoop-client and hadoop-common 2.0.0-cdh4.6.0 as dependencies and maven-shade-plugin is configured. I have CDH 4.6.0, spark-0.9.1-bin-cdh4 and JDK7 installed on my nodes. Here is my code: public class CountTool { public static void main(String[] args) { // validate input if (args.length != 1) { System.err.println("invalid usage!\ncount <input>"); return; } // input parameters String input = args[0]; // spark context SparkConf conf = new SparkConf() .setMaster("spark://server08.mydomain.com:7077") .setSparkHome("/opt/spark") .setJars(new String[] {"/home/hadoop/dist/mycode.jar"}) .setAppName(CountTool.class.getSimpleName()) .set("spark.executor.memory", "1g"); JavaSparkContext jsc = new JavaSparkContext(conf); // fix for cdh4 jsc.hadoopConfiguration().set("fs.hdfs.impl", org.apache.hadoop.hdfs.DistributedFileSystem.class.getName()); jsc.hadoopConfiguration().set("fs.file.impl", org.apache.hadoop.fs.LocalFileSystem.class.getName()); // load file and count JavaRDD<String> file = jsc.textFile(input); long count = file.count(); System.out.println(String.format("file %s %d lines", input, count)); } } I couldn't load any file either from local disk or hdfs with this code. When I try to load a file from local it gives "Loss was due to java.io.EOFException" error: [hadoop@server08 clustering]$ java -cp /home/hadoop/dist/mycode.jar com.gg.clustering.spark.CountTool /home/hadoop/scripts/clustering/deals.txt 2014-05-20 10:04:22 INFO Slf4jLogger:80 - Slf4jLogger started 2014-05-20 10:04:22 INFO Remoting:74 - Starting remoting 2014-05-20 10:04:23 INFO Remoting:74 - Remoting started; listening on addresses :[akka.tcp://sp...@server08.mydomain.com:34236] 2014-05-20 10:04:23 INFO Remoting:74 - Remoting now listens on addresses: [akka.tcp://sp...@server08.mydomain.com:34236] 2014-05-20 10:04:23 INFO SparkEnv:50 - Registering BlockManagerMaster 2014-05-20 10:04:23 INFO DiskBlockManager:50 - Created local directory at /tmp/spark-local-20140520100423-ae57 2014-05-20 10:04:23 INFO MemoryStore:50 - MemoryStore started with capacity 8.4 GB. 2014-05-20 10:04:23 INFO ConnectionManager:50 - Bound socket to port 58899 with id = ConnectionManagerId(server08.mydomain.com,58899) 2014-05-20 10:04:23 INFO BlockManagerMaster:50 - Trying to register BlockManager 2014-05-20 10:04:23 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server08.mydomain.com:58899 with 8.4 GB RAM 2014-05-20 10:04:23 INFO BlockManagerMaster:50 - Registered BlockManager 2014-05-20 10:04:23 INFO HttpServer:50 - Starting HTTP Server 2014-05-20 10:04:23 INFO Server:266 - jetty-7.x.y-SNAPSHOT 2014-05-20 10:04:23 INFO AbstractConnector:338 - Started SocketConnector@0.0.0.0:38679 2014-05-20 10:04:23 INFO HttpBroadcast:50 - Broadcast server started at http://10.35.74.148:38679 2014-05-20 10:04:23 INFO SparkEnv:50 - Registering MapOutputTracker 2014-05-20 10:04:23 INFO HttpFileServer:50 - HTTP File server directory is /tmp/spark-1e8e7419-9d16-446e-a7e2-5e0e6307f7d0 2014-05-20 10:04:23 INFO HttpServer:50 - Starting HTTP Server 2014-05-20 10:04:23 INFO Server:266 - jetty-7.x.y-SNAPSHOT 2014-05-20 10:04:23 INFO AbstractConnector:338 - Started SocketConnector@0.0.0.0:42538 2014-05-20 10:04:23 INFO Server:266 - jetty-7.x.y-SNAPSHOT 2014-05-20 10:04:23 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/storage/rdd,null} 2014-05-20 10:04:23 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/storage,null} 2014-05-20 10:04:23 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/stages/stage,null} 2014-05-20 10:04:23 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/stages/pool,null} 2014-05-20 10:04:23 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/stages,null} 2014-05-20 10:04:23 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/environment,null} 2014-05-20 10:04:23 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/executors,null} 2014-05-20 10:04:23 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/metrics/json,null} 2014-05-20 10:04:23 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/static,null} 2014-05-20 10:04:23 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/,null} 2014-05-20 10:04:23 INFO AbstractConnector:338 - Started SelectChannelConnector@0.0.0.0:4040 2014-05-20 10:04:23 INFO SparkUI:50 - Started Spark Web UI at http://server08.mydomain.com:4040 2014-05-20 10:04:23 WARN NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2014-05-20 10:04:24 INFO SparkContext:50 - Added JAR /home/hadoop/dist/mycode.jar at http://10.35.74.148:42538/jars/mycode.jar with timestamp 1400569464072 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Connecting to master spark://server08.mydomain.com:7077... 2014-05-20 10:04:24 INFO MemoryStore:50 - ensureFreeSpace(124827) called with curMem=0, maxMem=9006219264 2014-05-20 10:04:24 INFO MemoryStore:50 - Block broadcast_0 stored as values to memory (estimated size 121.9 KB, free 8.4 GB) 2014-05-20 10:04:24 INFO SparkDeploySchedulerBackend:50 - Connected to Spark cluster with app ID app-20140520100424-0005 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor added: app-20140520100424-0005/0 on worker-20140520085014-server02.mydomain.com-54319 (server02.mydomain.com:54319) with 16 cores 2014-05-20 10:04:24 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100424-0005/0 on hostPort server02.mydomain.com:54319 with 16 cores, 1024.0 MB RAM 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor added: app-20140520100424-0005/1 on worker-20140520085014-server06.mydomain.com-33290 (server06.mydomain.com:33290) with 16 cores 2014-05-20 10:04:24 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100424-0005/1 on hostPort server06.mydomain.com:33290 with 16 cores, 1024.0 MB RAM 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor added: app-20140520100424-0005/2 on worker-20140520085014-server03.mydomain.com-50690 (server03.mydomain.com:50690) with 16 cores 2014-05-20 10:04:24 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100424-0005/2 on hostPort server03.mydomain.com:50690 with 16 cores, 1024.0 MB RAM 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor added: app-20140520100424-0005/3 on worker-20140520085014-server04.mydomain.com-41041 (server04.mydomain.com:41041) with 16 cores 2014-05-20 10:04:24 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100424-0005/3 on hostPort server04.mydomain.com:41041 with 16 cores, 1024.0 MB RAM 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor added: app-20140520100424-0005/4 on worker-20140520085014-server07.mydomain.com-44641 (server07.mydomain.com:44641) with 16 cores 2014-05-20 10:04:24 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100424-0005/4 on hostPort server07.mydomain.com:44641 with 16 cores, 1024.0 MB RAM 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor added: app-20140520100424-0005/5 on worker-20140520085014-server05.mydomain.com-32908 (server05.mydomain.com:32908) with 16 cores 2014-05-20 10:04:24 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100424-0005/5 on hostPort server05.mydomain.com:32908 with 16 cores, 1024.0 MB RAM 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100424-0005/0 is now RUNNING 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100424-0005/1 is now RUNNING 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100424-0005/2 is now RUNNING 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100424-0005/3 is now RUNNING 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100424-0005/4 is now RUNNING 2014-05-20 10:04:24 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100424-0005/5 is now RUNNING 2014-05-20 10:04:24 INFO FileInputFormat:233 - Total input paths to process : 1 2014-05-20 10:04:24 INFO SparkContext:50 - Starting job: count at CountTool.java:44 2014-05-20 10:04:24 INFO DAGScheduler:50 - Got job 0 (count at CountTool.java:44) with 2 output partitions (allowLocal=false) 2014-05-20 10:04:24 INFO DAGScheduler:50 - Final stage: Stage 0 (count at CountTool.java:44) 2014-05-20 10:04:24 INFO DAGScheduler:50 - Parents of final stage: List() 2014-05-20 10:04:24 INFO DAGScheduler:50 - Missing parents: List() 2014-05-20 10:04:24 INFO DAGScheduler:50 - Submitting Stage 0 (MappedRDD[1] at textFile at CountTool.java:43), which has no missing parents 2014-05-20 10:04:24 INFO DAGScheduler:50 - Submitting 2 missing tasks from Stage 0 (MappedRDD[1] at textFile at CountTool.java:43) 2014-05-20 10:04:24 INFO TaskSchedulerImpl:50 - Adding task set 0.0 with 2 tasks 2014-05-20 10:04:25 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server03.mydomain.com:39494/user/Executor#1298593046] with ID 2 2014-05-20 10:04:25 INFO TaskSetManager:50 - Starting task 0.0:0 as TID 0 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 2014-05-20 10:04:25 INFO TaskSetManager:50 - Serialized task 0.0:0 as 1629 bytes in 4 ms 2014-05-20 10:04:25 INFO TaskSetManager:50 - Starting task 0.0:1 as TID 1 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 2014-05-20 10:04:25 INFO TaskSetManager:50 - Serialized task 0.0:1 as 1629 bytes in 0 ms 2014-05-20 10:04:25 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server02.mydomain.com:59496/user/Executor#1432291773] with ID 0 2014-05-20 10:04:25 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server06.mydomain.com:39547/user/Executor#78156672] with ID 1 2014-05-20 10:04:25 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server05.mydomain.com:34135/user/Executor#-777886910] with ID 5 2014-05-20 10:04:25 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server07.mydomain.com:41161/user/Executor#336519762] with ID 4 2014-05-20 10:04:25 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server04.mydomain.com:34174/user/Executor#-1382135714] with ID 3 2014-05-20 10:04:25 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server03.mydomain.com:58285 with 589.2 MB RAM 2014-05-20 10:04:26 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server02.mydomain.com:51425 with 589.2 MB RAM 2014-05-20 10:04:26 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server06.mydomain.com:46383 with 589.2 MB RAM 2014-05-20 10:04:26 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server05.mydomain.com:52167 with 589.2 MB RAM 2014-05-20 10:04:26 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server04.mydomain.com:41308 with 589.2 MB RAM 2014-05-20 10:04:26 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server07.mydomain.com:42993 with 589.2 MB RAM 2014-05-20 10:04:27 WARN TaskSetManager:62 - Lost TID 1 (task 0.0:1) 2014-05-20 10:04:27 WARN TaskSetManager:62 - Loss was due to java.io.EOFException java.io.EOFException at java.io.ObjectInputStream$BlockDataInputStream.readFully(ObjectInputStream.java:2744) at java.io.ObjectInputStream.readFully(ObjectInputStream.java:1032) at org.apache.hadoop.io.DataOutputBuffer$Buffer.write(DataOutputBuffer.java:68) at org.apache.hadoop.io.DataOutputBuffer.write(DataOutputBuffer.java:106) at org.apache.hadoop.io.UTF8.readChars(UTF8.java:258) at org.apache.hadoop.io.UTF8.readString(UTF8.java:250) at org.apache.hadoop.mapred.FileSplit.readFields(FileSplit.java:87) at org.apache.hadoop.io.ObjectWritable.readObject(ObjectWritable.java:280) at org.apache.hadoop.io.ObjectWritable.readFields(ObjectWritable.java:75) at org.apache.spark.SerializableWritable.readObject(SerializableWritable.scala:39) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) at org.apache.spark.scheduler.ResultTask.readExternal(ResultTask.scala:145) at java.io.ObjectInputStream.readExternalData(ObjectInputStream.java:1837) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1796) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) at org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:40) at org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:62) at org.apache.spark.executor.Executor$TaskRunner$$anonfun$run$1.apply$mcV$sp(Executor.scala:193) at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:42) at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:41) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) at org.apache.spark.deploy.SparkHadoopUtil.runAsUser(SparkHadoopUtil.scala:41) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:176) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:744) 2014-05-20 10:04:27 WARN TaskSetManager:62 - Lost TID 0 (task 0.0:0) 2014-05-20 10:04:27 INFO TaskSetManager:50 - Loss was due to java.io.EOFException [duplicate 1] 2014-05-20 10:04:27 INFO TaskSetManager:50 - Starting task 0.0:0 as TID 2 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 2014-05-20 10:04:27 INFO TaskSetManager:50 - Serialized task 0.0:0 as 1629 bytes in 0 ms 2014-05-20 10:04:27 INFO TaskSetManager:50 - Starting task 0.0:1 as TID 3 on executor 5: server05.mydomain.com (PROCESS_LOCAL) 2014-05-20 10:04:27 INFO TaskSetManager:50 - Serialized task 0.0:1 as 1629 bytes in 0 ms 2014-05-20 10:04:27 WARN TaskSetManager:62 - Lost TID 2 (task 0.0:0) 2014-05-20 10:04:27 INFO TaskSetManager:50 - Loss was due to java.io.EOFException [duplicate 2] 2014-05-20 10:04:27 INFO TaskSetManager:50 - Starting task 0.0:0 as TID 4 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 2014-05-20 10:04:27 INFO TaskSetManager:50 - Serialized task 0.0:0 as 1629 bytes in 0 ms 2014-05-20 10:04:27 WARN TaskSetManager:62 - Lost TID 4 (task 0.0:0) 2014-05-20 10:04:27 INFO TaskSetManager:50 - Loss was due to java.io.EOFException [duplicate 3] 2014-05-20 10:04:27 INFO TaskSetManager:50 - Starting task 0.0:0 as TID 5 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 2014-05-20 10:04:27 INFO TaskSetManager:50 - Serialized task 0.0:0 as 1629 bytes in 0 ms 2014-05-20 10:04:27 WARN TaskSetManager:62 - Lost TID 5 (task 0.0:0) 2014-05-20 10:04:27 INFO TaskSetManager:50 - Loss was due to java.io.EOFException [duplicate 4] 2014-05-20 10:04:27 ERROR TaskSetManager:66 - Task 0.0:0 failed 4 times; aborting job 2014-05-20 10:04:27 INFO DAGScheduler:50 - Failed to run count at CountTool.java:44 Exception in thread "main" org.apache.spark.SparkException: Job aborted: Task 0.0:0 failed 4 times (most recent failure: Exception failure: java.io.EOFException) at org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1020) at org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1018) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$abortStage(DAGScheduler.scala:1018) at org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) at org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) at scala.Option.foreach(Option.scala:236) at org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:604) at org.apache.spark.scheduler.DAGScheduler$$anonfun$start$1$$anon$2$$anonfun$receive$1.applyOrElse(DAGScheduler.scala:190) at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498) at akka.actor.ActorCell.invoke(ActorCell.scala:456) at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237) at akka.dispatch.Mailbox.run(Mailbox.scala:219) at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) I got same error with hdfs: [hadoop@server08 clustering]$ java -cp /home/hadoop/dist/mycode.jar com.gg.clustering.spark.CountTool hdfs://server01.mydomain.com:/user/hadoop/results/clustering/deals.txt 2014-05-20 10:05:26 INFO Slf4jLogger:80 - Slf4jLogger started 2014-05-20 10:05:26 INFO Remoting:74 - Starting remoting 2014-05-20 10:05:26 INFO Remoting:74 - Remoting started; listening on addresses :[akka.tcp://sp...@server08.mydomain.com:55747] 2014-05-20 10:05:26 INFO Remoting:74 - Remoting now listens on addresses: [akka.tcp://sp...@server08.mydomain.com:55747] 2014-05-20 10:05:26 INFO SparkEnv:50 - Registering BlockManagerMaster 2014-05-20 10:05:26 INFO DiskBlockManager:50 - Created local directory at /tmp/spark-local-20140520100526-e8cb 2014-05-20 10:05:26 INFO MemoryStore:50 - MemoryStore started with capacity 8.4 GB. 2014-05-20 10:05:26 INFO ConnectionManager:50 - Bound socket to port 39356 with id = ConnectionManagerId(server08.mydomain.com,39356) 2014-05-20 10:05:26 INFO BlockManagerMaster:50 - Trying to register BlockManager 2014-05-20 10:05:26 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server08.mydomain.com:39356 with 8.4 GB RAM 2014-05-20 10:05:26 INFO BlockManagerMaster:50 - Registered BlockManager 2014-05-20 10:05:26 INFO HttpServer:50 - Starting HTTP Server 2014-05-20 10:05:26 INFO Server:266 - jetty-7.x.y-SNAPSHOT 2014-05-20 10:05:26 INFO AbstractConnector:338 - Started SocketConnector@0.0.0.0:53241 2014-05-20 10:05:26 INFO HttpBroadcast:50 - Broadcast server started at http://10.35.74.148:53241 2014-05-20 10:05:26 INFO SparkEnv:50 - Registering MapOutputTracker 2014-05-20 10:05:26 INFO HttpFileServer:50 - HTTP File server directory is /tmp/spark-ba35f7bd-9741-40e3-a814-37dbb9e2592e 2014-05-20 10:05:26 INFO HttpServer:50 - Starting HTTP Server 2014-05-20 10:05:26 INFO Server:266 - jetty-7.x.y-SNAPSHOT 2014-05-20 10:05:26 INFO AbstractConnector:338 - Started SocketConnector@0.0.0.0:39988 2014-05-20 10:05:26 INFO Server:266 - jetty-7.x.y-SNAPSHOT 2014-05-20 10:05:26 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/storage/rdd,null} 2014-05-20 10:05:26 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/storage,null} 2014-05-20 10:05:26 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/stages/stage,null} 2014-05-20 10:05:26 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/stages/pool,null} 2014-05-20 10:05:26 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/stages,null} 2014-05-20 10:05:26 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/environment,null} 2014-05-20 10:05:26 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/executors,null} 2014-05-20 10:05:26 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/metrics/json,null} 2014-05-20 10:05:26 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/static,null} 2014-05-20 10:05:26 INFO ContextHandler:744 - started o.e.j.s.h.ContextHandler{/,null} 2014-05-20 10:05:26 INFO AbstractConnector:338 - Started SelectChannelConnector@0.0.0.0:4040 2014-05-20 10:05:26 INFO SparkUI:50 - Started Spark Web UI at http://server08.mydomain.com:4040 2014-05-20 10:05:26 WARN NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2014-05-20 10:05:27 INFO SparkContext:50 - Added JAR /home/hadoop/dist/mycode.jar at http://10.35.74.148:39988/jars/mycode.jar with timestamp 1400569527177 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Connecting to master spark://server08.mydomain.com:7077... 2014-05-20 10:05:27 INFO MemoryStore:50 - ensureFreeSpace(124827) called with curMem=0, maxMem=9006219264 2014-05-20 10:05:27 INFO MemoryStore:50 - Block broadcast_0 stored as values to memory (estimated size 121.9 KB, free 8.4 GB) 2014-05-20 10:05:27 INFO SparkDeploySchedulerBackend:50 - Connected to Spark cluster with app ID app-20140520100527-0006 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor added: app-20140520100527-0006/0 on worker-20140520085014-server02.mydomain.com-54319 (server02.mydomain.com:54319) with 16 cores 2014-05-20 10:05:27 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100527-0006/0 on hostPort server02.mydomain.com:54319 with 16 cores, 1024.0 MB RAM 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor added: app-20140520100527-0006/1 on worker-20140520085014-server06.mydomain.com-33290 (server06.mydomain.com:33290) with 16 cores 2014-05-20 10:05:27 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100527-0006/1 on hostPort server06.mydomain.com:33290 with 16 cores, 1024.0 MB RAM 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor added: app-20140520100527-0006/2 on worker-20140520085014-server03.mydomain.com-50690 (server03.mydomain.com:50690) with 16 cores 2014-05-20 10:05:27 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100527-0006/2 on hostPort server03.mydomain.com:50690 with 16 cores, 1024.0 MB RAM 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor added: app-20140520100527-0006/3 on worker-20140520085014-server04.mydomain.com-41041 (server04.mydomain.com:41041) with 16 cores 2014-05-20 10:05:27 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100527-0006/3 on hostPort server04.mydomain.com:41041 with 16 cores, 1024.0 MB RAM 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor added: app-20140520100527-0006/4 on worker-20140520085014-server07.mydomain.com-44641 (server07.mydomain.com:44641) with 16 cores 2014-05-20 10:05:27 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100527-0006/4 on hostPort server07.mydomain.com:44641 with 16 cores, 1024.0 MB RAM 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor added: app-20140520100527-0006/5 on worker-20140520085014-server05.mydomain.com-32908 (server05.mydomain.com:32908) with 16 cores 2014-05-20 10:05:27 INFO SparkDeploySchedulerBackend:50 - Granted executor ID app-20140520100527-0006/5 on hostPort server05.mydomain.com:32908 with 16 cores, 1024.0 MB RAM 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100527-0006/1 is now RUNNING 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100527-0006/0 is now RUNNING 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100527-0006/5 is now RUNNING 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100527-0006/2 is now RUNNING 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100527-0006/3 is now RUNNING 2014-05-20 10:05:27 INFO AppClient$ClientActor:50 - Executor updated: app-20140520100527-0006/4 is now RUNNING 2014-05-20 10:05:28 INFO FileInputFormat:233 - Total input paths to process : 1 2014-05-20 10:05:28 INFO SparkContext:50 - Starting job: count at CountTool.java:44 2014-05-20 10:05:28 INFO DAGScheduler:50 - Got job 0 (count at CountTool.java:44) with 2 output partitions (allowLocal=false) 2014-05-20 10:05:28 INFO DAGScheduler:50 - Final stage: Stage 0 (count at CountTool.java:44) 2014-05-20 10:05:28 INFO DAGScheduler:50 - Parents of final stage: List() 2014-05-20 10:05:28 INFO DAGScheduler:50 - Missing parents: List() 2014-05-20 10:05:28 INFO DAGScheduler:50 - Submitting Stage 0 (MappedRDD[1] at textFile at CountTool.java:43), which has no missing parents 2014-05-20 10:05:28 INFO DAGScheduler:50 - Submitting 2 missing tasks from Stage 0 (MappedRDD[1] at textFile at CountTool.java:43) 2014-05-20 10:05:28 INFO TaskSchedulerImpl:50 - Adding task set 0.0 with 2 tasks 2014-05-20 10:05:28 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server03.mydomain.com:40916/user/Executor#1474912] with ID 2 2014-05-20 10:05:28 INFO TaskSetManager:50 - Starting task 0.0:0 as TID 0 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 2014-05-20 10:05:28 INFO TaskSetManager:50 - Serialized task 0.0:0 as 1692 bytes in 5 ms 2014-05-20 10:05:28 INFO TaskSetManager:50 - Starting task 0.0:1 as TID 1 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 2014-05-20 10:05:28 INFO TaskSetManager:50 - Serialized task 0.0:1 as 1692 bytes in 0 ms 2014-05-20 10:05:28 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server02.mydomain.com:42834/user/Executor#388749186] with ID 0 2014-05-20 10:05:29 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server05.mydomain.com:49626/user/Executor#1815450843] with ID 5 2014-05-20 10:05:29 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server03.mydomain.com:49424 with 589.2 MB RAM 2014-05-20 10:05:29 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server07.mydomain.com:35109/user/Executor#-199036165] with ID 4 2014-05-20 10:05:29 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server06.mydomain.com:35535/user/Executor#-1925679390] with ID 1 2014-05-20 10:05:29 INFO SparkDeploySchedulerBackend:50 - Registered executor: Actor[akka.tcp://sparkexecu...@server04.mydomain.com:53806/user/Executor#725612946] with ID 3 2014-05-20 10:05:29 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server02.mydomain.com:35927 with 589.2 MB RAM 2014-05-20 10:05:29 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server05.mydomain.com:57374 with 589.2 MB RAM 2014-05-20 10:05:29 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server07.mydomain.com:50669 with 589.2 MB RAM 2014-05-20 10:05:29 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server06.mydomain.com:33799 with 589.2 MB RAM 2014-05-20 10:05:29 INFO BlockManagerMasterActor$BlockManagerInfo:50 - Registering block manager server04.mydomain.com:53457 with 589.2 MB RAM 2014-05-20 10:05:30 WARN TaskSetManager:62 - Lost TID 1 (task 0.0:1) 2014-05-20 10:05:30 WARN TaskSetManager:62 - Loss was due to java.io.EOFException java.io.EOFException at java.io.ObjectInputStream$BlockDataInputStream.readFully(ObjectInputStream.java:2744) at java.io.ObjectInputStream.readFully(ObjectInputStream.java:1032) at org.apache.hadoop.io.DataOutputBuffer$Buffer.write(DataOutputBuffer.java:68) at org.apache.hadoop.io.DataOutputBuffer.write(DataOutputBuffer.java:106) at org.apache.hadoop.io.UTF8.readChars(UTF8.java:258) at org.apache.hadoop.io.UTF8.readString(UTF8.java:250) at org.apache.hadoop.mapred.FileSplit.readFields(FileSplit.java:87) at org.apache.hadoop.io.ObjectWritable.readObject(ObjectWritable.java:280) at org.apache.hadoop.io.ObjectWritable.readFields(ObjectWritable.java:75) at org.apache.spark.SerializableWritable.readObject(SerializableWritable.scala:39) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) at org.apache.spark.scheduler.ResultTask.readExternal(ResultTask.scala:145) at java.io.ObjectInputStream.readExternalData(ObjectInputStream.java:1837) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1796) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) at org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:40) at org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:62) at org.apache.spark.executor.Executor$TaskRunner$$anonfun$run$1.apply$mcV$sp(Executor.scala:193) at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:42) at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:41) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) at org.apache.spark.deploy.SparkHadoopUtil.runAsUser(SparkHadoopUtil.scala:41) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:176) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:744) 2014-05-20 10:05:30 WARN TaskSetManager:62 - Lost TID 0 (task 0.0:0) 2014-05-20 10:05:30 INFO TaskSetManager:50 - Loss was due to java.io.EOFException [duplicate 1] 2014-05-20 10:05:30 INFO TaskSetManager:50 - Starting task 0.0:0 as TID 2 on executor 2: server03.mydomain.com (NODE_LOCAL) 2014-05-20 10:05:30 INFO TaskSetManager:50 - Serialized task 0.0:0 as 1692 bytes in 1 ms 2014-05-20 10:05:30 INFO TaskSetManager:50 - Starting task 0.0:1 as TID 3 on executor 5: server05.mydomain.com (NODE_LOCAL) 2014-05-20 10:05:30 INFO TaskSetManager:50 - Serialized task 0.0:1 as 1692 bytes in 0 ms 2014-05-20 10:05:30 WARN TaskSetManager:62 - Lost TID 2 (task 0.0:0) 2014-05-20 10:05:30 INFO TaskSetManager:50 - Loss was due to java.io.EOFException [duplicate 2] 2014-05-20 10:05:30 INFO TaskSetManager:50 - Starting task 0.0:0 as TID 4 on executor 2: server03.mydomain.com (NODE_LOCAL) 2014-05-20 10:05:30 INFO TaskSetManager:50 - Serialized task 0.0:0 as 1692 bytes in 0 ms 2014-05-20 10:05:30 WARN TaskSetManager:62 - Lost TID 4 (task 0.0:0) 2014-05-20 10:05:30 INFO TaskSetManager:50 - Loss was due to java.io.EOFException [duplicate 3] 2014-05-20 10:05:30 INFO TaskSetManager:50 - Starting task 0.0:0 as TID 5 on executor 2: server03.mydomain.com (NODE_LOCAL) 2014-05-20 10:05:30 INFO TaskSetManager:50 - Serialized task 0.0:0 as 1692 bytes in 0 ms 2014-05-20 10:05:30 WARN TaskSetManager:62 - Lost TID 5 (task 0.0:0) 2014-05-20 10:05:30 INFO TaskSetManager:50 - Loss was due to java.io.EOFException [duplicate 4] 2014-05-20 10:05:30 ERROR TaskSetManager:66 - Task 0.0:0 failed 4 times; aborting job 2014-05-20 10:05:30 INFO DAGScheduler:50 - Failed to run count at CountTool.java:44 Exception in thread "main" org.apache.spark.SparkException: Job aborted: Task 0.0:0 failed 4 times (most recent failure: Exception failure: java.io.EOFException) at org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1020) at org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1018) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$abortStage(DAGScheduler.scala:1018) at org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) at org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) at scala.Option.foreach(Option.scala:236) at org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:604) at org.apache.spark.scheduler.DAGScheduler$$anonfun$start$1$$anon$2$$anonfun$receive$1.applyOrElse(DAGScheduler.scala:190) at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498) at akka.actor.ActorCell.invoke(ActorCell.scala:456) at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237) at akka.dispatch.Mailbox.run(Mailbox.scala:219) at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) spark-shell also gives the same error: scala> sc.textFile("/home/hadoop/scripts/clustering/deals.txt").count() 14/05/20 10:00:22 INFO SparkDeploySchedulerBackend: Registered executor: Actor[akka.tcp://sparkexecu...@server03.mydomain.com:35256/user/Executor#-1346794213] with ID 2 14/05/20 10:00:22 INFO SparkDeploySchedulerBackend: Registered executor: Actor[akka.tcp://sparkexecu...@server02.mydomain.com:60855/user/Executor#-1316203186] with ID 0 14/05/20 10:00:22 INFO SparkDeploySchedulerBackend: Registered executor: Actor[akka.tcp://sparkexecu...@server07.mydomain.com:47643/user/Executor#-556408193] with ID 4 14/05/20 10:00:22 INFO SparkDeploySchedulerBackend: Registered executor: Actor[akka.tcp://sparkexecu...@server05.mydomain.com:55188/user/Executor#466888766] with ID 5 14/05/20 10:00:22 INFO SparkDeploySchedulerBackend: Registered executor: Actor[akka.tcp://sparkexecu...@server06.mydomain.com:34672/user/Executor#299947202] with ID 1 14/05/20 10:00:22 INFO BlockManagerMasterActor$BlockManagerInfo: Registering block manager server03.mydomain.com:57441 with 294.9 MB RAM 14/05/20 10:00:22 INFO BlockManagerMasterActor$BlockManagerInfo: Registering block manager server02.mydomain.com:55847 with 294.9 MB RAM 14/05/20 10:00:23 INFO BlockManagerMasterActor$BlockManagerInfo: Registering block manager server07.mydomain.com:52932 with 294.9 MB RAM 14/05/20 10:00:23 INFO BlockManagerMasterActor$BlockManagerInfo: Registering block manager server05.mydomain.com:45204 with 294.9 MB RAM 14/05/20 10:00:23 INFO BlockManagerMasterActor$BlockManagerInfo: Registering block manager server06.mydomain.com:57226 with 294.9 MB RAM 14/05/20 10:00:23 INFO SparkDeploySchedulerBackend: Registered executor: Actor[akka.tcp://sparkexecu...@server04.mydomain.com:37144/user/Executor#237933685] with ID 3 14/05/20 10:00:23 INFO BlockManagerMasterActor$BlockManagerInfo: Registering block manager server04.mydomain.com:55928 with 294.9 MB RAM 14/05/20 10:00:24 INFO MemoryStore: ensureFreeSpace(78267) called with curMem=0, maxMem=309225062 14/05/20 10:00:24 INFO MemoryStore: Block broadcast_0 stored as values to memory (estimated size 76.4 KB, free 294.8 MB) 14/05/20 10:00:24 INFO FileInputFormat: Total input paths to process : 1 14/05/20 10:00:24 INFO SparkContext: Starting job: count at <console>:13 14/05/20 10:00:24 INFO DAGScheduler: Got job 0 (count at <console>:13) with 2 output partitions (allowLocal=false) 14/05/20 10:00:24 INFO DAGScheduler: Final stage: Stage 0 (count at <console>:13) 14/05/20 10:00:24 INFO DAGScheduler: Parents of final stage: List() 14/05/20 10:00:24 INFO DAGScheduler: Missing parents: List() 14/05/20 10:00:24 INFO DAGScheduler: Submitting Stage 0 (MappedRDD[1] at textFile at <console>:13), which has no missing parents 14/05/20 10:00:24 INFO DAGScheduler: Submitting 2 missing tasks from Stage 0 (MappedRDD[1] at textFile at <console>:13) 14/05/20 10:00:24 INFO TaskSchedulerImpl: Adding task set 0.0 with 2 tasks 14/05/20 10:00:24 INFO TaskSetManager: Starting task 0.0:0 as TID 0 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 14/05/20 10:00:24 INFO TaskSetManager: Serialized task 0.0:0 as 1570 bytes in 5 ms 14/05/20 10:00:24 INFO TaskSetManager: Starting task 0.0:1 as TID 1 on executor 5: server05.mydomain.com (PROCESS_LOCAL) 14/05/20 10:00:24 INFO TaskSetManager: Serialized task 0.0:1 as 1570 bytes in 1 ms 14/05/20 10:00:25 WARN TaskSetManager: Lost TID 0 (task 0.0:0) 14/05/20 10:00:25 WARN TaskSetManager: Loss was due to java.io.FileNotFoundException java.io.FileNotFoundException: File file:/home/hadoop/scripts/clustering/deals.txt does not exist at org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:468) at org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:373) at org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSInputChecker.<init>(ChecksumFileSystem.java:138) at org.apache.hadoop.fs.ChecksumFileSystem.open(ChecksumFileSystem.java:334) at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:711) at org.apache.hadoop.mapred.LineRecordReader.<init>(LineRecordReader.java:93) at org.apache.hadoop.mapred.TextInputFormat.getRecordReader(TextInputFormat.java:54) at org.apache.spark.rdd.HadoopRDD$$anon$1.<init>(HadoopRDD.scala:156) at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:149) at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:64) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:241) at org.apache.spark.rdd.RDD.iterator(RDD.scala:232) at org.apache.spark.rdd.MappedRDD.compute(MappedRDD.scala:31) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:241) at org.apache.spark.rdd.RDD.iterator(RDD.scala:232) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:109) at org.apache.spark.scheduler.Task.run(Task.scala:53) at org.apache.spark.executor.Executor$TaskRunner$$anonfun$run$1.apply$mcV$sp(Executor.scala:211) at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:42) at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:41) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) at org.apache.spark.deploy.SparkHadoopUtil.runAsUser(SparkHadoopUtil.scala:41) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:176) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:744) 14/05/20 10:00:25 INFO TaskSetManager: Starting task 0.0:0 as TID 2 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 14/05/20 10:00:25 INFO TaskSetManager: Serialized task 0.0:0 as 1570 bytes in 0 ms 14/05/20 10:00:25 WARN TaskSetManager: Lost TID 2 (task 0.0:0) 14/05/20 10:00:25 INFO TaskSetManager: Loss was due to java.io.FileNotFoundException: File file:/home/hadoop/scripts/clustering/deals.txt does not exist [duplicate 1] 14/05/20 10:00:25 INFO TaskSetManager: Starting task 0.0:0 as TID 3 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 14/05/20 10:00:25 INFO TaskSetManager: Serialized task 0.0:0 as 1570 bytes in 0 ms 14/05/20 10:00:25 WARN TaskSetManager: Lost TID 1 (task 0.0:1) 14/05/20 10:00:25 INFO TaskSetManager: Loss was due to java.io.FileNotFoundException: File file:/home/hadoop/scripts/clustering/deals.txt does not exist [duplicate 2] 14/05/20 10:00:25 INFO TaskSetManager: Starting task 0.0:1 as TID 4 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 14/05/20 10:00:25 INFO TaskSetManager: Serialized task 0.0:1 as 1570 bytes in 1 ms 14/05/20 10:00:25 WARN TaskSetManager: Lost TID 3 (task 0.0:0) 14/05/20 10:00:25 INFO TaskSetManager: Loss was due to java.io.FileNotFoundException: File file:/home/hadoop/scripts/clustering/deals.txt does not exist [duplicate 3] 14/05/20 10:00:25 INFO TaskSetManager: Starting task 0.0:0 as TID 5 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 14/05/20 10:00:25 INFO TaskSetManager: Serialized task 0.0:0 as 1570 bytes in 1 ms 14/05/20 10:00:25 WARN TaskSetManager: Lost TID 4 (task 0.0:1) 14/05/20 10:00:25 INFO TaskSetManager: Loss was due to java.io.FileNotFoundException: File file:/home/hadoop/scripts/clustering/deals.txt does not exist [duplicate 4] 14/05/20 10:00:25 INFO TaskSetManager: Starting task 0.0:1 as TID 6 on executor 2: server03.mydomain.com (PROCESS_LOCAL) 14/05/20 10:00:25 INFO TaskSetManager: Serialized task 0.0:1 as 1570 bytes in 0 ms 14/05/20 10:00:25 WARN TaskSetManager: Lost TID 5 (task 0.0:0) 14/05/20 10:00:25 INFO TaskSetManager: Loss was due to java.io.FileNotFoundException: File file:/home/hadoop/scripts/clustering/deals.txt does not exist [duplicate 5] 14/05/20 10:00:25 ERROR TaskSetManager: Task 0.0:0 failed 4 times; aborting job 14/05/20 10:00:25 INFO DAGScheduler: Failed to run count at <console>:13 14/05/20 10:00:25 INFO TaskSetManager: Loss was due to java.io.FileNotFoundException: File file:/home/hadoop/scripts/clustering/deals.txt does not exist [duplicate 6] 14/05/20 10:00:25 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose tasks have all completed, from pool org.apache.spark.SparkException: Job aborted: Task 0.0:0 failed 4 times (most recent failure: Exception failure: java.io.FileNotFoundException: File file:/home/hadoop/scripts/clustering/deals.txt does not exist) at org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1020) at org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1018) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$abortStage(DAGScheduler.scala:1018) at org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) at org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) at scala.Option.foreach(Option.scala:236) at org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:604) at org.apache.spark.scheduler.DAGScheduler$$anonfun$start$1$$anon$2$$anonfun$receive$1.applyOrElse(DAGScheduler.scala:190) at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498) at akka.actor.ActorCell.invoke(ActorCell.scala:456) at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237) at akka.dispatch.Mailbox.run(Mailbox.scala:219) at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) But I can load a hdfs file from scala-shell: scala> sc.textFile("hdfs://server01.mydomain.com:/user/hadoop/results/clustering/deals.txt").count() 14/05/20 10:02:57 INFO MemoryStore: ensureFreeSpace(111825) called with curMem=681339, maxMem=309225062 14/05/20 10:02:57 INFO MemoryStore: Block broadcast_7 stored as values to memory (estimated size 109.2 KB, free 294.1 MB) 14/05/20 10:02:57 INFO FileInputFormat: Total input paths to process : 1 14/05/20 10:02:57 INFO SparkContext: Starting job: count at <console>:13 14/05/20 10:02:57 INFO DAGScheduler: Got job 6 (count at <console>:13) with 2 output partitions (allowLocal=false) 14/05/20 10:02:57 INFO DAGScheduler: Final stage: Stage 6 (count at <console>:13) 14/05/20 10:02:57 INFO DAGScheduler: Parents of final stage: List() 14/05/20 10:02:57 INFO DAGScheduler: Missing parents: List() 14/05/20 10:02:57 INFO DAGScheduler: Submitting Stage 6 (MappedRDD[15] at textFile at <console>:13), which has no missing parents 14/05/20 10:02:57 INFO DAGScheduler: Submitting 2 missing tasks from Stage 6 (MappedRDD[15] at textFile at <console>:13) 14/05/20 10:02:57 INFO TaskSchedulerImpl: Adding task set 6.0 with 2 tasks 14/05/20 10:02:57 INFO TaskSetManager: Starting task 6.0:0 as TID 47 on executor 2: server03.mydomain.com (NODE_LOCAL) 14/05/20 10:02:57 INFO TaskSetManager: Serialized task 6.0:0 as 1688 bytes in 0 ms 14/05/20 10:02:57 INFO TaskSetManager: Starting task 6.0:1 as TID 48 on executor 5: server05.mydomain.com (NODE_LOCAL) 14/05/20 10:02:57 INFO TaskSetManager: Serialized task 6.0:1 as 1688 bytes in 1 ms 14/05/20 10:02:58 INFO DAGScheduler: Completed ResultTask(6, 0) 14/05/20 10:02:58 INFO TaskSetManager: Finished TID 47 in 454 ms on server03.mydomain.com (progress: 1/2) 14/05/20 10:02:58 INFO TaskSetManager: Finished TID 48 in 504 ms on server05.mydomain.com (progress: 2/2) 14/05/20 10:02:58 INFO DAGScheduler: Completed ResultTask(6, 1) 14/05/20 10:02:58 INFO TaskSchedulerImpl: Removed TaskSet 6.0, whose tasks have all completed, from pool 14/05/20 10:02:58 INFO DAGScheduler: Stage 6 (count at <console>:13) finished in 0.506 s 14/05/20 10:02:58 INFO SparkContext: Job finished: count at <console>:13, took 0.513880747 s res8: Long = 101 I also tried to load file into an ArrayList and use parallelize(), it worked too. But I don't understand why I can't just load a file and count it? Could someone please help me understand the problem? Thanks in advance... -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Problem-with-loading-files-Loss-was-due-to-java-io-EOFException-java-io-EOFException-tp6090.html Sent from the Apache Spark User List mailing list archive at Nabble.com.