Dear Community,

       Please ignore my last post about Spark SQL.

       When I run:
val file = sc.textFile("./README.md")
val count = file.flatMap(line => line.split(" ")).map(word => (word, 
1)).reduceByKey(_+_)
count.collect()
‍
        it happends too.

        is there any possible reason for that? we make have some adjustment in 
network last night

                                                                                
             Chen Weikeng
14/10/09 20:45:23 ERROR Executor: Exception in task ID 1 
java.lang.NullPointerException         at 
org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:571)    at 
org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:571)    at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)     at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)      at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:229)     at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)     at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)      at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:229)     at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)     at 
org.apache.spark.sql.SchemaRDD.compute(SchemaRDD.scala:116)  at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)      at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:229)     at 
org.apache.spark.rdd.MappedRDD.compute(MappedRDD.scala:31)   at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)      at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:229)     at 
org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:111)  at 
org.apache.spark.scheduler.Task.run(Task.scala:51)   at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:183)        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) 
     at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) 
     at java.lang.Thread.run(Thread.java:745) 14/10/09 20:45:23 INFO 
CoarseGrainedExecutorBackend: Got assigned task 2 14/10/09 20:45:23 INFO 
Executor: Running task ID 2 14/10/09 20:45:23 DEBUG BlockManager: Getting local 
block broadcast_0 14/10/09 20:45:23 DEBUG BlockManager: Level for block 
broadcast_0 is StorageLevel(true, true, false, true, 1) 14/10/09 20:45:23 DEBUG 
BlockManager: Getting block broadcast_0 from memory 14/10/09 20:45:23 INFO 
BlockManager: Found block broadcast_0 locally 14/10/09 20:45:23 DEBUG Executor: 
Task 2's epoch is 0 14/10/09 20:45:23 INFO HadoopRDD: Input split: 
file:/public/rdma14/app/spark-rdma/examples/src/main/resources/people.txt:16+16 
14/10/09 20:45:23 ERROR Executor: Exception in task ID 2 
java.lang.NullPointerException  at 
org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:571)    at 
org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:571)    at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)     at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)      at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:229)     at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)     at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)      at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:229)     at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)     at 
org.apache.spark.sql.SchemaRDD.compute(SchemaRDD.scala:116)  at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)      at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:229)     at 
org.apache.spark.rdd.MappedRDD.compute(MappedRDD.scala:31)   at 
org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)      at 
org.apache.spark.rdd.RDD.iterator(RDD.scala:229)     at 
org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:111)  at 
org.apache.spark.scheduler.Task.run(Task.scala:51)   at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:183)        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) 
     at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) 
     at java.lang.Thread.run(Thread.java:745)‍

Reply via email to