Hi all,

I've a single machine with 8 cores and 8g mem. I've deployed the standalone spark on the machine and successfully run the examples.

Now I'm trying to write some simple java codes. I just read a local file (23M) into string list and use JavaRDD<String> rdds = sparkContext.paralellize() method to get the corresponding rdd. And I asked to run rdds.count(). But the program just stopped on the count(). The last log info is:

   14/05/27 14:13:16 INFO SparkContext: Starting job: count at
   RDDTest.java:40
   14/05/27 14:13:16 INFO DAGScheduler: Got job 0 (count at
   RDDTest.java:40) with 2 output partitions (allowLocal=false)
   14/05/27 14:13:16 INFO DAGScheduler: Final stage: Stage 0 (count at
   RDDTest.java:40)
   14/05/27 14:13:16 INFO DAGScheduler: Parents of final stage: List()
   14/05/27 14:13:16 INFO DAGScheduler: Missing parents: List()
   14/05/27 14:13:16 INFO DAGScheduler: Submitting Stage 0
   (ParallelCollectionRDD[0] at parallelize at RDDTest.java:37), which
   has no missing parents
   14/05/27 14:13:16 INFO SparkDeploySchedulerBackend: Connected to
   Spark cluster with app ID app-20140527141316-0003
   14/05/27 14:13:16 INFO AppClient$ClientActor: Executor added:
   app-20140527141316-0003/0 on worker-20140526221107-spark-35303
   (spark:35303) with 8 cores
   14/05/27 14:13:16 INFO SparkDeploySchedulerBackend: Granted executor
   ID app-20140527141316-0003/0 on hostPort spark:35303 with 8 cores,
   1024.0 MB RAM
   14/05/27 14:13:16 INFO AppClient$ClientActor: Executor updated:
   app-20140527141316-0003/0 is now RUNNING
   14/05/27 14:13:16 INFO DAGScheduler: Submitting 2 missing tasks from
   Stage 0 (ParallelCollectionRDD[0] at parallelize at RDDTest.java:37)
   14/05/27 14:13:16 INFO TaskSchedulerImpl: Adding task set 0.0 with 2
   tasks
   14/05/27 14:13:17 INFO SparkDeploySchedulerBackend: Registered
   executor:
   Actor[akka.tcp://sparkExecutor@spark:34279/user/Executor#196489168]
   with ID 0
   14/05/27 14:13:17 INFO TaskSetManager: Starting task 0.0:0 as TID 0
   on executor 0: spark (PROCESS_LOCAL)
   14/05/27 14:13:17 INFO TaskSetManager: Serialized task 0.0:0 as
   12993529 bytes in 127 ms
   14/05/27 14:13:17 INFO TaskSetManager: Starting task 0.0:1 as TID 1
   on executor 0: spark (PROCESS_LOCAL)
   14/05/27 14:13:17 INFO TaskSetManager: Serialized task 0.0:1 as
   13006417 bytes in 74 ms
   14/05/27 14:13:17 INFO BlockManagerMasterActor$BlockManagerInfo:
   Registering block manager spark:37617 with 589.2 MB RAM

I tried to figure out what's going on, but just can't. Could any please give me some suggestions and point out some possible issues?

Best Regards,
Min

Reply via email to