Replying to my own email in case someone has the same or similar issue. On a hunch I ran this against my Linux (Ubuntu 14.04 with JDK 8) box. Not only did "bin/run-example SparkPi" run without any problems, it also provided a very helpful message in the output.
15/05/23 08:35:15 WARN Utils: Your hostname, tsunami resolves to a loopback address: 127.0.1.1; using 10.0.0.10 instead (on interface wlan0) 15/05/23 08:35:15 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address So I went back to my Mac, set SPARK_LOCAL_IP=127.0.0.1 and everything runs fine now. To make this permanent I put this in conf/spark-env.sh. -sujit On Sat, May 23, 2015 at 8:14 AM, Sujit Pal <sujitatgt...@gmail.com> wrote: > Hello all, > > This is probably me doing something obviously wrong, would really > appreciate some pointers on how to fix this. > > I installed spark-1.3.1-bin-hadoop2.6.tgz from the Spark download page [ > https://spark.apache.org/downloads.html] and just untarred it on a local > drive. I am on Mac OSX 10.9.5 and the JDK is 1.8.0_40. > > I ran the following commands (the first 3 run succesfully, I mention it > here to rule out any possibility of it being an obviously bad install). > > 1) laptop$ bin/spark-shell > > scala> sc.parallelize(1 to 100).count() > > res0: Long = 100 > > scala> exit > > 2) laptop$ bin/pyspark > > >>> sc.parallelize(range(100)).count() > > 100 > > >>> quit() > > 3) laptop$ bin/spark-submit examples/src/main/python/pi.py > > Pi is roughly 3.142800 > > 4) laptop$ bin/run-example SparkPi > > This hangs at this line (full stack trace is provided at the end of this > mail) > > 15/05/23 07:52:10 INFO Executor: Fetching > http://10.0.0.5:51575/jars/spark-examples-1.3.1-hadoop2.6.0.jar with > timestamp 1432392670140 > > 15/05/23 07:52:10 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID > 0) > > java.net.SocketTimeoutException: connect timed out > > ... > > and finally dies with this message: > > Exception in thread "main" org.apache.spark.SparkException: Job aborted > due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent > failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): > java.net.SocketTimeoutException: connect timed out > > > I checked with ifconfig -a on my box, 10.0.0.5 is my IP address on my > local network. > > en0: flags=8863<UP,BROADCAST,SMART,RUNNING,SIMPLEX,MULTICAST> mtu 1500 > > ether 34:36:3b:d2:b0:f4 > > inet 10.0.0.5 netmask 0xffffff00 broadcast 10.0.0.255 > > media: autoselect > > status: active > > > I think perhaps there may be some configuration I am missing. Being able > to run jobs locally (without HDFS or creating a cluster) is essential for > development, and the examples come from the Spark 1.3.1 Quick Start page [ > https://spark.apache.org/docs/latest/quick-start.html], so this is > probably something to do with my environment. > > > Thanks in advance for any help you can provide. > > -sujit > > ===== > > Full output of SparkPi run (including stack trace) follows: > > Using Spark's default log4j profile: > org/apache/spark/log4j-defaults.properties > > 15/05/23 08:08:55 INFO SparkContext: Running Spark version 1.3.1 > > 15/05/23 08:08:57 WARN NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > > 15/05/23 08:08:57 INFO SecurityManager: Changing view acls to: palsujit > > 15/05/23 08:08:57 INFO SecurityManager: Changing modify acls to: palsujit > > 15/05/23 08:08:57 INFO SecurityManager: SecurityManager: authentication > disabled; ui acls disabled; users with view permissions: Set(palsujit); > users with modify permissions: Set(palsujit) > > 15/05/23 08:08:57 INFO Slf4jLogger: Slf4jLogger started > > 15/05/23 08:08:57 INFO Remoting: Starting remoting > > 15/05/23 08:08:58 INFO Remoting: Remoting started; listening on addresses > :[akka.tcp://sparkDriver@10.0.0.5:52008] > > 15/05/23 08:08:58 INFO Utils: Successfully started service 'sparkDriver' > on port 52008. > > 15/05/23 08:08:58 INFO SparkEnv: Registering MapOutputTracker > > 15/05/23 08:08:58 INFO SparkEnv: Registering BlockManagerMaster > > 15/05/23 08:08:58 INFO DiskBlockManager: Created local directory at > /var/folders/z8/s_crq_2j2rqb9mv_4j8djsjnx359l2/T/spark-d97baddf-1b6f-41db-92bb-f82ab5184cb7/blockmgr-4ef3a194-1929-4dd3-a0e5-215175d8e41a > > 15/05/23 08:08:58 INFO MemoryStore: MemoryStore started with capacity > 265.1 MB > > 15/05/23 08:08:58 INFO HttpFileServer: HTTP File server directory is > /var/folders/z8/s_crq_2j2rqb9mv_4j8djsjnx359l2/T/spark-fdf36480-def0-44b7-9942-098d9ef3e2b4/httpd-e494852a-7d61-4441-8b80-566d9f820afb > > 15/05/23 08:08:58 INFO HttpServer: Starting HTTP Server > > 15/05/23 08:08:58 INFO Server: jetty-8.y.z-SNAPSHOT > > 15/05/23 08:08:58 INFO AbstractConnector: Started > SocketConnector@0.0.0.0:52009 > > 15/05/23 08:08:58 INFO Utils: Successfully started service 'HTTP file > server' on port 52009. > > 15/05/23 08:08:58 INFO SparkEnv: Registering OutputCommitCoordinator > > 15/05/23 08:08:58 INFO Server: jetty-8.y.z-SNAPSHOT > > 15/05/23 08:08:58 INFO AbstractConnector: Started > SelectChannelConnector@0.0.0.0:4040 > > 15/05/23 08:08:58 INFO Utils: Successfully started service 'SparkUI' on > port 4040. > > 15/05/23 08:08:58 INFO SparkUI: Started SparkUI at http://10.0.0.5:4040 > > 15/05/23 08:08:58 INFO SparkContext: Added JAR > file:/Users/palsujit/Software/spark-1.3.1-bin-hadoop2.6/lib/spark-examples-1.3.1-hadoop2.6.0.jar > at http://10.0.0.5:52009/jars/spark-examples-1.3.1-hadoop2.6.0.jar with > timestamp 1432393738514 > > 15/05/23 08:08:58 INFO Executor: Starting executor ID <driver> on host > localhost > > 15/05/23 08:08:58 INFO AkkaUtils: Connecting to HeartbeatReceiver: > akka.tcp://sparkDriver@10.0.0.5:52008/user/HeartbeatReceiver > > 15/05/23 08:08:58 INFO NettyBlockTransferService: Server created on 52010 > > 15/05/23 08:08:58 INFO BlockManagerMaster: Trying to register BlockManager > > 15/05/23 08:08:58 INFO BlockManagerMasterActor: Registering block manager > localhost:52010 with 265.1 MB RAM, BlockManagerId(<driver>, localhost, > 52010) > > 15/05/23 08:08:58 INFO BlockManagerMaster: Registered BlockManager > > 15/05/23 08:08:58 INFO SparkContext: Starting job: reduce at > SparkPi.scala:35 > > 15/05/23 08:08:58 INFO DAGScheduler: Got job 0 (reduce at > SparkPi.scala:35) with 2 output partitions (allowLocal=false) > > 15/05/23 08:08:58 INFO DAGScheduler: Final stage: Stage 0(reduce at > SparkPi.scala:35) > > 15/05/23 08:08:58 INFO DAGScheduler: Parents of final stage: List() > > 15/05/23 08:08:58 INFO DAGScheduler: Missing parents: List() > > 15/05/23 08:08:58 INFO DAGScheduler: Submitting Stage 0 > (MapPartitionsRDD[1] at map at SparkPi.scala:31), which has no missing > parents > > 15/05/23 08:08:58 INFO MemoryStore: ensureFreeSpace(1848) called with > curMem=0, maxMem=278019440 > > 15/05/23 08:08:58 INFO MemoryStore: Block broadcast_0 stored as values in > memory (estimated size 1848.0 B, free 265.1 MB) > > 15/05/23 08:08:58 INFO MemoryStore: ensureFreeSpace(1296) called with > curMem=1848, maxMem=278019440 > > 15/05/23 08:08:58 INFO MemoryStore: Block broadcast_0_piece0 stored as > bytes in memory (estimated size 1296.0 B, free 265.1 MB) > > 15/05/23 08:08:58 INFO BlockManagerInfo: Added broadcast_0_piece0 in > memory on localhost:52010 (size: 1296.0 B, free: 265.1 MB) > > 15/05/23 08:08:58 INFO BlockManagerMaster: Updated info of block > broadcast_0_piece0 > > 15/05/23 08:08:58 INFO SparkContext: Created broadcast 0 from broadcast at > DAGScheduler.scala:839 > > 15/05/23 08:08:58 INFO DAGScheduler: Submitting 2 missing tasks from Stage > 0 (MapPartitionsRDD[1] at map at SparkPi.scala:31) > > 15/05/23 08:08:58 INFO TaskSchedulerImpl: Adding task set 0.0 with 2 tasks > > 15/05/23 08:08:58 INFO TaskSetManager: Starting task 0.0 in stage 0.0 (TID > 0, localhost, PROCESS_LOCAL, 1333 bytes) > > 15/05/23 08:08:58 INFO TaskSetManager: Starting task 1.0 in stage 0.0 (TID > 1, localhost, PROCESS_LOCAL, 1333 bytes) > > 15/05/23 08:08:58 INFO Executor: Running task 0.0 in stage 0.0 (TID 0) > > 15/05/23 08:08:58 INFO Executor: Running task 1.0 in stage 0.0 (TID 1) > > 15/05/23 08:08:58 INFO Executor: Fetching > http://10.0.0.5:52009/jars/spark-examples-1.3.1-hadoop2.6.0.jar with > timestamp 1432393738514 > > 15/05/23 08:09:58 INFO Executor: Fetching > http://10.0.0.5:52009/jars/spark-examples-1.3.1-hadoop2.6.0.jar with > timestamp 1432393738514 > > 15/05/23 08:09:58 ERROR Executor: Exception in task 1.0 in stage 0.0 (TID > 1) > > java.net.SocketTimeoutException: connect timed out > > at java.net.PlainSocketImpl.socketConnect(Native Method) > > at > java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:345) > > at > java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206) > > at > java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188) > > at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) > > at java.net.Socket.connect(Socket.java:589) > > at sun.net.NetworkClient.doConnect(NetworkClient.java:175) > > at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) > > at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) > > at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) > > at sun.net.www.http.HttpClient.New(HttpClient.java:308) > > at sun.net.www.http.HttpClient.New(HttpClient.java:326) > > at > sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1168) > > at > sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1104) > > at > sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:998) > > at > sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:932) > > at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:610) > > at org.apache.spark.util.Utils$.fetchFile(Utils.scala:431) > > at > org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:374) > > at > org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:366) > > at > scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) > > at > scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) > > at > scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) > > at > scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226) > > at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) > > at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) > > at > scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) > > at org.apache.spark.executor.Executor.org > $apache$spark$executor$Executor$$updateDependencies(Executor.scala:366) > > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:184) > > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > > at java.lang.Thread.run(Thread.java:745) > > 15/05/23 08:09:58 WARN TaskSetManager: Lost task 1.0 in stage 0.0 (TID 1, > localhost): java.net.SocketTimeoutException: connect timed out > > at java.net.PlainSocketImpl.socketConnect(Native Method) > > at > java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:345) > > at > java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206) > > at > java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188) > > at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) > > at java.net.Socket.connect(Socket.java:589) > > at sun.net.NetworkClient.doConnect(NetworkClient.java:175) > > at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) > > at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) > > at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) > > at sun.net.www.http.HttpClient.New(HttpClient.java:308) > > at sun.net.www.http.HttpClient.New(HttpClient.java:326) > > at > sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1168) > > at > sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1104) > > at > sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:998) > > at > sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:932) > > at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:610) > > at org.apache.spark.util.Utils$.fetchFile(Utils.scala:431) > > at > org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:374) > > at > org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:366) > > at > scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) > > at > scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) > > at > scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) > > at > scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226) > > at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) > > at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) > > at > scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) > > at org.apache.spark.executor.Executor.org > $apache$spark$executor$Executor$$updateDependencies(Executor.scala:366) > > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:184) > > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > > at java.lang.Thread.run(Thread.java:745) > > > 15/05/23 08:09:58 ERROR TaskSetManager: Task 1 in stage 0.0 failed 1 > times; aborting job > > 15/05/23 08:09:58 INFO TaskSchedulerImpl: Cancelling stage 0 > > 15/05/23 08:09:58 INFO Executor: Executor is trying to kill task 0.0 in > stage 0.0 (TID 0) > > 15/05/23 08:09:58 INFO TaskSchedulerImpl: Stage 0 was cancelled > > 15/05/23 08:09:58 INFO DAGScheduler: Stage 0 (reduce at SparkPi.scala:35) > failed in 60.065 s > > 15/05/23 08:09:58 INFO DAGScheduler: Job 0 failed: reduce at > SparkPi.scala:35, took 60.191508 s > > Exception in thread "main" org.apache.spark.SparkException: Job aborted > due to stage failure: Task 1 in stage 0.0 failed 1 times, most recent > failure: Lost task 1.0 in stage 0.0 (TID 1, localhost): > java.net.SocketTimeoutException: connect timed out > > at java.net.PlainSocketImpl.socketConnect(Native Method) > > at > java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:345) > > at > java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206) > > at > java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188) > > at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) > > at java.net.Socket.connect(Socket.java:589) > > at sun.net.NetworkClient.doConnect(NetworkClient.java:175) > > at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) > > at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) > > at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) > > at sun.net.www.http.HttpClient.New(HttpClient.java:308) > > at sun.net.www.http.HttpClient.New(HttpClient.java:326) > > at > sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1168) > > at > sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1104) > > at > sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:998) > > at > sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:932) > > at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:610) > > at org.apache.spark.util.Utils$.fetchFile(Utils.scala:431) > > at > org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:374) > > at > org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:366) > > at > scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) > > at > scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) > > at > scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) > > at > scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226) > > at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) > > at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) > > at > scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) > > at org.apache.spark.executor.Executor.org > $apache$spark$executor$Executor$$updateDependencies(Executor.scala:366) > > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:184) > > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > > at java.lang.Thread.run(Thread.java:745) > > > Driver stacktrace: > > at org.apache.spark.scheduler.DAGScheduler.org > $apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1204) > > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1193) > > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1192) > > at > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) > > at > org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1192) > > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:693) > > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:693) > > at scala.Option.foreach(Option.scala:236) > > at > org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:693) > > at > org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1393) > > at > org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1354) > > at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) > > >