Hi Jacek/All, I restarted my terminal and then i try spark-submit and again getting those errors. How do i see how many "runtimes" are running and how to have only one? some how my spark 1.6 and spark 2.0 are conflicting. how to fix it?
i installed spark 1.6 earlier using this steps http://genomegeek.blogspot.com/2014/11/how-to-install-apache-spark-on-mac-os-x.html i installed spark 2.0 using these steps http://blog.weetech.co/2015/08/light-learning-apache-spark.html Here is the for run-example m-C02KL0B1FFT4:bin vr$ ./run-example SparkPi Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties 16/09/26 09:11:00 INFO SparkContext: Running Spark version 2.0.0 16/09/26 09:11:00 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 16/09/26 09:11:00 INFO SecurityManager: Changing view acls to: vr 16/09/26 09:11:00 INFO SecurityManager: Changing modify acls to: vr 16/09/26 09:11:00 INFO SecurityManager: Changing view acls groups to: 16/09/26 09:11:00 INFO SecurityManager: Changing modify acls groups to: 16/09/26 09:11:00 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(vr); groups with view permissions: Set(); users with modify permissions: Set(vr); groups with modify permissions: Set() 16/09/26 09:11:01 INFO Utils: Successfully started service 'sparkDriver' on port 59323. 16/09/26 09:11:01 INFO SparkEnv: Registering MapOutputTracker 16/09/26 09:11:01 INFO SparkEnv: Registering BlockManagerMaster 16/09/26 09:11:01 INFO DiskBlockManager: Created local directory at /private/var/folders/23/ycbtxh8s551gzlsgj8q647d88gsjgb/T/blockmgr-d0d6dfea-2c97-4337-8e7d-0bbcb141f4c9 16/09/26 09:11:01 INFO MemoryStore: MemoryStore started with capacity 366.3 MB 16/09/26 09:11:01 INFO SparkEnv: Registering OutputCommitCoordinator 16/09/26 09:11:01 WARN Utils: Service 'SparkUI' could not bind on port 4040. Attempting port 4041. 16/09/26 09:11:01 INFO Utils: Successfully started service 'SparkUI' on port 4041. 16/09/26 09:11:01 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at http://192.168.1.3:4041 16/09/26 09:11:01 INFO SparkContext: Added JAR file:/Users/vr/Downloads/spark-2.0.0/examples/target/scala-2.11/jars/scopt_2.11-3.3.0.jar at spark://192.168.1.3:59323/jars/scopt_2.11-3.3.0.jar with timestamp 1474906261472 16/09/26 09:11:01 INFO SparkContext: Added JAR file:/Users/vr/Downloads/spark-2.0.0/examples/target/scala-2.11/jars/spark-examples_2.11-2.0.0.jar at spark://192.168.1.3:59323/jars/spark-examples_2.11-2.0.0.jar with timestamp 1474906261473 16/09/26 09:11:01 INFO Executor: Starting executor ID driver on host localhost 16/09/26 09:11:01 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 59324. 16/09/26 09:11:01 INFO NettyBlockTransferService: Server created on 192.168.1.3:59324 16/09/26 09:11:01 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(driver, 192.168.1.3, 59324) 16/09/26 09:11:01 INFO BlockManagerMasterEndpoint: Registering block manager 192.168.1.3:59324 with 366.3 MB RAM, BlockManagerId(driver, 192.168.1.3, 59324) 16/09/26 09:11:01 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(driver, 192.168.1.3, 59324) 16/09/26 09:11:01 WARN SparkContext: Use an existing SparkContext, some configuration may not take effect. 16/09/26 09:11:01 INFO SharedState: Warehouse path is 'file:/Users/vr/Downloads/spark-2.0.0/bin/spark-warehouse'. 16/09/26 09:11:01 INFO SparkContext: Starting job: reduce at SparkPi.scala:38 16/09/26 09:11:02 INFO DAGScheduler: Got job 0 (reduce at SparkPi.scala:38) with 2 output partitions 16/09/26 09:11:02 INFO DAGScheduler: Final stage: ResultStage 0 (reduce at SparkPi.scala:38) 16/09/26 09:11:02 INFO DAGScheduler: Parents of final stage: List() 16/09/26 09:11:02 INFO DAGScheduler: Missing parents: List() 16/09/26 09:11:02 INFO DAGScheduler: Submitting ResultStage 0 (MapPartitionsRDD[1] at map at SparkPi.scala:34), which has no missing parents 16/09/26 09:11:02 INFO MemoryStore: Block broadcast_0 stored as values in memory (estimated size 1832.0 B, free 366.3 MB) 16/09/26 09:11:02 INFO MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 1169.0 B, free 366.3 MB) 16/09/26 09:11:02 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory on 192.168.1.3:59324 (size: 1169.0 B, free: 366.3 MB) 16/09/26 09:11:02 INFO SparkContext: Created broadcast 0 from broadcast at DAGScheduler.scala:1012 16/09/26 09:11:02 INFO DAGScheduler: Submitting 2 missing tasks from ResultStage 0 (MapPartitionsRDD[1] at map at SparkPi.scala:34) 16/09/26 09:11:02 INFO TaskSchedulerImpl: Adding task set 0.0 with 2 tasks 16/09/26 09:11:02 INFO TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0, localhost, partition 0, PROCESS_LOCAL, 5474 bytes) 16/09/26 09:11:02 INFO TaskSetManager: Starting task 1.0 in stage 0.0 (TID 1, localhost, partition 1, PROCESS_LOCAL, 5474 bytes) 16/09/26 09:11:02 INFO Executor: Running task 1.0 in stage 0.0 (TID 1) 16/09/26 09:11:02 INFO Executor: Running task 0.0 in stage 0.0 (TID 0) 16/09/26 09:11:02 INFO Executor: Fetching spark:// 192.168.1.3:59323/jars/scopt_2.11-3.3.0.jar with timestamp 1474906261472 16/09/26 09:12:17 INFO Executor: Fetching spark:// 192.168.1.3:59323/jars/scopt_2.11-3.3.0.jar with timestamp 1474906261472 16/09/26 09:12:17 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID 0) java.io.IOException: Failed to connect to /192.168.1.3:59323 java.io.IOException: Failed to connect to /192.168.1.3:59323 at org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:228) at org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:179) at org.apache.spark.rpc.netty.NettyRpcEnv.downloadClient(NettyRpcEnv.scala:358) at org.apache.spark.rpc.netty.NettyRpcEnv.openChannel(NettyRpcEnv.scala:324) at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:633) at org.apache.spark.util.Utils$.fetchFile(Utils.scala:459) at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:488) at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:480) at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733) at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99) at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:99) at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230) at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40) at scala.collection.mutable.HashMap.foreach(HashMap.scala:99) at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732) at org.apache.spark.executor.Executor.org $apache$spark$executor$Executor$$updateDependencies(Executor.scala:480) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:252) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Caused by: java.net.ConnectException: Operation timed out: / 192.168.1.3:59323 On Sun, Sep 25, 2016 at 8:32 AM, Jacek Laskowski <ja...@japila.pl> wrote: > Hi, > > How did you install Spark 1.6? It's usually as simple as rm -rf > $SPARK_1.6_HOME, but it really depends on how you installed it in the > first place. > > Pozdrawiam, > Jacek Laskowski > ---- > https://medium.com/@jaceklaskowski/ > Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark > Follow me at https://twitter.com/jaceklaskowski > > > On Sun, Sep 25, 2016 at 4:32 PM, vr spark <vrspark...@gmail.com> wrote: > > yes, i have both spark 1.6 and spark 2.0. > > I unset the spark home environment variable and pointed spark submit to > 2.0. > > Its working now. > > > > How do i uninstall/remove spark 1.6 from mac? > > > > Thanks > > > > > > On Sun, Sep 25, 2016 at 4:28 AM, Jacek Laskowski <ja...@japila.pl> > wrote: > >> > >> Hi, > >> > >> Can you execute run-example SparkPi with your Spark installation? > >> > >> Also, see the logs: > >> > >> 16/09/24 23:15:15 WARN Utils: Service 'SparkUI' could not bind on port > >> 4040. Attempting port 4041. > >> > >> 16/09/24 23:15:15 INFO Utils: Successfully started service 'SparkUI' > >> on port 4041. > >> > >> You've got two Spark runtimes up that may or may not contribute to the > >> issue. > >> > >> Pozdrawiam, > >> Jacek Laskowski > >> ---- > >> https://medium.com/@jaceklaskowski/ > >> Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark > >> Follow me at https://twitter.com/jaceklaskowski > >> > >> > >> On Sun, Sep 25, 2016 at 8:36 AM, vr spark <vrspark...@gmail.com> wrote: > >> > Hi, > >> > I have this simple scala app which works fine when i run it as scala > >> > application from the scala IDE for eclipse. > >> > But when i export is as jar and run it from spark-submit i am getting > >> > below > >> > error. Please suggest > >> > > >> > bin/spark-submit --class com.x.y.vr.spark.first.SimpleApp test.jar > >> > > >> > 16/09/24 23:15:15 WARN Utils: Service 'SparkUI' could not bind on port > >> > 4040. > >> > Attempting port 4041. > >> > > >> > 16/09/24 23:15:15 INFO Utils: Successfully started service 'SparkUI' > on > >> > port > >> > 4041. > >> > > >> > 16/09/24 23:15:15 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started > at > >> > http://192.168.1.3:4041 > >> > > >> > 16/09/24 23:15:15 INFO SparkContext: Added JAR > >> > file:/Users/vr/Downloads/spark-2.0.0/test.jar at > >> > spark://192.168.1.3:59263/jars/test.jar with timestamp 1474784115210 > >> > > >> > 16/09/24 23:15:15 INFO Executor: Starting executor ID driver on host > >> > localhost > >> > > >> > 16/09/24 23:15:15 INFO Utils: Successfully started service > >> > 'org.apache.spark.network.netty.NettyBlockTransferService' on port > >> > 59264. > >> > > >> > 16/09/24 23:15:15 INFO NettyBlockTransferService: Server created on > >> > 192.168.1.3:59264 > >> > > >> > 16/09/24 23:15:16 INFO TaskSetManager: Starting task 0.0 in stage 0.0 > >> > (TID > >> > 0, localhost, partition 0, PROCESS_LOCAL, 5354 bytes) > >> > > >> > 16/09/24 23:15:16 INFO TaskSetManager: Starting task 1.0 in stage 0.0 > >> > (TID > >> > 1, localhost, partition 1, PROCESS_LOCAL, 5354 bytes) > >> > > >> > 16/09/24 23:15:16 INFO Executor: Running task 0.0 in stage 0.0 (TID 0) > >> > > >> > 16/09/24 23:15:16 INFO Executor: Running task 1.0 in stage 0.0 (TID 1) > >> > > >> > 16/09/24 23:15:16 INFO Executor: Fetching > >> > spark://192.168.1.3:59263/jars/test.jar with timestamp 1474784115210 > >> > > >> > 16/09/24 23:16:31 INFO Executor: Fetching > >> > spark://192.168.1.3:59263/jars/test.jar with timestamp 1474784115210 > >> > > >> > 16/09/24 23:16:31 ERROR Executor: Exception in task 1.0 in stage 0.0 > >> > (TID 1) > >> > > >> > java.io.IOException: Failed to connect to /192.168.1.3:59263 > >> > > >> > at > >> > > >> > org.apache.spark.network.client.TransportClientFactory.createClient( > TransportClientFactory.java:228) > >> > > >> > at > >> > > >> > org.apache.spark.network.client.TransportClientFactory.createClient( > TransportClientFactory.java:179) > >> > > >> > at > >> > > >> > org.apache.spark.rpc.netty.NettyRpcEnv.downloadClient( > NettyRpcEnv.scala:358) > >> > > >> > at > >> > org.apache.spark.rpc.netty.NettyRpcEnv.openChannel( > NettyRpcEnv.scala:324) > >> > > >> > at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:633) > >> > > >> > at org.apache.spark.util.Utils$.fetchFile(Utils.scala:459) > >> > > >> > at > >> > > >> > org.apache.spark.executor.Executor$$anonfun$org$apache$ > spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:488) > >> > > >> > at > >> > > >> > org.apache.spark.executor.Executor$$anonfun$org$apache$ > spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:480) > >> > > >> > at > >> > > >> > scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply( > TraversableLike.scala:733) > >> > > >> > at > >> > > >> > scala.collection.mutable.HashMap$$anonfun$foreach$1. > apply(HashMap.scala:99) > >> > > >> > at > >> > > >> > scala.collection.mutable.HashMap$$anonfun$foreach$1. > apply(HashMap.scala:99) > >> > > >> > at > >> > > >> > scala.collection.mutable.HashTable$class.foreachEntry( > HashTable.scala:230) > >> > > >> > at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40) > >> > > >> > at scala.collection.mutable.HashMap.foreach(HashMap.scala:99) > >> > > >> > at > >> > > >> > scala.collection.TraversableLike$WithFilter. > foreach(TraversableLike.scala:732) > >> > > >> > at > >> > > >> > org.apache.spark.executor.Executor.org$apache$spark$ > executor$Executor$$updateDependencies(Executor.scala:480) > >> > > >> > at org.apache.spark.executor.Executor$TaskRunner.run( > Executor.scala:252) > >> > > >> > at > >> > > >> > java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1142) > >> > > >> > at > >> > > >> > java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:617) > >> > > >> > at java.lang.Thread.run(Thread.java:745) > >> > > >> > > >> > > >> > > >> > > >> > My Scala code > >> > > >> > > >> > package com.x.y.vr.spark.first > >> > > >> > /* SimpleApp.scala */ > >> > > >> > import org.apache.spark.SparkContext > >> > > >> > import org.apache.spark.SparkContext._ > >> > > >> > import org.apache.spark.SparkConf > >> > > >> > object SimpleApp { > >> > > >> > def main(args: Array[String]) { > >> > > >> > val logFile = "/Users/vttrich/Downloads/spark-2.0.0/README.md" // > >> > Should > >> > be some file on your system > >> > > >> > val conf = new SparkConf().setAppName("Simple Application") > >> > > >> > val sc = new SparkContext("local[*]", "RatingsCounter") > >> > > >> > //val sc = new SparkContext(conf) > >> > > >> > val logData = sc.textFile(logFile, 2).cache() > >> > > >> > val numAs = logData.filter(line => line.contains("a")).count() > >> > > >> > val numBs = logData.filter(line => line.contains("b")).count() > >> > > >> > println("Lines with a: %s, Lines with b: %s".format(numAs, numBs)) > >> > > >> > } > >> > > >> > } > > > > >