This do happens sometimes, but it is a warning because Spark is designed try successive ports until it succeeds. So unless a crazzzzy number of successive ports are blocked (runaway processes?? insufficient clearing of ports by OS??), these errors should not be a problem for tests passing.
On Wed, May 21, 2014 at 2:31 PM, Adrian Mocanu <amoc...@verticalscope.com>wrote: > Just found this at the top of the log: > > > > 17:14:41.124 [pool-7-thread-3-ScalaTest-running-StreamingSpikeSpec] WARN > o.e.j.u.component.AbstractLifeCycle - FAILED > SelectChannelConnector@0.0.0.0:4040: java.net.BindException: Address > already in use > > build 21-May-2014 17:14:41 java.net.BindException: Address already in > use > > > > Is there a way to set these connection up so that they don’t all start on > the same port (that’s my guess for the root cause of the issue) > > > > *From:* Adrian Mocanu [mailto:amoc...@verticalscope.com] > *Sent:* May-21-14 4:58 PM > *To:* u...@spark.incubator.apache.org; user@spark.apache.org > *Subject:* tests that run locally fail when run through bamboo > > > > I have a few test cases for Spark which extend TestSuiteBase from > org.apache.spark.streaming. > > The tests run fine on my machine but when I commit to repo and run the > tests automatically with bamboo the test cases fail with these errors. > > > > How to fix? > > > > > > 21-May-2014 16:33:09 > > [info] StreamingZigZagSpec: > > 21-May-2014 16:33:09 > > [info] - compute zigzag indicator in stream *** FAILED *** > > 21-May-2014 16:33:09 > > [info] org.apache.spark.SparkException: Job aborted: Task 1.0:1 failed 1 > times (most recent failure: Exception failure: > java.io.StreamCorruptedException: invalid type code: AC) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1020) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1018) > > 21-May-2014 16:33:09 > > [info] at > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > > 21-May-2014 16:33:09 > > [info] at > scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) > > 21-May-2014 16:33:09 > > [info] at org.apache.spark.scheduler.DAGScheduler.org > $apache$spark$scheduler$DAGScheduler$$abortStage(DAGScheduler.scala:1018) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) > > 21-May-2014 16:33:09 > > [info] at scala.Option.foreach(Option.scala:236) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:604) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$start$1$$anon$2$$anonfun$receive$1.applyOrElse(DAGScheduler.scala:190) > > 21-May-2014 16:33:09 > > [info] ... > > 21-May-2014 16:33:09 > > [info] - compute zigzag indicator in stream with intermittent empty RDDs > *** FAILED *** > > 21-May-2014 16:33:09 > > [info] Operation timed out after 10042 ms (TestSuiteBase.scala:283) > > 21-May-2014 16:33:09 > > [info] - compute zigzag indicator in stream with 3 empty RDDs *** FAILED > *** > > 21-May-2014 16:33:09 > > [info] org.apache.spark.SparkException: Job aborted: Task 1.0:1 failed 1 > times (most recent failure: Exception failure: > java.io.FileNotFoundException: > /tmp/spark-local-20140521163241-1707/0f/shuffle_1_1_1 (No such file or > directory)) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1020) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1018) > > 21-May-2014 16:33:09 > > [info] at > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > > 21-May-2014 16:33:09 > > [info] at > scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) > > 21-May-2014 16:33:09 > > [info] at org.apache.spark.scheduler.DAGScheduler.org > $apache$spark$scheduler$DAGScheduler$$abortStage(DAGScheduler.scala:1018) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) > > 21-May-2014 16:33:09 > > [info] at scala.Option.foreach(Option.scala:236) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:604) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$start$1$$anon$2$$anonfun$receive$1.applyOrElse(DAGScheduler.scala:190) > > 21-May-2014 16:33:09 > > [info] ... > > 21-May-2014 16:33:09 > > [info] - compute zigzag indicator in stream w notification for each > change *** FAILED *** > > 21-May-2014 16:33:09 > > [info] org.apache.spark.SparkException: Job aborted: Task 141.0:0 failed > 1 times (most recent failure: Exception failure: > java.io.FileNotFoundException: http://10.10.1.9:62793/broadcast_1) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1020) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1018) > > 21-May-2014 16:33:09 > > [info] at > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > > 21-May-2014 16:33:09 > > [info] at > scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) > > 21-May-2014 16:33:09 > > [info] at org.apache.spark.scheduler.DAGScheduler.org > $apache$spark$scheduler$DAGScheduler$$abortStage(DAGScheduler.scala:1018) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:604) > > 21-May-2014 16:33:09 > > [info] at scala.Option.foreach(Option.scala:236) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:604) > > 21-May-2014 16:33:09 > > [info] at > org.apache.spark.scheduler.DAGScheduler$$anonfun$start$1$$anon$2$$anonfun$receive$1.applyOrElse(DAGScheduler.scala:190) > > 21-May-2014 16:33:09 > > [info] ... > > 21-May-2014 16:33:09 > > [info] - compute zigzag in stream where there is 1 key/RDDs but multiple > key types exist > > 21-May-2014 16:33:09 > > [info] - compute zigzag in stream where RDDs have more than 1 key > > 21-May-2014 16:33:09 > > 16:33:09.819 [spark-akka.actor.default-dispatcher-15] INFO Remoting - > Remoting shut down > > 21-May-2014 16:33:09 > > 16:33:09.819 [spark-akka.actor.default-dispatcher-4] INFO > a.r.RemoteActorRefProvider$RemotingTerminator - Remoting shut down. > > 21-May-2014 16:33:09 > > [info] Run completed in 52 seconds, 792 milliseconds. > > 21-May-2014 16:33:09 > > [info] Total number of tests run: 36 > > 21-May-2014 16:33:09 > > [info] Suites: completed 8, aborted 0 > > 21-May-2014 16:33:09 > > [info] Tests: succeeded 25, failed 11, canceled 0, ignored 0, pending 0 > > > > > > > > -Adrian > > >