Github user Leemoonsoo commented on the pull request:
https://github.com/apache/incubator-zeppelin/pull/208#issuecomment-173423103
I've made some progress on testing CI,
at branch
https://github.com/Leemoonsoo/incubator-zeppelin/tree/rinterpreter_jan
at that passes one testing profile (spark-1.6) on CI, but fails on the
other profiles (spark-1.5, spark-1.4 ...)
https://travis-ci.org/Leemoonsoo/incubator-zeppelin/builds/103512174
Any people familiar with following exception?
```
16/01/20 02:52:10 INFO SparkContext: Running Spark version 1.5.2
16/01/20 02:52:10 INFO SecurityManager: Changing view acls to: travis
16/01/20 02:52:10 INFO SecurityManager: Changing modify acls to: travis
16/01/20 02:52:10 INFO SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions: Set(travis); users
with modify permissions: Set(travis)
16/01/20 02:52:11 INFO Slf4jLogger: Slf4jLogger started
16/01/20 02:52:11 INFO Remoting: Starting remoting
16/01/20 02:52:11 INFO Remoting: Remoting started; listening on addresses
:[akka.tcp://[email protected]:47557]
16/01/20 02:52:11 INFO Utils: Successfully started service 'sparkDriver' on
port 47557.
16/01/20 02:52:11 INFO SparkEnv: Registering MapOutputTracker
16/01/20 02:52:11 INFO SparkEnv: Registering BlockManagerMaster
16/01/20 02:52:11 INFO DiskBlockManager: Created local directory at
/tmp/blockmgr-d5f6db46-c354-457f-ac04-514cb750f5f7
16/01/20 02:52:11 INFO MemoryStore: MemoryStore started with capacity 530.3
MB
16/01/20 02:52:11 INFO HttpFileServer: HTTP File server directory is
/tmp/spark-70746e88-64b6-4e5f-a89a-0acc469f220c/httpd-ee337ade-6d23-4fcb-ac42-1e649e11c626
16/01/20 02:52:11 INFO HttpServer: Starting HTTP Server
16/01/20 02:52:11 INFO Utils: Successfully started service 'HTTP file
server' on port 42899.
16/01/20 02:52:11 INFO SparkEnv: Registering OutputCommitCoordinator
16/01/20 02:52:11 INFO Utils: Successfully started service 'SparkUI' on
port 4040.
16/01/20 02:52:11 INFO SparkUI: Started SparkUI at http://10.240.0.163:4040
16/01/20 02:52:11 INFO SparkContext: Added JAR
file:/home/travis/build/Leemoonsoo/incubator-zeppelin/interpreter/spark/zeppelin-spark-0.6.0-incubating-SNAPSHOT.jar
at http://10.240.0.163:42899/jars/zeppelin-spark-0.6.0-incubating-SNAPSHOT.jar
with timestamp 1453258331807
16/01/20 02:52:11 INFO FairSchedulableBuilder: Created default pool
default, schedulingMode: FIFO, minShare: 0, weight: 1
16/01/20 02:52:11 WARN MetricsSystem: Using default name DAGScheduler for
source because spark.app.id is not set.
16/01/20 02:52:11 INFO AppClient$ClientEndpoint: Connecting to master
spark://testing-gce-b9b06e11-646b-46d3-9860-831cc900b4f7.c.eco-emissary-99515.internal:7071...
16/01/20 02:52:31 ERROR SparkUncaughtExceptionHandler: Uncaught exception
in thread Thread[appclient-registration-retry-thread,5,main]
java.util.concurrent.RejectedExecutionException: Task
java.util.concurrent.FutureTask@12241de8 rejected from
java.util.concurrent.ThreadPoolExecutor@12773a6[Running, pool size = 1, active
threads = 0, queued tasks = 0, completed tasks = 1]
at
java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2048)
at
java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:821)
at
java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1372)
at
java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:110)
at
org.apache.spark.deploy.client.AppClient$ClientEndpoint$$anonfun$tryRegisterAllMasters$1.apply(AppClient.scala:96)
at
org.apache.spark.deploy.client.AppClient$ClientEndpoint$$anonfun$tryRegisterAllMasters$1.apply(AppClient.scala:95)
at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
at
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
at
scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:108)
at
org.apache.spark.deploy.client.AppClient$ClientEndpoint.tryRegisterAllMasters(AppClient.scala:95)
at
org.apache.spark.deploy.client.AppClient$ClientEndpoint.org$apache$spark$deploy$client$AppClient$ClientEndpoint$$registerWithMaster(AppClient.scala:121)
at
org.apache.spark.deploy.client.AppClient$ClientEndpoint$$anon$2$$anonfun$run$1.apply$mcV$sp(AppClient.scala:132)
at org.apache.spark.util.Utils$.tryOrExit(Utils.scala:1119)
at
org.apache.spark.deploy.client.AppClient$ClientEndpoint$$anon$2.run(AppClient.scala:124)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:304)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:178)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
```
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---