Hi Kevin, Cool! Please let me know if you need any assistance.
On Aug 8, 2016 1:42 PM, "Coakley, Kevin" <[email protected]> wrote: > Hi Wail, > > I figure out the problem, AsterixDB was configured for 127.0.0.1. The > notebook at https://github.com/Nullification/asterixdb-spark- > connector/blob/master/zeppelin-notebook/asterixdb-spark-example/note.json > ran successfully once I recreated the AsterixDB instance to use the > external IP. > > I have not ran any of my own queries but I did get both of the examples > https://github.com/Nullification/asterixdb-spark-connector to run > successfully. > > Thank you! > > -Kevin > > > > On 8/3/16, 10:23 AM, "Wail Alkowaileet" <[email protected]> wrote: > > One more thing: > Can you paste your cluster configuration as well? > > Thanks > > On Wed, Aug 3, 2016 at 12:32 PM, Wail Alkowaileet <[email protected]> > wrote: > > > Hi Kevin, > > > > Thanks for testing it! I really appreciate it. > > Definitely I tested it on my network (KACST) and I just tried to > reproduce > > the same problem you have at MIT network. It seems that I didn't get > the > > same problem. > > > > Can you paste the full logs? I just want to know if the connector > got the > > ResultLocations correctly ? > > > > > > Thanks again :-) > > > > On Tue, Aug 2, 2016 at 4:25 PM, Coakley, Kevin <[email protected]> > wrote: > > > >> Hi Wail, > >> > >> I was able to get the asterixdb-spark-connector to work as long as > >> asterixdb, zeppelin and spark are all running on the same server. > >> > >> When I try to access the asterixdb on a remote server, I receive the > >> org.apache.hyracks.api.exceptions.HyracksDataException: Connection > fail > >> error at the bottom of this email. > >> > >> I don’t believe there are any firewalls between the two systems so > I am > >> unsure why I am receiving a connection failure. I looked at the > hyracks > >> documentation at > >> https://github.com/apache/asterixdb/tree/master/hyracks- > fullstack/hyracks/hyracks-documentation/src/books/user-guide > >> it didn’t mention anything about how to access hyracks remotely. I > couldn’t > >> find any additional documentation by searching Google. > >> > >> > >> $ /opt/spark/bin/spark-shell --packages > >> org.apache.asterix:asterixdb-spark-connector_2.10:1.6.0 --conf > >> spark.asterix.connection.host=10.128.5.192 --conf > >> spark.asterix.connection.port=19002 --conf > spark.asterix.frame.size=131072 > >> > >> … > >> > >> scala> rddAql.collect().foreach(println) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: +++ Cleaning closure > <function1> > >> (org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12) +++ > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + declared fields: 2 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: public static final > long > >> org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12. > serialVersionUID > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: private final > >> org.apache.spark.rdd.RDD$$anonfun$collect$1 > >> org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$1 > >> 2.$outer > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + declared methods: 2 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: public final > >> java.lang.Object > >> org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12. > apply(java.lang.Object) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: public final > >> java.lang.Object > >> org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12. > apply(scala.collection.Ite > >> rator) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + inner classes: 0 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + outer classes: 2 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: > >> org.apache.spark.rdd.RDD$$anonfun$collect$1 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: > org.apache.spark.rdd.RDD > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + outer objects: 2 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: <function0> > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: AsterixRDD[0] at RDD at > >> AsterixRDD.scala:38 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + populating accessed > fields > >> because this is the starting closure > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + fields accessed by > starting > >> closure: 2 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: (class > >> org.apache.spark.rdd.RDD$$anonfun$collect$1,Set($outer)) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: (class > >> org.apache.spark.rdd.RDD,Set(org$apache$spark$rdd$RDD$$evidence$1)) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + outermost object is not a > >> closure, so do not clone it: (class org.apache.spark.rdd.RDD, > AsterixRDD[0] > >> at RDD at Ast > >> erixRDD.scala:38) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + cloning the object > <function0> > >> of class org.apache.spark.rdd.RDD$$anonfun$collect$1 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + cleaning cloned closure > >> <function0> recursively (org.apache.spark.rdd.RDD$$ > anonfun$collect$1) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: +++ Cleaning closure > <function0> > >> (org.apache.spark.rdd.RDD$$anonfun$collect$1) +++ > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + declared fields: 2 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: public static final > long > >> org.apache.spark.rdd.RDD$$anonfun$collect$1.serialVersionUID > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: private final > >> org.apache.spark.rdd.RDD org.apache.spark.rdd.RDD$$ > anonfun$collect$1.$outer > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + declared methods: 2 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: public > >> org.apache.spark.rdd.RDD org.apache.spark.rdd.RDD$$anonfun$collect$ > 1.org > >> $apache$spark$rdd$RDD$$anonfun$$$ > >> outer() > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: public final > >> java.lang.Object org.apache.spark.rdd.RDD$$ > anonfun$collect$1.apply() > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + inner classes: 1 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: > >> org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + outer classes: 1 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: > org.apache.spark.rdd.RDD > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + outer objects: 1 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: AsterixRDD[0] at RDD at > >> AsterixRDD.scala:38 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + fields accessed by > starting > >> closure: 2 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: (class > >> org.apache.spark.rdd.RDD$$anonfun$collect$1,Set($outer)) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: (class > >> org.apache.spark.rdd.RDD,Set(org$apache$spark$rdd$RDD$$evidence$1)) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + outermost object is not a > >> closure, so do not clone it: (class org.apache.spark.rdd.RDD, > AsterixRDD[0] > >> at RDD at Ast > >> erixRDD.scala:38) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: +++ closure <function0> > >> (org.apache.spark.rdd.RDD$$anonfun$collect$1) is now cleaned +++ > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: +++ closure <function1> > >> (org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12) is now > cleaned +++ > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: +++ Cleaning closure > <function2> > >> (org.apache.spark.SparkContext$$anonfun$runJob$5) +++ > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + declared fields: 2 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: public static final > long > >> org.apache.spark.SparkContext$$anonfun$runJob$5.serialVersionUID > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: private final > >> scala.Function1 > >> org.apache.spark.SparkContext$$anonfun$runJob$5.cleanedFunc$1 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + declared methods: 2 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: public final > >> java.lang.Object > >> org.apache.spark.SparkContext$$anonfun$runJob$5.apply(java. > lang.Object,java.lang.O > >> bject) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: public final > >> java.lang.Object > >> org.apache.spark.SparkContext$$anonfun$runJob$5.apply(org. > apache.spark.TaskContext > >> ,scala.collection.Iterator) > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + inner classes: 0 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + outer classes: 0 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + outer objects: 0 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + populating accessed > fields > >> because this is the starting closure > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + fields accessed by > starting > >> closure: 0 > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: + there are no enclosing > objects! > >> 16/08/02 20:18:49 DEBUG ClosureCleaner: +++ closure <function2> > >> (org.apache.spark.SparkContext$$anonfun$runJob$5) is now cleaned > +++ > >> 16/08/02 20:18:49 INFO SparkContext: Starting job: collect at > <console>:42 > >> 16/08/02 20:18:49 INFO DAGScheduler: Got job 0 (collect at > <console>:42) > >> with 1 output partitions > >> 16/08/02 20:18:49 INFO DAGScheduler: Final stage: ResultStage 0 > (collect > >> at <console>:42) > >> 16/08/02 20:18:49 INFO DAGScheduler: Parents of final stage: List() > >> 16/08/02 20:18:49 INFO DAGScheduler: Missing parents: List() > >> 16/08/02 20:18:49 DEBUG DAGScheduler: submitStage(ResultStage 0) > >> 16/08/02 20:18:49 DEBUG DAGScheduler: missing: List() > >> 16/08/02 20:18:49 INFO DAGScheduler: Submitting ResultStage 0 > >> (AsterixRDD[0] at RDD at AsterixRDD.scala:38), which has no missing > parents > >> 16/08/02 20:18:49 DEBUG DAGScheduler: > submitMissingTasks(ResultStage 0) > >> 16/08/02 20:18:49 INFO MemoryStore: Block broadcast_0 stored as > values in > >> memory (estimated size 1312.0 B, free 1312.0 B) > >> 16/08/02 20:18:49 DEBUG BlockManager: Put block broadcast_0 locally > took > >> 102 ms > >> 16/08/02 20:18:49 DEBUG BlockManager: Putting block broadcast_0 > without > >> replication took 103 ms > >> 16/08/02 20:18:49 INFO MemoryStore: Block broadcast_0_piece0 stored > as > >> bytes in memory (estimated size 912.0 B, free 2.2 KB) > >> 16/08/02 20:18:49 INFO BlockManagerInfo: Added broadcast_0_piece0 in > >> memory on localhost:42758 (size: 912.0 B, free: 517.4 MB) > >> 16/08/02 20:18:49 DEBUG BlockManagerMaster: Updated info of block > >> broadcast_0_piece0 > >> 16/08/02 20:18:49 DEBUG BlockManager: Told master about block > >> broadcast_0_piece0 > >> 16/08/02 20:18:49 DEBUG BlockManager: Put block broadcast_0_piece0 > >> locally took 5 ms > >> 16/08/02 20:18:49 DEBUG BlockManager: Putting block > broadcast_0_piece0 > >> without replication took 5 ms > >> 16/08/02 20:18:49 INFO SparkContext: Created broadcast 0 from > broadcast > >> at DAGScheduler.scala:1006 > >> 16/08/02 20:18:49 INFO DAGScheduler: Submitting 1 missing tasks from > >> ResultStage 0 (AsterixRDD[0] at RDD at AsterixRDD.scala:38) > >> 16/08/02 20:18:49 DEBUG DAGScheduler: New pending partitions: Set(0) > >> 16/08/02 20:18:49 INFO TaskSchedulerImpl: Adding task set 0.0 with > 1 tasks > >> 16/08/02 20:18:49 DEBUG TaskSetManager: Epoch for TaskSet 0.0: 0 > >> 16/08/02 20:18:49 DEBUG TaskSetManager: Valid locality levels for > TaskSet > >> 0.0: ANY > >> 16/08/02 20:18:49 DEBUG TaskSchedulerImpl: parentName: , name: > TaskSet_0, > >> runningTasks: 0 > >> 16/08/02 20:18:49 DEBUG TaskSetManager: Valid locality levels for > TaskSet > >> 0.0: ANY > >> 16/08/02 20:18:49 INFO TaskSetManager: Starting task 0.0 in stage > 0.0 > >> (TID 0, localhost, partition 0,ANY, 5872 bytes) > >> 16/08/02 20:18:49 INFO Executor: Running task 0.0 in stage 0.0 (TID > 0) > >> 16/08/02 20:18:49 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache. > httpcomponents_httpclient-4.5.jar > >> with timestamp 1470169042777 > >> 16/08/02 20:18:49 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:49 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache. > httpcomponents_httpclient-4.5.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb > >> > >> 128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp3860074775671300011.tmp > >> 16/08/02 20:18:49 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.httpco > >> mponents_httpclient-4.5.jar to class loader > >> 16/08/02 20:18:49 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.scala-lang_scala- > compiler-2.10.4.jar > >> with timestamp 1470169042864 > >> 16/08/02 20:18:49 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:49 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.scala-lang_scala- > compiler-2.10.4.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb1288 > >> > >> 10f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp3262286823800108479.tmp > >> 16/08/02 20:18:49 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.scala-lang_sc > >> ala-compiler-2.10.4.jar to class loader > >> 16/08/02 20:18:49 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.slf4j_slf4j-api-1.6.1.jar with > >> timestamp 1470169042806 > >> 16/08/02 20:18:49 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:49 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.slf4j_slf4j-api-1.6.1.jar to > >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/userFi > >> > >> les-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp4983533479624179412.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.slf4j_slf4j-a > >> pi-1.6.1.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/net.liftweb_lift-json_2.10-2.6.2.jar > with > >> timestamp 1470169042778 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/net.liftweb_lift-json_2.10-2.6.2.jar > to > >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2 > >> > >> /userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp7676771515275148134.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/net.liftweb_lift- > >> json_2.10-2.6.2.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/commons-logging_commons- > logging-1.2.jar > >> with timestamp 1470169042807 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/commons-logging_commons- > logging-1.2.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12881 > >> > >> 0f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp8603065512312611872.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/commons-logging_c > >> ommons-logging-1.2.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.scala-lang_scalap-2.10.4.jar > with > >> timestamp 1470169042810 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.scala-lang_scalap-2.10.4.jar to > >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/use > >> > >> rFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp1356133302927190413.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.scala-lang_sc > >> alap-2.10.4.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.asterix_ > asterixdb-spark-connector_2.10-1.6.0.jar > >> with timestamp 147016904 > >> 2761 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.asterix_ > asterixdb-spark-connector_2.10-1.6.0.jar > >> to /tmp/spark-67977d02-e7fd > >> > >> -4237-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d- > a326-fb0e0fb72af0/fetchFileTemp5674715681903620103.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.asteri > >> x_asterixdb-spark-connector_2.10-1.6.0.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-util-0.2.18-SNAPSHOT.jar > >> with timestamp 1470169042780 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-util-0.2.18-SNAPSHOT.jar > >> to /tmp/spark-67977d02-e7fd-4237-b9 > >> > >> 11-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp960508663840606601.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack > >> s_hyracks-util-0.2.18-SNAPSHOT.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > servlet-8.0.0.RC0.jar > >> with timestamp 1470169042804 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > servlet-8.0.0.RC0.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25f > >> > >> b128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp6629368718502114303.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty > >> _jetty-servlet-8.0.0.RC0.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket- > request-1.5.2.jar > >> with timestamp 1470169042806 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket- > request-1.5.2.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12 > >> > >> 8810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp1723007311578117804.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.wicket > >> _wicket-request-1.5.2.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.scala-lang_scala- > reflect-2.10.4.jar > >> with timestamp 1470169042881 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.scala-lang_scala- > reflect-2.10.4.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12881 > >> > >> 0f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp7325758342859460972.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.scala-lang_sc > >> ala-reflect-2.10.4.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/commons-io_commons-io-2.4.jar with > >> timestamp 1470169042783 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/commons-io_commons-io-2.4.jar to > >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/userFi > >> > >> les-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp5519211317591426960.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/commons-io_common > >> s-io-2.4.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-control-nc-0.2.18-SNAPSHOT.jar > >> with timestamp 14701690427 > >> 64 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-control-nc-0.2.18-SNAPSHOT.jar > >> to /tmp/spark-67977d02-e7fd-4 > >> > >> 237-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d- > a326-fb0e0fb72af0/fetchFileTemp3919109807459653847.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack > >> s_hyracks-control-nc-0.2.18-SNAPSHOT.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > io-8.0.0.RC0.jar > >> with timestamp 1470169042803 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > io-8.0.0.RC0.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb1288 > >> > >> 10f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp4524176199382352705.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty > >> _jetty-io-8.0.0.RC0.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-control-cc-0.2.18-SNAPSHOT.jar > >> with timestamp 14701690427 > >> 82 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-control-cc-0.2.18-SNAPSHOT.jar > >> to /tmp/spark-67977d02-e7fd-4 > >> > >> 237-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d- > a326-fb0e0fb72af0/fetchFileTemp4072935671878616979.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack > >> s_hyracks-control-cc-0.2.18-SNAPSHOT.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > security-8.0.0.RC0.jar > >> with timestamp 1470169042805 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > security-8.0.0.RC0.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25 > >> > >> fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp1641593757599706804.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty > >> _jetty-security-8.0.0.RC0.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > continuation-8.0.0.RC0.jar > >> with timestamp 1470169042802 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > continuation-8.0.0.RC0.jar > >> to /tmp/spark-67977d02-e7fd-4237-b91 > >> > >> 1-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp6705043369408149452.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty > >> _jetty-continuation-8.0.0.RC0.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-data-std-0.2.18-SNAPSHOT.jar > >> with timestamp 1470169042782 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-data-std-0.2.18-SNAPSHOT.jar > >> to /tmp/spark-67977d02-e7fd-423 > >> > >> 7-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp7285196531917051222.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack > >> s_hyracks-data-std-0.2.18-SNAPSHOT.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket- > util-1.5.2.jar > >> with timestamp 1470169042806 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket- > util-1.5.2.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12881 > >> > >> 0f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp2425095420520230824.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.wicket > >> _wicket-util-1.5.2.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.json_json-20090211.jar with > timestamp > >> 1470169042779 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.json_json-20090211.jar to > >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/userFiles > >> -601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp7246507731018592850.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.json_json-200 > >> 90211.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-net-0.2.18-SNAPSHOT.jar > >> with timestamp 1470169042780 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-net-0.2.18-SNAPSHOT.jar > >> to /tmp/spark-67977d02-e7fd-4237-b91 > >> > >> 1-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp4617288623442337252.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack > >> s_hyracks-net-0.2.18-SNAPSHOT.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/com.thoughtworks. > paranamer_paranamer-2.4.1.jar > >> with timestamp 1470169042810 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/com.thoughtworks. > paranamer_paranamer-2.4.1.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25 > >> > >> fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp4990079550464975335.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/com.thoughtworks. > >> paranamer_paranamer-2.4.1.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-ipc-0.2.18-SNAPSHOT.jar > >> with timestamp 1470169042779 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-ipc-0.2.18-SNAPSHOT.jar > >> to /tmp/spark-67977d02-e7fd-4237-b91 > >> > >> 1-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp7836301704668276963.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack > >> s_hyracks-ipc-0.2.18-SNAPSHOT.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-control-common-0.2.18-SNAPSHOT.jar > >> with timestamp 1470169 > >> 042783 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-control-common-0.2.18-SNAPSHOT.jar > >> to /tmp/spark-67977d02-e7 > >> > >> fd-4237-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d- > a326-fb0e0fb72af0/fetchFileTemp611546527112857923.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack > >> s_hyracks-control-common-0.2.18-SNAPSHOT.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > http-8.0.0.RC0.jar > >> with timestamp 1470169042803 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > http-8.0.0.RC0.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12 > >> > >> 8810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp1312243926013444902.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty > >> _jetty-http-8.0.0.RC0.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.ini4j_ini4j-0.5.4.jar with > timestamp > >> 1470169042801 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.ini4j_ini4j-0.5.4.jar to > >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/userFiles- > >> 601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/fetchFileTemp62954716978914462 > 78.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.ini4j_ini4j-0 > >> .5.4.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-comm-0.2.18-SNAPSHOT.jar > >> with timestamp 1470169042781 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-comm-0.2.18-SNAPSHOT.jar > >> to /tmp/spark-67977d02-e7fd-4237-b9 > >> > >> 11-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp7808003918156444019.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack > >> s_hyracks-comm-0.2.18-SNAPSHOT.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-api-0.2.18-SNAPSHOT.jar > >> with timestamp 1470169042763 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-api-0.2.18-SNAPSHOT.jar > >> to /tmp/spark-67977d02-e7fd-4237-b91 > >> > >> 1-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp1674943860545785458.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack > >> s_hyracks-api-0.2.18-SNAPSHOT.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-dataflow-common-0.2.18-SNAPSHOT.jar > >> with timestamp 147016 > >> 9042764 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-dataflow-common-0.2.18-SNAPSHOT.jar > >> to /tmp/spark-67977d02-e > >> > >> 7fd-4237-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d- > a326-fb0e0fb72af0/fetchFileTemp8322251742856356113.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack > >> s_hyracks-dataflow-common-0.2.18-SNAPSHOT.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.commons_ > commons-lang3-3.1.jar > >> with timestamp 1470169042780 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.commons_ > commons-lang3-3.1.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb1288 > >> > >> 10f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp1348552964800068025.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.common > >> s_commons-lang3-3.1.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-client-0.2.18-SNAPSHOT.jar > >> with timestamp 1470169042763 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.hyracks_ > hyracks-client-0.2.18-SNAPSHOT.jar > >> to /tmp/spark-67977d02-e7fd-4237- > >> > >> b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp2029035581834026042.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack > >> s_hyracks-client-0.2.18-SNAPSHOT.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache. > httpcomponents_httpcore-4.4.1.jar > >> with timestamp 1470169042807 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache. > httpcomponents_httpcore-4.4.1.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb > >> > >> 128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp1299168202947951583.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.httpco > >> mponents_httpcore-4.4.1.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > webapp-8.0.0.RC0.jar > >> with timestamp 1470169042784 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > webapp-8.0.0.RC0.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb > >> > >> 128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp6872334740679608070.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty > >> _jetty-webapp-8.0.0.RC0.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > util-8.0.0.RC0.jar > >> with timestamp 1470169042804 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > util-8.0.0.RC0.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12 > >> > >> 8810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp1130775847969234836.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty > >> _jetty-util-8.0.0.RC0.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/commons-codec_commons-codec-1.9.jar > with > >> timestamp 1470169042808 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/commons-codec_commons-codec-1.9.jar > to > >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > >> > >> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp6734358121338695352.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/commons-codec_com > >> mons-codec-1.9.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/com.googlecode.json- > simple_json-simple-1.1.jar > >> with timestamp 1470169042782 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/com.googlecode.json- > simple_json-simple-1.1.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25 > >> > >> fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp1101986196715574441.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/com.googlecode.js > >> on-simple_json-simple-1.1.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.mortbay.jetty_ > servlet-api-3.0.20100224.jar > >> with timestamp 1470169042802 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.mortbay.jetty_ > servlet-api-3.0.20100224.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25 > >> > >> fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp8117640740880704102.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.mortbay.jetty > >> _servlet-api-3.0.20100224.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > xml-8.0.0.RC0.jar > >> with timestamp 1470169042804 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > xml-8.0.0.RC0.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb128 > >> > >> 810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp6442266218584598106.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty > >> _jetty-xml-8.0.0.RC0.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket- > core-1.5.2.jar > >> with timestamp 1470169042800 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket- > core-1.5.2.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12881 > >> > >> 0f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp1041242714193765966.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.wicket > >> _wicket-core-1.5.2.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/args4j_args4j-2.0.12.jar with > timestamp > >> 1470169042779 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/args4j_args4j-2.0.12.jar to > >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/userFiles-6 > >> 01aa2fd-61b6-4f3d-a326-fb0e0fb72af0/fetchFileTemp56660284197390001 > 74.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/args4j_args4j-2.0 > >> .12.jar to class loader > >> 16/08/02 20:18:50 INFO Executor: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > server-8.0.0.RC0.jar > >> with timestamp 1470169042784 > >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security > >> 16/08/02 20:18:50 INFO Utils: Fetching > >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty- > server-8.0.0.RC0.jar > >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb > >> > >> 128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/ > fetchFileTemp1033020651873699094.tmp > >> 16/08/02 20:18:50 INFO Executor: Adding > >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/ > userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty > >> _jetty-server-8.0.0.RC0.jar to class loader > >> 16/08/02 20:18:50 DEBUG Executor: Task 0's epoch is 0 > >> 16/08/02 20:18:50 DEBUG BlockManager: Getting local block > broadcast_0 > >> 16/08/02 20:18:50 DEBUG BlockManager: Level for block broadcast_0 is > >> StorageLevel(true, true, false, true, 1) > >> 16/08/02 20:18:50 DEBUG BlockManager: Getting block broadcast_0 from > >> memory > >> java.net.ConnectException: Connection refused > >> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > >> at > >> sun.nio.ch.SocketChannelImpl.finishConnect( > SocketChannelImpl.java:717) > >> at > >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$ > IOThread.run(TCPEndpoint.java:190) > >> java.net.ConnectException: Connection refused > >> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > >> at > >> sun.nio.ch.SocketChannelImpl.finishConnect( > SocketChannelImpl.java:717) > >> at > >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$ > IOThread.run(TCPEndpoint.java:190) > >> java.net.ConnectException: Connection refused > >> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > >> at > >> sun.nio.ch.SocketChannelImpl.finishConnect( > SocketChannelImpl.java:717) > >> at > >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$ > IOThread.run(TCPEndpoint.java:190) > >> java.net.ConnectException: Connection refused > >> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > >> at > >> sun.nio.ch.SocketChannelImpl.finishConnect( > SocketChannelImpl.java:717) > >> at > >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$ > IOThread.run(TCPEndpoint.java:190) > >> java.net.ConnectException: Connection refused > >> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > >> at > >> sun.nio.ch.SocketChannelImpl.finishConnect( > SocketChannelImpl.java:717) > >> at > >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$ > IOThread.run(TCPEndpoint.java:190) > >> java.net.ConnectException: Connection refused > >> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > >> at > >> sun.nio.ch.SocketChannelImpl.finishConnect( > SocketChannelImpl.java:717) > >> at > >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$ > IOThread.run(TCPEndpoint.java:190) > >> java.net.ConnectException: Connection refused > >> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > >> at > >> sun.nio.ch.SocketChannelImpl.finishConnect( > SocketChannelImpl.java:717) > >> at > >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$ > IOThread.run(TCPEndpoint.java:190) > >> 16/08/02 20:18:50 ERROR Executor: Exception in task 0.0 in stage > 0.0 (TID > >> 0) > >> org.apache.hyracks.api.exceptions.HyracksDataException: Connection > failure > >> at > >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open( > DatasetNetworkInputChannel.java:105) > >> at > >> org.apache.asterix.connector.result.AsterixResultReader.< > init>(AsterixResultReader.scala:59) > >> at > >> org.apache.asterix.connector.rdd.AsterixRDD.compute( > AsterixRDD.scala:77) > >> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD. > scala:306) > >> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > >> at > >> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) > >> at org.apache.spark.scheduler.Task.run(Task.scala:89) > >> at > >> org.apache.spark.executor.Executor$TaskRunner.run( > Executor.scala:213) > >> at > >> java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1142) > >> at > >> java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:617) > >> at java.lang.Thread.run(Thread.java:745) > >> Caused by: org.apache.hyracks.net.exceptions.NetException: > Connection > >> failure > >> at > >> org.apache.hyracks.net.protocols.muxdemux.MultiplexedConnection. > waitUntilConnected(MultiplexedConnection.java:119) > >> at > >> org.apache.hyracks.net.protocols.muxdemux.MuxDemux. > connect(MuxDemux.java:141) > >> at > >> org.apache.hyracks.client.net.ClientNetworkManager.connect( > ClientNetworkManager.java:53) > >> at > >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open( > DatasetNetworkInputChannel.java:103) > >> ... 10 more > >> 16/08/02 20:18:50 DEBUG TaskSchedulerImpl: parentName: , name: > TaskSet_0, > >> runningTasks: 0 > >> 16/08/02 20:18:50 WARN TaskSetManager: Lost task 0.0 in stage 0.0 > (TID 0, > >> localhost): org.apache.hyracks.api.exceptions.HyracksDataException: > >> Connection fail > >> ure > >> at > >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open( > DatasetNetworkInputChannel.java:105) > >> at > >> org.apache.asterix.connector.result.AsterixResultReader.< > init>(AsterixResultReader.scala:59) > >> at > >> org.apache.asterix.connector.rdd.AsterixRDD.compute( > AsterixRDD.scala:77) > >> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD. > scala:306) > >> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > >> at > >> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) > >> at org.apache.spark.scheduler.Task.run(Task.scala:89) > >> at > >> org.apache.spark.executor.Executor$TaskRunner.run( > Executor.scala:213) > >> at > >> java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1142) > >> at > >> java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:617) > >> at java.lang.Thread.run(Thread.java:745) > >> Caused by: org.apache.hyracks.net.exceptions.NetException: > Connection > >> failure > >> at > >> org.apache.hyracks.net.protocols.muxdemux.MultiplexedConnection. > waitUntilConnected(MultiplexedConnection.java:119) > >> at > >> org.apache.hyracks.net.protocols.muxdemux.MuxDemux. > connect(MuxDemux.java:141) > >> at > >> org.apache.hyracks.client.net.ClientNetworkManager.connect( > ClientNetworkManager.java:53) > >> at > >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open( > DatasetNetworkInputChannel.java:103) > >> ... 10 more > >> > >> 16/08/02 20:18:50 ERROR TaskSetManager: Task 0 in stage 0.0 failed 1 > >> times; aborting job > >> 16/08/02 20:18:50 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose > >> tasks have all completed, from pool > >> 16/08/02 20:18:50 INFO TaskSchedulerImpl: Cancelling stage 0 > >> 16/08/02 20:18:50 INFO DAGScheduler: ResultStage 0 (collect at > >> <console>:42) failed in 1.099 s > >> 16/08/02 20:18:50 DEBUG DAGScheduler: After removal of stage 0, > remaining > >> stages = 0 > >> 16/08/02 20:18:50 INFO DAGScheduler: Job 0 failed: collect at > >> <console>:42, took 1.412302 s > >> org.apache.spark.SparkException: Job aborted due to stage failure: > Task 0 > >> in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in > stage > >> 0.0 (TID 0 > >> , localhost): org.apache.hyracks.api.exceptions. > HyracksDataException: > >> Connection failure > >> at > >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open( > DatasetNetworkInputChannel.java:105) > >> at > >> org.apache.asterix.connector.result.AsterixResultReader.< > init>(AsterixResultReader.scala:59) > >> at > >> org.apache.asterix.connector.rdd.AsterixRDD.compute( > AsterixRDD.scala:77) > >> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD. > scala:306) > >> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > >> at > >> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) > >> at org.apache.spark.scheduler.Task.run(Task.scala:89) > >> at > >> org.apache.spark.executor.Executor$TaskRunner.run( > Executor.scala:213) > >> at > >> java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1142) > >> at > >> java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:617) > >> at java.lang.Thread.run(Thread.java:745) > >> Caused by: org.apache.hyracks.net.exceptions.NetException: > Connection > >> failure > >> at > >> org.apache.hyracks.net.protocols.muxdemux.MultiplexedConnection. > waitUntilConnected(MultiplexedConnection.java:119) > >> at > >> org.apache.hyracks.net.protocols.muxdemux.MuxDemux. > connect(MuxDemux.java:141) > >> at > >> org.apache.hyracks.client.net.ClientNetworkManager.connect( > ClientNetworkManager.java:53) > >> at > >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open( > DatasetNetworkInputChannel.java:103) > >> ... 10 more > >> > >> Driver stacktrace: > >> at org.apache.spark.scheduler.DAGScheduler.org > >> $apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages( > DAGScheduler.scala:1431) > >> at > >> org.apache.spark.scheduler.DAGScheduler$$anonfun$ > abortStage$1.apply(DAGScheduler.scala:1419) > >> at > >> org.apache.spark.scheduler.DAGScheduler$$anonfun$ > abortStage$1.apply(DAGScheduler.scala:1418) > >> at > >> scala.collection.mutable.ResizableArray$class.foreach( > ResizableArray.scala:59) > >> at > >> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) > >> at > >> org.apache.spark.scheduler.DAGScheduler.abortStage( > DAGScheduler.scala:1418) > >> at > >> org.apache.spark.scheduler.DAGScheduler$$anonfun$ > handleTaskSetFailed$1.apply(DAGScheduler.scala:799) > >> at > >> org.apache.spark.scheduler.DAGScheduler$$anonfun$ > handleTaskSetFailed$1.apply(DAGScheduler.scala:799) > >> at scala.Option.foreach(Option.scala:236) > >> at > >> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed( > DAGScheduler.scala:799) > >> at > >> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop. > doOnReceive(DAGScheduler.scala:1640) > >> at > >> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop. > onReceive(DAGScheduler.scala:1599) > >> at > >> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop. > onReceive(DAGScheduler.scala:1588) > >> at org.apache.spark.util.EventLoop$$anon$1.run( > EventLoop.scala:48) > >> at > >> org.apache.spark.scheduler.DAGScheduler.runJob( > DAGScheduler.scala:620) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala: > 1832) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala: > 1845) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala: > 1858) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala: > 1929) > >> at > >> org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope( > RDDOperationScope.scala:150) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope( > RDDOperationScope.scala:111) > >> at org.apache.spark.rdd.RDD.withScope(RDD.scala:316) > >> at org.apache.spark.rdd.RDD.collect(RDD.scala:926) > >> at > >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC. > <init>(<console>:42) > >> at > >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.< > init>(<console>:47) > >> at > >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(< > console>:49) > >> at > >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:51) > >> at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$ > $iwC$$iwC.<init>(<console>:53) > >> at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:55) > >> at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:57) > >> at $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:59) > >> at $iwC$$iwC$$iwC$$iwC.<init>(<console>:61) > >> at $iwC$$iwC$$iwC.<init>(<console>:63) > >> at $iwC$$iwC.<init>(<console>:65) > >> at $iwC.<init>(<console>:67) > >> at <init>(<console>:69) > >> at .<init>(<console>:73) > >> at .<clinit>(<console>) > >> at .<init>(<console>:7) > >> at .<clinit>(<console>) > >> at $print(<console>) > >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) > >> at > >> sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:62) > >> at > >> sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > >> at java.lang.reflect.Method.invoke(Method.java:498) > >> at > >> org.apache.spark.repl.SparkIMain$ReadEvalPrint.call( > SparkIMain.scala:1065) > >> at > >> org.apache.spark.repl.SparkIMain$Request.loadAndRun( > SparkIMain.scala:1346) > >> at > >> org.apache.spark.repl.SparkIMain.loadAndRunReq$1( > SparkIMain.scala:840) > >> at > >> org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871) > >> at > >> org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819) > >> at > >> org.apache.spark.repl.SparkILoop.reallyInterpret$1( > SparkILoop.scala:857) > >> at > >> org.apache.spark.repl.SparkILoop.interpretStartingWith( > SparkILoop.scala:902) > >> at org.apache.spark.repl.SparkILoop.command(SparkILoop. > scala:814) > >> at > >> org.apache.spark.repl.SparkILoop.processLine$1( > SparkILoop.scala:657) > >> at > >> org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665) > >> at org.apache.spark.repl.SparkILoop.org > >> $apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670) > >> at > >> org.apache.spark.repl.SparkILoop$$anonfun$org$ > apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997) > >> at > >> org.apache.spark.repl.SparkILoop$$anonfun$org$ > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945) > >> at > >> org.apache.spark.repl.SparkILoop$$anonfun$org$ > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945) > >> at > >> scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader( > ScalaClassLoader.scala:135) > >> at org.apache.spark.repl.SparkILoop.org > >> $apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945) > >> at org.apache.spark.repl.SparkILoop.process(SparkILoop. > scala:1059) > >> at org.apache.spark.repl.Main$.main(Main.scala:31) > >> at org.apache.spark.repl.Main.main(Main.scala) > >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) > >> at > >> sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:62) > >> at > >> sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > >> at java.lang.reflect.Method.invoke(Method.java:498) > >> at > >> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) > >> at > >> org.apache.spark.deploy.SparkSubmit$.doRunMain$1( > SparkSubmit.scala:181) > >> at > >> org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) > >> at > >> org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) > >> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit. > scala) > >> Caused by: org.apache.hyracks.api.exceptions.HyracksDataException: > >> Connection failure > >> at > >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open( > DatasetNetworkInputChannel.java:105) > >> at > >> org.apache.asterix.connector.result.AsterixResultReader.< > init>(AsterixResultReader.scala:59) > >> at > >> org.apache.asterix.connector.rdd.AsterixRDD.compute( > AsterixRDD.scala:77) > >> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD. > scala:306) > >> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) > >> at > >> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) > >> at org.apache.spark.scheduler.Task.run(Task.scala:89) > >> at > >> org.apache.spark.executor.Executor$TaskRunner.run( > Executor.scala:213) > >> at > >> java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1142) > >> at > >> java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:617) > >> at java.lang.Thread.run(Thread.java:745) > >> Caused by: org.apache.hyracks.net.exceptions.NetException: > Connection > >> failure > >> at > >> org.apache.hyracks.net.protocols.muxdemux.MultiplexedConnection. > waitUntilConnected(MultiplexedConnection.java:119) > >> at > >> org.apache.hyracks.net.protocols.muxdemux.MuxDemux. > connect(MuxDemux.java:141) > >> at > >> org.apache.hyracks.client.net.ClientNetworkManager.connect( > ClientNetworkManager.java:53) > >> at > >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open( > DatasetNetworkInputChannel.java:103) > >> ... 10 more > >> > >> > >> > >> > >> On 7/29/16, 12:31 PM, "Wail Alkowaileet" <[email protected]> > wrote: > >> > >> Hi Illdar and Kevin, > >> > >> Sorry for the late reply. Yes the servlet that provides the > result > >> locations is still not in the codebase. > >> If you can apply the changes < > >> https://asterix-gerrit.ics.uci.edu/#/c/1003/> > >> into your AsterixDB and try it again I would be thankful. > >> > >> I'm still working on the Sonar comments. But it should still > work > >> fine. > >> > >> Thanks. > >> > >> On Fri, Jul 29, 2016 at 1:21 AM, Ildar Absalyamov < > >> [email protected]> wrote: > >> > >> > I also got the same error with similar using connector inside > >> zeppelin. > >> > > >> > > On Jul 26, 2016, at 16:40, Coakley, Kevin < > [email protected]> > >> wrote: > >> > > > >> > > Hi Wail, > >> > > > >> > > I am running the contents of > >> > > >> https://github.com/Nullification/asterixdb-spark- > connector/blob/master/zeppelin-notebook/asterixdb-spark-example/note.json > >> > using spark-shell and I get a 404 error when trying to access > >> > /query/result/location?handle=%7B%22handle%22%3A%5B13%2C0% > 5D%7D > >> HTTP/1.1. I > >> > don’t know if this is an error with what I am doing, an > AsterixDB > >> error or > >> > an error with asterixdb-spark-connector. Thank you for any > help > >> that you > >> > can provide. > >> > > > >> > > Below is the command that I used and the error (I removed > >> everything > >> > else): > >> > > > >> > > /opt/spark/bin/spark-shell --packages > >> > org.apache.asterix:asterixdb-spark-connector_2.10:1.6.0 > --conf > >> > spark.asterix.connection.host=10.128.5.170 --conf > >> > spark.asterix.connection.port=19002 --conf > >> spark.asterix.frame.size=131072 > >> > > > >> > > ….. > >> > > scala> val df = sqlContext.aql(aqlQuery,infer = true, > >> printCaseClasses = > >> > true) > >> > > 16/07/26 23:27:11 INFO SparkContextFunctions: > >> > spark.asterix.connection.host 10.128.5.170 > >> > > 16/07/26 23:27:11 INFO SparkContextFunctions: > >> > spark.asterix.connection.port 19002 > >> > > 16/07/26 23:27:11 INFO SparkContextFunctions: > >> spark.asterix.frame.size > >> > 131072 > >> > > 16/07/26 23:27:11 INFO SparkContextFunctions: > >> spark.asterix.frame.number > >> > 1 > >> > > 16/07/26 23:27:11 INFO SparkContextFunctions: > >> > spark.asterix.reader.number 2 > >> > > 16/07/26 23:27:11 INFO SparkContextFunctions: > >> > spark.asterix.prefetch.threshold 2 > >> > > 16/07/26 23:27:11 DEBUG RequestAddCookies: CookieSpec > selected: > >> > best-match > >> > > 16/07/26 23:27:11 DEBUG RequestAuthCache: Auth cache not > set in > >> the > >> > context > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > request: [route: {}->http://10.128.5.170:19002][total kept > alive: > >> 0; > >> > route allocated: > >> > > 0 of 2; total allocated: 0 of 20] > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > leased: [id: 0][route: {}->http://10.128.5.170:19002][total > kept > >> alive: > >> > 0; route allo > >> > > cated: 1 of 2; total allocated: 1 of 20] > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Opening connection > {}-> > >> > http://10.128.5.170:19002 > >> > > 16/07/26 23:27:11 DEBUG HttpClientConnectionManager: > Connecting > >> to / > >> > 10.128.5.170:19002 > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Executing request > POST > >> > /aql?mode=asynchronous&schema-inferencer=Spark HTTP/1.1 > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Target auth state: > >> UNCHALLENGED > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Proxy auth state: > >> UNCHALLENGED > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >> POST > >> > /aql?mode=asynchronous&schema-inferencer=Spark HTTP/1.1 > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >> > >> Content-Length: 386 > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >> > Content-Type: > >> > text/plain; charset=UTF-8 > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >> Host: > >> > 10.128.5.170:19002 > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >> > Connection: > >> > Keep-Alive > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >> > User-Agent: > >> > Apache-HttpClient/4.3.2 (java 1.5) > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >> > >> Accept-Encoding: > >> > gzip,deflate > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "POST > >> > /aql?mode=asynchronous&schema-inferencer=Spark > HTTP/1.1[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> > "Content-Length: > >> > 386[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> > "Content-Type: > >> > text/plain; charset=UTF-8[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "Host: > >> > 10.128.5.170:19002[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> > "Connection: > >> > Keep-Alive[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> > "User-Agent: > >> > Apache-HttpClient/4.3.2 (java 1.5)[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> > "Accept-Encoding: > >> > gzip,deflate[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "[\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> " > let > >> > $exampleSet := [[\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> " > >> {"name" : > >> > "Ann", "age" : 20, "salary" : 100000},[\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> " > >> {"name" : > >> > "Bob", "age" : 30, "salary" : 200000},[\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> " > >> {"name" : > >> > "Cat", "age" : 40, "salary" : 300000, "dependents" : [1, 2, > >> 3]},[\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> " > >> {"name" : > >> > "Cat", "age" : 50, "salary" : 400000}[\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> " > >> ][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> " > for > >> $x in > >> > $exampleSet[\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> " > >> return > >> > $x[\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> " " > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << "HTTP/1.1 > 200 > >> > OK[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << > "Content-Type: > >> > application/json;charset=utf-8[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << > >> "Transfer-Encoding: > >> > chunked[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << "Server: > >> > Jetty(8.0.0.RC0)[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << "[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << > "11[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << > >> > "{"handle":[13,0]}[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << "0[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << "[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 << > HTTP/1.1 200 > >> OK > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 << > Content-Type: > >> > application/json;charset=utf-8 > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 << > >> Transfer-Encoding: > >> > chunked > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 << Server: > >> > Jetty(8.0.0.RC0) > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Connection can be > kept > >> alive > >> > indefinitely > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > [id: 0][route: {}->http://10.128.5.170:19002] can be kept > alive > >> > indefinitely > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > released: [id: 0][route: {}->http://10.128.5.170:19002][total > kept > >> alive: > >> > 1; route al > >> > > located: 1 of 2; total allocated: 1 of 20] > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > manager is shutting down > >> > > 16/07/26 23:27:11 DEBUG DefaultManagedHttpClientConnection: > >> > http-outgoing-0: Close connection > >> > > 16/07/26 23:27:11 DEBUG DefaultManagedHttpClientConnection: > >> > http-outgoing-0: Close connection > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > manager shut down > >> > > 16/07/26 23:27:11 INFO AsterixHttpAPI: Response Handle: > >> {"handle":[13,0]} > >> > > 16/07/26 23:27:11 INFO AsterixHttpAPI: Handle(JID:13,RSID:0) > >> > > 16/07/26 23:27:11 DEBUG AsterixHttpAPI: Get status of: > >> > %7B%22handle%22%3A%5B13%2C0%5D%7D > >> > > 16/07/26 23:27:11 DEBUG RequestAddCookies: CookieSpec > selected: > >> > best-match > >> > > 16/07/26 23:27:11 DEBUG RequestAuthCache: Auth cache not > set in > >> the > >> > context > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > request: [route: {}->http://10.128.5.170:19002][total kept > alive: > >> 0; > >> > route allocated: > >> > > 0 of 2; total allocated: 0 of 20] > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > leased: [id: 1][route: {}->http://10.128.5.170:19002][total > kept > >> alive: > >> > 0; route allo > >> > > cated: 1 of 2; total allocated: 1 of 20] > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Opening connection > {}-> > >> > http://10.128.5.170:19002 > >> > > 16/07/26 23:27:11 DEBUG HttpClientConnectionManager: > Connecting > >> to / > >> > 10.128.5.170:19002 > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Executing request > GET > >> > /query/status?handle=%7B%22handle%22%3A%5B13%2C0%5D%7D > HTTP/1.1 > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Target auth state: > >> UNCHALLENGED > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Proxy auth state: > >> UNCHALLENGED > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 >> GET > >> > /query/status?handle=%7B%22handle%22%3A%5B13%2C0%5D%7D > HTTP/1.1 > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 >> Host: > >> > 10.128.5.170:19002 > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 >> > Connection: > >> > Keep-Alive > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 >> > User-Agent: > >> > Apache-HttpClient/4.3.2 (java 1.5) > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 >> > >> Accept-Encoding: > >> > gzip,deflate > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >> "GET > >> > /query/status?handle=%7B%22handle%22%3A%5B13%2C0%5D%7D > >> HTTP/1.1[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >> "Host: > >> > 10.128.5.170:19002[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >> > "Connection: > >> > Keep-Alive[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >> > "User-Agent: > >> > Apache-HttpClient/4.3.2 (java 1.5)[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >> > "Accept-Encoding: > >> > gzip,deflate[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >> "[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 << "HTTP/1.1 > 200 > >> > OK[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 << > "Content-Type: > >> > text/html;charset=UTF-8[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 << > "Content-Length: > >> > 20[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 << "Server: > >> > Jetty(8.0.0.RC0)[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 << "[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 << > >> "{"status":"SUCCESS"}" > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 << > HTTP/1.1 200 > >> OK > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 << > Content-Type: > >> > text/html;charset=UTF-8 > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 << > >> Content-Length: 20 > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 << Server: > >> > Jetty(8.0.0.RC0) > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Connection can be > kept > >> alive > >> > indefinitely > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > [id: 1][route: {}->http://10.128.5.170:19002] can be kept > alive > >> > indefinitely > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > released: [id: 1][route: {}->http://10.128.5.170:19002][total > kept > >> alive: > >> > 1; route al > >> > > located: 1 of 2; total allocated: 1 of 20] > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > manager is shutting down > >> > > 16/07/26 23:27:11 DEBUG DefaultManagedHttpClientConnection: > >> > http-outgoing-1: Close connection > >> > > 16/07/26 23:27:11 DEBUG DefaultManagedHttpClientConnection: > >> > http-outgoing-1: Close connection > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > manager shut down > >> > > 16/07/26 23:27:11 DEBUG AsterixHttpAPI: Get locations of: > >> > %7B%22handle%22%3A%5B13%2C0%5D%7D > >> > > 16/07/26 23:27:11 DEBUG RequestAddCookies: CookieSpec > selected: > >> > best-match > >> > > 16/07/26 23:27:11 DEBUG RequestAuthCache: Auth cache not > set in > >> the > >> > context > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > request: [route: {}->http://10.128.5.170:19002][total kept > alive: > >> 0; > >> > route allocated: > >> > > 0 of 2; total allocated: 0 of 20] > >> > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > leased: [id: 2][route: {}->http://10.128.5.170:19002][total > kept > >> alive: > >> > 0; route allo > >> > > cated: 1 of 2; total allocated: 1 of 20] > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Opening connection > {}-> > >> > http://10.128.5.170:19002 > >> > > 16/07/26 23:27:11 DEBUG HttpClientConnectionManager: > Connecting > >> to / > >> > 10.128.5.170:19002 > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Executing request > GET > >> > /query/result/location?handle=%7B%22handle%22%3A%5B13%2C0% > 5D%7D > >> HTTP/1.1 > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Target auth state: > >> UNCHALLENGED > >> > > 16/07/26 23:27:11 DEBUG MainClientExec: Proxy auth state: > >> UNCHALLENGED > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-2 >> GET > >> > /query/result/location?handle=%7B%22handle%22%3A%5B13%2C0% > 5D%7D > >> HTTP/1.1 > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-2 >> Host: > >> > 10.128.5.170:19002 > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-2 >> > Connection: > >> > Keep-Alive > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-2 >> > User-Agent: > >> > Apache-HttpClient/4.3.2 (java 1.5) > >> > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-2 >> > >> Accept-Encoding: > >> > gzip,deflate > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >> "GET > >> > /query/result/location?handle=%7B%22handle%22%3A%5B13%2C0% > 5D%7D > >> > HTTP/1.1[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >> "Host: > >> > 10.128.5.170:19002[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >> > "Connection: > >> > Keep-Alive[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >> > "User-Agent: > >> > Apache-HttpClient/4.3.2 (java 1.5)[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >> > "Accept-Encoding: > >> > gzip,deflate[\r][\n]" > >> > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >> "[\r][\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "HTTP/1.1 > 404 Not > >> > Found[\r][\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << > "Cache-Control: > >> > must-revalidate,no-cache,no-store[\r][\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << > "Content-Type: > >> > text/html;charset=ISO-8859-1[\r][\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << > "Content-Length: > >> > 1288[\r][\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "Server: > >> > Jetty(8.0.0.RC0)[\r][\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "[\r][\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << > "<html>[\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << > "<head>[\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "<meta > >> > http-equiv="Content-Type" > >> content="text/html;charset=ISO-8859-1"/>[\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << > "<title>Error > >> 404 Not > >> > Found</title>[\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << > "</head>[\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << > "<body>[\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "<h2>HTTP > ERROR: > >> > 404</h2>[\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "<p>Problem > >> accessing > >> > /query/result/location. Reason:[\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "<pre> > Not > >> > Found</pre></p>[\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "<hr > >> > /><i><small>Powered by Jetty://</small></i>[\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << " > >> > [\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << > "</body>[\n]" > >> > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << > "</html>[\n]" > >> > > 16/07/26 23:27:12 DEBUG headers: http-outgoing-2 << > HTTP/1.1 404 > >> Not > >> > Found > >> > > 16/07/26 23:27:12 DEBUG headers: http-outgoing-2 << > Cache-Control: > >> > must-revalidate,no-cache,no-store > >> > > 16/07/26 23:27:12 DEBUG headers: http-outgoing-2 << > Content-Type: > >> > text/html;charset=ISO-8859-1 > >> > > 16/07/26 23:27:12 DEBUG headers: http-outgoing-2 << > >> Content-Length: 1288 > >> > > 16/07/26 23:27:12 DEBUG headers: http-outgoing-2 << Server: > >> > Jetty(8.0.0.RC0) > >> > > 16/07/26 23:27:12 DEBUG MainClientExec: Connection can be > kept > >> alive > >> > indefinitely > >> > > 16/07/26 23:27:12 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > [id: 2][route: {}->http://10.128.5.170:19002] can be kept > alive > >> > indefinitely > >> > > 16/07/26 23:27:12 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > released: [id: 2][route: {}->http://10.128.5.170:19002][total > kept > >> alive: > >> > 1; route al > >> > > located: 1 of 2; total allocated: 1 of 20] > >> > > 16/07/26 23:27:12 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > manager is shutting down > >> > > 16/07/26 23:27:12 DEBUG DefaultManagedHttpClientConnection: > >> > http-outgoing-2: Close connection > >> > > 16/07/26 23:27:12 DEBUG DefaultManagedHttpClientConnection: > >> > http-outgoing-2: Close connection > >> > > 16/07/26 23:27:12 DEBUG PoolingHttpClientConnectionManager: > >> Connection > >> > manager shut down > >> > > 16/07/26 23:27:12 INFO AsterixHttpAPI: Result Locations: > <html> > >> > > <head> > >> > > <meta http-equiv="Content-Type" > >> content="text/html;charset=ISO-8859-1"/> > >> > > <title>Error 404 Not Found</title> > >> > > </head> > >> > > <body> > >> > > <h2>HTTP ERROR: 404</h2> > >> > > <p>Problem accessing /query/result/location. Reason: > >> > > <pre> Not Found</pre></p> > >> > > <hr /><i><small>Powered by Jetty://</small></i> > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > </body> > >> > > </html> > >> > > > >> > > net.liftweb.json.JsonParser$ParseException: unknown token < > >> > > Near: <h > >> > > at > >> net.liftweb.json.JsonParser$Parser.fail(JsonParser.scala:234) > >> > > at > >> > net.liftweb.json.JsonParser$Parser.nextToken(JsonParser. > scala:321) > >> > > at > >> > net.liftweb.json.JsonParser$$anonfun$2.apply(JsonParser. > scala:188) > >> > > at > >> > net.liftweb.json.JsonParser$$anonfun$2.apply(JsonParser. > scala:141) > >> > > at net.liftweb.json.JsonParser$. > parse(JsonParser.scala:80) > >> > > at net.liftweb.json.JsonParser$. > parse(JsonParser.scala:45) > >> > > at net.liftweb.json.package$.parse(package.scala:41) > >> > > at > >> net.liftweb.json.Serialization$.read(Serialization.scala:58) > >> > > at > >> > > >> org.apache.asterix.connector.AsterixHttpAPI.getResultLocations( > AsterixHttpAPI.scala:129) > >> > > at > >> > > >> org.apache.asterix.connector.SparkContextFunctions.executeQuery( > SparkContextFunctions.scala:103) > >> > > at > >> > > >> org.apache.asterix.connector.SparkContextFunctions.aql( > SparkContextFunctions.scala:80) > >> > > at > >> > > >> org.apache.spark.sql.asterix.SQLContextFunctions.executeQuery( > SQLContextFunctions.scala:103) > >> > > at > >> > > >> org.apache.spark.sql.asterix.SQLContextFunctions.aql( > SQLContextFunctions.scala:84) > >> > > at > >> > > >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC. > <init>(<console>:35) > >> > > at > >> > > >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.< > init>(<console>:40) > >> > > at > >> > > >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(< > console>:42) > >> > > at > >> > $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(< > console>:44) > >> > > at > >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:46) > >> > > at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$ > $iwC.<init>(<console>:48) > >> > > at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC. > <init>(<console>:50) > >> > > at $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52) > >> > > at $iwC$$iwC$$iwC$$iwC.<init>(<console>:54) > >> > > at $iwC$$iwC$$iwC.<init>(<console>:56) > >> > > at $iwC$$iwC.<init>(<console>:58) > >> > > at $iwC.<init>(<console>:60) > >> > > at <init>(<console>:62) > >> > > at .<init>(<console>:66) > >> > > at .<clinit>(<console>) > >> > > at .<init>(<console>:7) > >> > > at .<clinit>(<console>) > >> > > at $print(<console>) > >> > > at sun.reflect.NativeMethodAccessorImpl. > invoke0(Native > >> Method) > >> > > at > >> > > >> sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:62) > >> > > at > >> > > >> sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > >> > > at java.lang.reflect.Method.invoke(Method.java:498) > >> > > at > >> > > >> org.apache.spark.repl.SparkIMain$ReadEvalPrint.call( > SparkIMain.scala:1065) > >> > > at > >> > > >> org.apache.spark.repl.SparkIMain$Request.loadAndRun( > SparkIMain.scala:1346) > >> > > at > >> > > >> org.apache.spark.repl.SparkIMain.loadAndRunReq$1( > SparkIMain.scala:840) > >> > > at > >> > org.apache.spark.repl.SparkIMain.interpret( > SparkIMain.scala:871) > >> > > at > >> > org.apache.spark.repl.SparkIMain.interpret( > SparkIMain.scala:819) > >> > > at > >> > > >> org.apache.spark.repl.SparkILoop.reallyInterpret$1( > SparkILoop.scala:857) > >> > > at > >> > > >> org.apache.spark.repl.SparkILoop.interpretStartingWith( > SparkILoop.scala:902) > >> > > at > >> org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814) > >> > > at > >> > org.apache.spark.repl.SparkILoop.processLine$1( > SparkILoop.scala:657) > >> > > at > >> > org.apache.spark.repl.SparkILoop.innerLoop$1( > SparkILoop.scala:665) > >> > > at org.apache.spark.repl.SparkILoop.org > >> > $apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670) > >> > > at > >> > > >> org.apache.spark.repl.SparkILoop$$anonfun$org$ > apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997) > >> > > at > >> > > >> org.apache.spark.repl.SparkILoop$$anonfun$org$ > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945) > >> > > at > >> > > >> org.apache.spark.repl.SparkILoop$$anonfun$org$ > apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945) > >> > > at > >> > > >> scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader( > ScalaClassLoader.scala:135) > >> > > at org.apache.spark.repl.SparkILoop.org > >> > $apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945) > >> > > at > >> org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059) > >> > > at org.apache.spark.repl.Main$.main(Main.scala:31) > >> > > at org.apache.spark.repl.Main.main(Main.scala) > >> > > at sun.reflect.NativeMethodAccessorImpl. > invoke0(Native > >> Method) > >> > > at > >> > > >> sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:62) > >> > > at > >> > > >> sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > >> > > at java.lang.reflect.Method.invoke(Method.java:498) > >> > > at > >> > > >> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) > >> > > at > >> > > >> org.apache.spark.deploy.SparkSubmit$.doRunMain$1( > SparkSubmit.scala:181) > >> > > at > >> > org.apache.spark.deploy.SparkSubmit$.submit( > SparkSubmit.scala:206) > >> > > at > >> > org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit. > scala:121) > >> > > at > >> org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > From: Wail Alkowaileet <[email protected]> > >> > > Reply-To: "[email protected]" < > >> [email protected]> > >> > > Date: Monday, July 18, 2016 at 5:26 AM > >> > > To: "[email protected]" <[email protected]>, > " > >> > [email protected]" <[email protected]> > >> > > Subject: Re: Trio: AsterixDB, Spark and Zeppelin. > >> > > > >> > > Sorry. Here's the link for the connecot: > >> > > https://github.com/Nullification/asterixdb-spark-connector > >> > > > >> > > On Mon, Jul 18, 2016 at 2:34 PM, Wail Alkowaileet < > >> [email protected]> > >> > wrote: > >> > > Dears, > >> > > > >> > > Finally I finished cleaning and documenting the > AsterixDB-Spark > >> > connector and finalize Zeppelin interpreter for AQL and SQL++. > >> > > > >> > > AsterixDB-Spark Connector: > >> > > • Supports both AQL and SQL++ queries. > >> > > • Much cleaner code now. > >> > > • Please if you have ANY problem with it, create an issue > in the > >> project > >> > repo. > >> > > • I'm working on a tutorial-video from the build to use it > in > >> Zeppelin. > >> > > • I recommend you to use Zeppelin. (you can import the > connector > >> example > >> > notebook) > >> > > Source Code: > >> https://github.com/Nullification/astreixdb-spark-connector > >> > > > >> > > Apache Zeppelin with AsterixDB interpreter: > >> > > • Supports JSON-flattening (which will allow zeppelin to > visualize > >> > results). > >> > > • See attached screenshots. > >> > > • Will try to initiate pull request to merge it to Zeppelin > >> master. > >> > > Source Code: https://github.com/Nullification/zeppelin > >> > > > >> > > Finally, I just submitted Schema Inferencer. I have work on > some > >> Sonar > >> > comments and it should be ready soon. > >> > > > >> > > Thanks! > >> > > > >> > > -- > >> > > > >> > > Regards, > >> > > Wail Alkowaileet > >> > > > >> > > > >> > > > >> > > > >> > > -- > >> > > > >> > > Regards, > >> > > Wail Alkowaileet > >> > > > >> > > > >> > > >> > Best regards, > >> > Ildar > >> > > >> > > >> > >> > >> -- > >> > >> *Regards,* > >> Wail Alkowaileet > >> > >> > >> > > > > > > -- > > > > *Regards,* > > Wail Alkowaileet > > > > > > -- > > *Regards,* > Wail Alkowaileet > > >
