asterixdb-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Wail Alkowaileet <wael....@gmail.com>
Subject Re: Trio: AsterixDB, Spark and Zeppelin.
Date Wed, 10 Aug 2016 15:36:17 GMT
Hi Kevin,

Cool!
Please let me know if you need any assistance.

On Aug 8, 2016 1:42 PM, "Coakley, Kevin" <kcoakley@sdsc.edu> wrote:

> Hi Wail,
>
> I figure out the problem, AsterixDB was configured for 127.0.0.1. The
> notebook at https://github.com/Nullification/asterixdb-spark-
> connector/blob/master/zeppelin-notebook/asterixdb-spark-example/note.json
> ran successfully once I recreated the AsterixDB instance to use the
> external IP.
>
> I have not ran any of my own queries but I did get both of the examples
> https://github.com/Nullification/asterixdb-spark-connector to run
> successfully.
>
> Thank you!
>
> -Kevin
>
>
>
> On 8/3/16, 10:23 AM, "Wail Alkowaileet" <wael.y.k@gmail.com> wrote:
>
>     One more thing:
>     Can you paste your cluster configuration as well?
>
>     Thanks
>
>     On Wed, Aug 3, 2016 at 12:32 PM, Wail Alkowaileet <wael.y.k@gmail.com>
>     wrote:
>
>     > Hi Kevin,
>     >
>     > Thanks for testing it! I really appreciate it.
>     > Definitely I tested it on my network (KACST) and I just tried to
> reproduce
>     > the same problem you have at MIT network. It seems that I didn't get
> the
>     > same problem.
>     >
>     > Can you paste the full logs? I just want to know if the connector
> got the
>     > ResultLocations correctly ?
>     >
>     >
>     > Thanks again :-)
>     >
>     > On Tue, Aug 2, 2016 at 4:25 PM, Coakley, Kevin <kcoakley@sdsc.edu>
> wrote:
>     >
>     >> Hi Wail,
>     >>
>     >> I was able to get the asterixdb-spark-connector to work as long as
>     >> asterixdb, zeppelin and spark are all running on the same server.
>     >>
>     >> When I try to access the asterixdb on a remote server, I receive the
>     >> org.apache.hyracks.api.exceptions.HyracksDataException: Connection
> fail
>     >> error at the bottom of this email.
>     >>
>     >> I don’t believe there are any firewalls between the two systems so
> I am
>     >> unsure why I am receiving a connection failure. I looked at the
> hyracks
>     >> documentation at
>     >> https://github.com/apache/asterixdb/tree/master/hyracks-
> fullstack/hyracks/hyracks-documentation/src/books/user-guide
>     >> it didn’t mention anything about how to access hyracks remotely. I
> couldn’t
>     >> find any additional documentation by searching Google.
>     >>
>     >>
>     >> $ /opt/spark/bin/spark-shell --packages
>     >> org.apache.asterix:asterixdb-spark-connector_2.10:1.6.0 --conf
>     >> spark.asterix.connection.host=10.128.5.192 --conf
>     >> spark.asterix.connection.port=19002 --conf
> spark.asterix.frame.size=131072
>     >>
>     >> …
>     >>
>     >> scala>       rddAql.collect().foreach(println)
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner: +++ Cleaning closure
> <function1>
>     >> (org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12) +++
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + declared fields: 2
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      public static final
> long
>     >> org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12.
> serialVersionUID
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      private final
>     >> org.apache.spark.rdd.RDD$$anonfun$collect$1
>     >> org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$1
>     >> 2.$outer
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + declared methods: 2
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      public final
>     >> java.lang.Object
>     >> org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12.
> apply(java.lang.Object)
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      public final
>     >> java.lang.Object
>     >> org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12.
> apply(scala.collection.Ite
>     >> rator)
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + inner classes: 0
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + outer classes: 2
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:
>     >> org.apache.spark.rdd.RDD$$anonfun$collect$1
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:
> org.apache.spark.rdd.RDD
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + outer objects: 2
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      <function0>
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      AsterixRDD[0] at RDD at
>     >> AsterixRDD.scala:38
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + populating accessed
> fields
>     >> because this is the starting closure
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + fields accessed by
> starting
>     >> closure: 2
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      (class
>     >> org.apache.spark.rdd.RDD$$anonfun$collect$1,Set($outer))
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      (class
>     >> org.apache.spark.rdd.RDD,Set(org$apache$spark$rdd$RDD$$evidence$1))
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + outermost object is not a
>     >> closure, so do not clone it: (class org.apache.spark.rdd.RDD,
> AsterixRDD[0]
>     >> at RDD at Ast
>     >> erixRDD.scala:38)
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + cloning the object
> <function0>
>     >> of class org.apache.spark.rdd.RDD$$anonfun$collect$1
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + cleaning cloned closure
>     >> <function0> recursively (org.apache.spark.rdd.RDD$$
> anonfun$collect$1)
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner: +++ Cleaning closure
> <function0>
>     >> (org.apache.spark.rdd.RDD$$anonfun$collect$1) +++
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + declared fields: 2
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      public static final
> long
>     >> org.apache.spark.rdd.RDD$$anonfun$collect$1.serialVersionUID
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      private final
>     >> org.apache.spark.rdd.RDD org.apache.spark.rdd.RDD$$
> anonfun$collect$1.$outer
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + declared methods: 2
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      public
>     >> org.apache.spark.rdd.RDD org.apache.spark.rdd.RDD$$anonfun$collect$
> 1.org
>     >> $apache$spark$rdd$RDD$$anonfun$$$
>     >> outer()
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      public final
>     >> java.lang.Object org.apache.spark.rdd.RDD$$
> anonfun$collect$1.apply()
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + inner classes: 1
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:
>     >> org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + outer classes: 1
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:
> org.apache.spark.rdd.RDD
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + outer objects: 1
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      AsterixRDD[0] at RDD at
>     >> AsterixRDD.scala:38
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + fields accessed by
> starting
>     >> closure: 2
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      (class
>     >> org.apache.spark.rdd.RDD$$anonfun$collect$1,Set($outer))
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      (class
>     >> org.apache.spark.rdd.RDD,Set(org$apache$spark$rdd$RDD$$evidence$1))
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + outermost object is not a
>     >> closure, so do not clone it: (class org.apache.spark.rdd.RDD,
> AsterixRDD[0]
>     >> at RDD at Ast
>     >> erixRDD.scala:38)
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  +++ closure <function0>
>     >> (org.apache.spark.rdd.RDD$$anonfun$collect$1) is now cleaned +++
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  +++ closure <function1>
>     >> (org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12) is now
> cleaned +++
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner: +++ Cleaning closure
> <function2>
>     >> (org.apache.spark.SparkContext$$anonfun$runJob$5) +++
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + declared fields: 2
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      public static final
> long
>     >> org.apache.spark.SparkContext$$anonfun$runJob$5.serialVersionUID
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      private final
>     >> scala.Function1
>     >> org.apache.spark.SparkContext$$anonfun$runJob$5.cleanedFunc$1
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + declared methods: 2
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      public final
>     >> java.lang.Object
>     >> org.apache.spark.SparkContext$$anonfun$runJob$5.apply(java.
> lang.Object,java.lang.O
>     >> bject)
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:      public final
>     >> java.lang.Object
>     >> org.apache.spark.SparkContext$$anonfun$runJob$5.apply(org.
> apache.spark.TaskContext
>     >> ,scala.collection.Iterator)
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + inner classes: 0
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + outer classes: 0
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + outer objects: 0
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + populating accessed
> fields
>     >> because this is the starting closure
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + fields accessed by
> starting
>     >> closure: 0
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  + there are no enclosing
> objects!
>     >> 16/08/02 20:18:49 DEBUG ClosureCleaner:  +++ closure <function2>
>     >> (org.apache.spark.SparkContext$$anonfun$runJob$5) is now cleaned
> +++
>     >> 16/08/02 20:18:49 INFO SparkContext: Starting job: collect at
> <console>:42
>     >> 16/08/02 20:18:49 INFO DAGScheduler: Got job 0 (collect at
> <console>:42)
>     >> with 1 output partitions
>     >> 16/08/02 20:18:49 INFO DAGScheduler: Final stage: ResultStage 0
> (collect
>     >> at <console>:42)
>     >> 16/08/02 20:18:49 INFO DAGScheduler: Parents of final stage: List()
>     >> 16/08/02 20:18:49 INFO DAGScheduler: Missing parents: List()
>     >> 16/08/02 20:18:49 DEBUG DAGScheduler: submitStage(ResultStage 0)
>     >> 16/08/02 20:18:49 DEBUG DAGScheduler: missing: List()
>     >> 16/08/02 20:18:49 INFO DAGScheduler: Submitting ResultStage 0
>     >> (AsterixRDD[0] at RDD at AsterixRDD.scala:38), which has no missing
> parents
>     >> 16/08/02 20:18:49 DEBUG DAGScheduler:
> submitMissingTasks(ResultStage 0)
>     >> 16/08/02 20:18:49 INFO MemoryStore: Block broadcast_0 stored as
> values in
>     >> memory (estimated size 1312.0 B, free 1312.0 B)
>     >> 16/08/02 20:18:49 DEBUG BlockManager: Put block broadcast_0 locally
> took
>     >> 102 ms
>     >> 16/08/02 20:18:49 DEBUG BlockManager: Putting block broadcast_0
> without
>     >> replication took  103 ms
>     >> 16/08/02 20:18:49 INFO MemoryStore: Block broadcast_0_piece0 stored
> as
>     >> bytes in memory (estimated size 912.0 B, free 2.2 KB)
>     >> 16/08/02 20:18:49 INFO BlockManagerInfo: Added broadcast_0_piece0 in
>     >> memory on localhost:42758 (size: 912.0 B, free: 517.4 MB)
>     >> 16/08/02 20:18:49 DEBUG BlockManagerMaster: Updated info of block
>     >> broadcast_0_piece0
>     >> 16/08/02 20:18:49 DEBUG BlockManager: Told master about block
>     >> broadcast_0_piece0
>     >> 16/08/02 20:18:49 DEBUG BlockManager: Put block broadcast_0_piece0
>     >> locally took  5 ms
>     >> 16/08/02 20:18:49 DEBUG BlockManager: Putting block
> broadcast_0_piece0
>     >> without replication took  5 ms
>     >> 16/08/02 20:18:49 INFO SparkContext: Created broadcast 0 from
> broadcast
>     >> at DAGScheduler.scala:1006
>     >> 16/08/02 20:18:49 INFO DAGScheduler: Submitting 1 missing tasks from
>     >> ResultStage 0 (AsterixRDD[0] at RDD at AsterixRDD.scala:38)
>     >> 16/08/02 20:18:49 DEBUG DAGScheduler: New pending partitions: Set(0)
>     >> 16/08/02 20:18:49 INFO TaskSchedulerImpl: Adding task set 0.0 with
> 1 tasks
>     >> 16/08/02 20:18:49 DEBUG TaskSetManager: Epoch for TaskSet 0.0: 0
>     >> 16/08/02 20:18:49 DEBUG TaskSetManager: Valid locality levels for
> TaskSet
>     >> 0.0: ANY
>     >> 16/08/02 20:18:49 DEBUG TaskSchedulerImpl: parentName: , name:
> TaskSet_0,
>     >> runningTasks: 0
>     >> 16/08/02 20:18:49 DEBUG TaskSetManager: Valid locality levels for
> TaskSet
>     >> 0.0: ANY
>     >> 16/08/02 20:18:49 INFO TaskSetManager: Starting task 0.0 in stage
> 0.0
>     >> (TID 0, localhost, partition 0,ANY, 5872 bytes)
>     >> 16/08/02 20:18:49 INFO Executor: Running task 0.0 in stage 0.0 (TID
> 0)
>     >> 16/08/02 20:18:49 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.
> httpcomponents_httpclient-4.5.jar
>     >> with timestamp 1470169042777
>     >> 16/08/02 20:18:49 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:49 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.
> httpcomponents_httpclient-4.5.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb
>     >>
>     >> 128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp3860074775671300011.tmp
>     >> 16/08/02 20:18:49 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.httpco
>     >> mponents_httpclient-4.5.jar to class loader
>     >> 16/08/02 20:18:49 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.scala-lang_scala-
> compiler-2.10.4.jar
>     >> with timestamp 1470169042864
>     >> 16/08/02 20:18:49 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:49 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.scala-lang_scala-
> compiler-2.10.4.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb1288
>     >>
>     >> 10f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp3262286823800108479.tmp
>     >> 16/08/02 20:18:49 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.scala-lang_sc
>     >> ala-compiler-2.10.4.jar to class loader
>     >> 16/08/02 20:18:49 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.slf4j_slf4j-api-1.6.1.jar with
>     >> timestamp 1470169042806
>     >> 16/08/02 20:18:49 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:49 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.slf4j_slf4j-api-1.6.1.jar to
>     >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/userFi
>     >>
>     >> les-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp4983533479624179412.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.slf4j_slf4j-a
>     >> pi-1.6.1.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/net.liftweb_lift-json_2.10-2.6.2.jar
> with
>     >> timestamp 1470169042778
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/net.liftweb_lift-json_2.10-2.6.2.jar
> to
>     >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2
>     >>
>     >> /userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp7676771515275148134.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/net.liftweb_lift-
>     >> json_2.10-2.6.2.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/commons-logging_commons-
> logging-1.2.jar
>     >> with timestamp 1470169042807
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/commons-logging_commons-
> logging-1.2.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12881
>     >>
>     >> 0f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp8603065512312611872.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/commons-logging_c
>     >> ommons-logging-1.2.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.scala-lang_scalap-2.10.4.jar
> with
>     >> timestamp 1470169042810
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.scala-lang_scalap-2.10.4.jar to
>     >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/use
>     >>
>     >> rFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp1356133302927190413.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.scala-lang_sc
>     >> alap-2.10.4.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.asterix_
> asterixdb-spark-connector_2.10-1.6.0.jar
>     >> with timestamp 147016904
>     >> 2761
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.asterix_
> asterixdb-spark-connector_2.10-1.6.0.jar
>     >> to /tmp/spark-67977d02-e7fd
>     >>
>     >> -4237-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-
> a326-fb0e0fb72af0/fetchFileTemp5674715681903620103.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.asteri
>     >> x_asterixdb-spark-connector_2.10-1.6.0.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-util-0.2.18-SNAPSHOT.jar
>     >> with timestamp 1470169042780
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-util-0.2.18-SNAPSHOT.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b9
>     >>
>     >> 11-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp960508663840606601.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack
>     >> s_hyracks-util-0.2.18-SNAPSHOT.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> servlet-8.0.0.RC0.jar
>     >> with timestamp 1470169042804
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> servlet-8.0.0.RC0.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25f
>     >>
>     >> b128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp6629368718502114303.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty
>     >> _jetty-servlet-8.0.0.RC0.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket-
> request-1.5.2.jar
>     >> with timestamp 1470169042806
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket-
> request-1.5.2.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12
>     >>
>     >> 8810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp1723007311578117804.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.wicket
>     >> _wicket-request-1.5.2.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.scala-lang_scala-
> reflect-2.10.4.jar
>     >> with timestamp 1470169042881
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.scala-lang_scala-
> reflect-2.10.4.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12881
>     >>
>     >> 0f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp7325758342859460972.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.scala-lang_sc
>     >> ala-reflect-2.10.4.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/commons-io_commons-io-2.4.jar with
>     >> timestamp 1470169042783
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/commons-io_commons-io-2.4.jar to
>     >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/userFi
>     >>
>     >> les-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp5519211317591426960.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/commons-io_common
>     >> s-io-2.4.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-control-nc-0.2.18-SNAPSHOT.jar
>     >> with timestamp 14701690427
>     >> 64
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-control-nc-0.2.18-SNAPSHOT.jar
>     >> to /tmp/spark-67977d02-e7fd-4
>     >>
>     >> 237-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-
> a326-fb0e0fb72af0/fetchFileTemp3919109807459653847.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack
>     >> s_hyracks-control-nc-0.2.18-SNAPSHOT.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> io-8.0.0.RC0.jar
>     >> with timestamp 1470169042803
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> io-8.0.0.RC0.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb1288
>     >>
>     >> 10f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp4524176199382352705.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty
>     >> _jetty-io-8.0.0.RC0.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-control-cc-0.2.18-SNAPSHOT.jar
>     >> with timestamp 14701690427
>     >> 82
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-control-cc-0.2.18-SNAPSHOT.jar
>     >> to /tmp/spark-67977d02-e7fd-4
>     >>
>     >> 237-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-
> a326-fb0e0fb72af0/fetchFileTemp4072935671878616979.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack
>     >> s_hyracks-control-cc-0.2.18-SNAPSHOT.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> security-8.0.0.RC0.jar
>     >> with timestamp 1470169042805
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> security-8.0.0.RC0.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25
>     >>
>     >> fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp1641593757599706804.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty
>     >> _jetty-security-8.0.0.RC0.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> continuation-8.0.0.RC0.jar
>     >> with timestamp 1470169042802
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> continuation-8.0.0.RC0.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b91
>     >>
>     >> 1-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp6705043369408149452.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty
>     >> _jetty-continuation-8.0.0.RC0.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-data-std-0.2.18-SNAPSHOT.jar
>     >> with timestamp 1470169042782
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-data-std-0.2.18-SNAPSHOT.jar
>     >> to /tmp/spark-67977d02-e7fd-423
>     >>
>     >> 7-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp7285196531917051222.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack
>     >> s_hyracks-data-std-0.2.18-SNAPSHOT.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket-
> util-1.5.2.jar
>     >> with timestamp 1470169042806
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket-
> util-1.5.2.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12881
>     >>
>     >> 0f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp2425095420520230824.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.wicket
>     >> _wicket-util-1.5.2.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.json_json-20090211.jar with
> timestamp
>     >> 1470169042779
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.json_json-20090211.jar to
>     >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/userFiles
>     >> -601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp7246507731018592850.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.json_json-200
>     >> 90211.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-net-0.2.18-SNAPSHOT.jar
>     >> with timestamp 1470169042780
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-net-0.2.18-SNAPSHOT.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b91
>     >>
>     >> 1-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp4617288623442337252.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack
>     >> s_hyracks-net-0.2.18-SNAPSHOT.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/com.thoughtworks.
> paranamer_paranamer-2.4.1.jar
>     >> with timestamp 1470169042810
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/com.thoughtworks.
> paranamer_paranamer-2.4.1.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25
>     >>
>     >> fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp4990079550464975335.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/com.thoughtworks.
>     >> paranamer_paranamer-2.4.1.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-ipc-0.2.18-SNAPSHOT.jar
>     >> with timestamp 1470169042779
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-ipc-0.2.18-SNAPSHOT.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b91
>     >>
>     >> 1-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp7836301704668276963.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack
>     >> s_hyracks-ipc-0.2.18-SNAPSHOT.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-control-common-0.2.18-SNAPSHOT.jar
>     >> with timestamp 1470169
>     >> 042783
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-control-common-0.2.18-SNAPSHOT.jar
>     >> to /tmp/spark-67977d02-e7
>     >>
>     >> fd-4237-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-
> a326-fb0e0fb72af0/fetchFileTemp611546527112857923.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack
>     >> s_hyracks-control-common-0.2.18-SNAPSHOT.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> http-8.0.0.RC0.jar
>     >> with timestamp 1470169042803
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> http-8.0.0.RC0.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12
>     >>
>     >> 8810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp1312243926013444902.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty
>     >> _jetty-http-8.0.0.RC0.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.ini4j_ini4j-0.5.4.jar with
> timestamp
>     >> 1470169042801
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.ini4j_ini4j-0.5.4.jar to
>     >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/userFiles-
>     >> 601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/fetchFileTemp62954716978914462
> 78.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.ini4j_ini4j-0
>     >> .5.4.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-comm-0.2.18-SNAPSHOT.jar
>     >> with timestamp 1470169042781
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-comm-0.2.18-SNAPSHOT.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b9
>     >>
>     >> 11-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp7808003918156444019.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack
>     >> s_hyracks-comm-0.2.18-SNAPSHOT.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-api-0.2.18-SNAPSHOT.jar
>     >> with timestamp 1470169042763
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-api-0.2.18-SNAPSHOT.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b91
>     >>
>     >> 1-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp1674943860545785458.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack
>     >> s_hyracks-api-0.2.18-SNAPSHOT.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-dataflow-common-0.2.18-SNAPSHOT.jar
>     >> with timestamp 147016
>     >> 9042764
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-dataflow-common-0.2.18-SNAPSHOT.jar
>     >> to /tmp/spark-67977d02-e
>     >>
>     >> 7fd-4237-b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-
> a326-fb0e0fb72af0/fetchFileTemp8322251742856356113.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack
>     >> s_hyracks-dataflow-common-0.2.18-SNAPSHOT.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.commons_
> commons-lang3-3.1.jar
>     >> with timestamp 1470169042780
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.commons_
> commons-lang3-3.1.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb1288
>     >>
>     >> 10f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp1348552964800068025.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.common
>     >> s_commons-lang3-3.1.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-client-0.2.18-SNAPSHOT.jar
>     >> with timestamp 1470169042763
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.hyracks_
> hyracks-client-0.2.18-SNAPSHOT.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-
>     >>
>     >> b911-25fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp2029035581834026042.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.hyrack
>     >> s_hyracks-client-0.2.18-SNAPSHOT.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.
> httpcomponents_httpcore-4.4.1.jar
>     >> with timestamp 1470169042807
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.
> httpcomponents_httpcore-4.4.1.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb
>     >>
>     >> 128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp1299168202947951583.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.httpco
>     >> mponents_httpcore-4.4.1.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> webapp-8.0.0.RC0.jar
>     >> with timestamp 1470169042784
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> webapp-8.0.0.RC0.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb
>     >>
>     >> 128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp6872334740679608070.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty
>     >> _jetty-webapp-8.0.0.RC0.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> util-8.0.0.RC0.jar
>     >> with timestamp 1470169042804
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> util-8.0.0.RC0.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12
>     >>
>     >> 8810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp1130775847969234836.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty
>     >> _jetty-util-8.0.0.RC0.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/commons-codec_commons-codec-1.9.jar
> with
>     >> timestamp 1470169042808
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/commons-codec_commons-codec-1.9.jar
> to
>     >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
>     >>
>     >> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp6734358121338695352.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/commons-codec_com
>     >> mons-codec-1.9.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/com.googlecode.json-
> simple_json-simple-1.1.jar
>     >> with timestamp 1470169042782
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/com.googlecode.json-
> simple_json-simple-1.1.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25
>     >>
>     >> fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp1101986196715574441.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/com.googlecode.js
>     >> on-simple_json-simple-1.1.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.mortbay.jetty_
> servlet-api-3.0.20100224.jar
>     >> with timestamp 1470169042802
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.mortbay.jetty_
> servlet-api-3.0.20100224.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25
>     >>
>     >> fb128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp8117640740880704102.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.mortbay.jetty
>     >> _servlet-api-3.0.20100224.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> xml-8.0.0.RC0.jar
>     >> with timestamp 1470169042804
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> xml-8.0.0.RC0.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb128
>     >>
>     >> 810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp6442266218584598106.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty
>     >> _jetty-xml-8.0.0.RC0.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket-
> core-1.5.2.jar
>     >> with timestamp 1470169042800
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.apache.wicket_wicket-
> core-1.5.2.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb12881
>     >>
>     >> 0f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp1041242714193765966.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.apache.wicket
>     >> _wicket-core-1.5.2.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/args4j_args4j-2.0.12.jar with
> timestamp
>     >> 1470169042779
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/args4j_args4j-2.0.12.jar to
>     >> /tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/userFiles-6
>     >> 01aa2fd-61b6-4f3d-a326-fb0e0fb72af0/fetchFileTemp56660284197390001
> 74.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/args4j_args4j-2.0
>     >> .12.jar to class loader
>     >> 16/08/02 20:18:50 INFO Executor: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> server-8.0.0.RC0.jar
>     >> with timestamp 1470169042784
>     >> 16/08/02 20:18:50 DEBUG Utils: fetchFile not using security
>     >> 16/08/02 20:18:50 INFO Utils: Fetching
>     >> http://10.128.5.183:38874/jars/org.eclipse.jetty_jetty-
> server-8.0.0.RC0.jar
>     >> to /tmp/spark-67977d02-e7fd-4237-b911-25fb
>     >>
>     >> 128810f2/userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/
> fetchFileTemp1033020651873699094.tmp
>     >> 16/08/02 20:18:50 INFO Executor: Adding
>     >> file:/tmp/spark-67977d02-e7fd-4237-b911-25fb128810f2/
> userFiles-601aa2fd-61b6-4f3d-a326-fb0e0fb72af0/org.eclipse.jetty
>     >> _jetty-server-8.0.0.RC0.jar to class loader
>     >> 16/08/02 20:18:50 DEBUG Executor: Task 0's epoch is 0
>     >> 16/08/02 20:18:50 DEBUG BlockManager: Getting local block
> broadcast_0
>     >> 16/08/02 20:18:50 DEBUG BlockManager: Level for block broadcast_0 is
>     >> StorageLevel(true, true, false, true, 1)
>     >> 16/08/02 20:18:50 DEBUG BlockManager: Getting block broadcast_0 from
>     >> memory
>     >> java.net.ConnectException: Connection refused
>     >>         at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>     >>         at
>     >> sun.nio.ch.SocketChannelImpl.finishConnect(
> SocketChannelImpl.java:717)
>     >>         at
>     >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$
> IOThread.run(TCPEndpoint.java:190)
>     >> java.net.ConnectException: Connection refused
>     >>         at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>     >>         at
>     >> sun.nio.ch.SocketChannelImpl.finishConnect(
> SocketChannelImpl.java:717)
>     >>         at
>     >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$
> IOThread.run(TCPEndpoint.java:190)
>     >> java.net.ConnectException: Connection refused
>     >>         at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>     >>         at
>     >> sun.nio.ch.SocketChannelImpl.finishConnect(
> SocketChannelImpl.java:717)
>     >>         at
>     >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$
> IOThread.run(TCPEndpoint.java:190)
>     >> java.net.ConnectException: Connection refused
>     >>         at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>     >>         at
>     >> sun.nio.ch.SocketChannelImpl.finishConnect(
> SocketChannelImpl.java:717)
>     >>         at
>     >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$
> IOThread.run(TCPEndpoint.java:190)
>     >> java.net.ConnectException: Connection refused
>     >>         at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>     >>         at
>     >> sun.nio.ch.SocketChannelImpl.finishConnect(
> SocketChannelImpl.java:717)
>     >>         at
>     >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$
> IOThread.run(TCPEndpoint.java:190)
>     >> java.net.ConnectException: Connection refused
>     >>         at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>     >>         at
>     >> sun.nio.ch.SocketChannelImpl.finishConnect(
> SocketChannelImpl.java:717)
>     >>         at
>     >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$
> IOThread.run(TCPEndpoint.java:190)
>     >> java.net.ConnectException: Connection refused
>     >>         at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>     >>         at
>     >> sun.nio.ch.SocketChannelImpl.finishConnect(
> SocketChannelImpl.java:717)
>     >>         at
>     >> org.apache.hyracks.net.protocols.tcp.TCPEndpoint$
> IOThread.run(TCPEndpoint.java:190)
>     >> 16/08/02 20:18:50 ERROR Executor: Exception in task 0.0 in stage
> 0.0 (TID
>     >> 0)
>     >> org.apache.hyracks.api.exceptions.HyracksDataException: Connection
> failure
>     >>         at
>     >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open(
> DatasetNetworkInputChannel.java:105)
>     >>         at
>     >> org.apache.asterix.connector.result.AsterixResultReader.<
> init>(AsterixResultReader.scala:59)
>     >>         at
>     >> org.apache.asterix.connector.rdd.AsterixRDD.compute(
> AsterixRDD.scala:77)
>     >>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.
> scala:306)
>     >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>     >>         at
>     >> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>     >>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>     >>         at
>     >> org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:213)
>     >>         at
>     >> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
>     >>         at
>     >> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
>     >>         at java.lang.Thread.run(Thread.java:745)
>     >> Caused by: org.apache.hyracks.net.exceptions.NetException:
> Connection
>     >> failure
>     >>         at
>     >> org.apache.hyracks.net.protocols.muxdemux.MultiplexedConnection.
> waitUntilConnected(MultiplexedConnection.java:119)
>     >>         at
>     >> org.apache.hyracks.net.protocols.muxdemux.MuxDemux.
> connect(MuxDemux.java:141)
>     >>         at
>     >> org.apache.hyracks.client.net.ClientNetworkManager.connect(
> ClientNetworkManager.java:53)
>     >>         at
>     >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open(
> DatasetNetworkInputChannel.java:103)
>     >>         ... 10 more
>     >> 16/08/02 20:18:50 DEBUG TaskSchedulerImpl: parentName: , name:
> TaskSet_0,
>     >> runningTasks: 0
>     >> 16/08/02 20:18:50 WARN TaskSetManager: Lost task 0.0 in stage 0.0
> (TID 0,
>     >> localhost): org.apache.hyracks.api.exceptions.HyracksDataException:
>     >> Connection fail
>     >> ure
>     >>         at
>     >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open(
> DatasetNetworkInputChannel.java:105)
>     >>         at
>     >> org.apache.asterix.connector.result.AsterixResultReader.<
> init>(AsterixResultReader.scala:59)
>     >>         at
>     >> org.apache.asterix.connector.rdd.AsterixRDD.compute(
> AsterixRDD.scala:77)
>     >>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.
> scala:306)
>     >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>     >>         at
>     >> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>     >>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>     >>         at
>     >> org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:213)
>     >>         at
>     >> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
>     >>         at
>     >> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
>     >>         at java.lang.Thread.run(Thread.java:745)
>     >> Caused by: org.apache.hyracks.net.exceptions.NetException:
> Connection
>     >> failure
>     >>         at
>     >> org.apache.hyracks.net.protocols.muxdemux.MultiplexedConnection.
> waitUntilConnected(MultiplexedConnection.java:119)
>     >>         at
>     >> org.apache.hyracks.net.protocols.muxdemux.MuxDemux.
> connect(MuxDemux.java:141)
>     >>         at
>     >> org.apache.hyracks.client.net.ClientNetworkManager.connect(
> ClientNetworkManager.java:53)
>     >>         at
>     >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open(
> DatasetNetworkInputChannel.java:103)
>     >>         ... 10 more
>     >>
>     >> 16/08/02 20:18:50 ERROR TaskSetManager: Task 0 in stage 0.0 failed 1
>     >> times; aborting job
>     >> 16/08/02 20:18:50 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose
>     >> tasks have all completed, from pool
>     >> 16/08/02 20:18:50 INFO TaskSchedulerImpl: Cancelling stage 0
>     >> 16/08/02 20:18:50 INFO DAGScheduler: ResultStage 0 (collect at
>     >> <console>:42) failed in 1.099 s
>     >> 16/08/02 20:18:50 DEBUG DAGScheduler: After removal of stage 0,
> remaining
>     >> stages = 0
>     >> 16/08/02 20:18:50 INFO DAGScheduler: Job 0 failed: collect at
>     >> <console>:42, took 1.412302 s
>     >> org.apache.spark.SparkException: Job aborted due to stage failure:
> Task 0
>     >> in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in
> stage
>     >> 0.0 (TID 0
>     >> , localhost): org.apache.hyracks.api.exceptions.
> HyracksDataException:
>     >> Connection failure
>     >>         at
>     >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open(
> DatasetNetworkInputChannel.java:105)
>     >>         at
>     >> org.apache.asterix.connector.result.AsterixResultReader.<
> init>(AsterixResultReader.scala:59)
>     >>         at
>     >> org.apache.asterix.connector.rdd.AsterixRDD.compute(
> AsterixRDD.scala:77)
>     >>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.
> scala:306)
>     >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>     >>         at
>     >> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>     >>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>     >>         at
>     >> org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:213)
>     >>         at
>     >> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
>     >>         at
>     >> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
>     >>         at java.lang.Thread.run(Thread.java:745)
>     >> Caused by: org.apache.hyracks.net.exceptions.NetException:
> Connection
>     >> failure
>     >>         at
>     >> org.apache.hyracks.net.protocols.muxdemux.MultiplexedConnection.
> waitUntilConnected(MultiplexedConnection.java:119)
>     >>         at
>     >> org.apache.hyracks.net.protocols.muxdemux.MuxDemux.
> connect(MuxDemux.java:141)
>     >>         at
>     >> org.apache.hyracks.client.net.ClientNetworkManager.connect(
> ClientNetworkManager.java:53)
>     >>         at
>     >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open(
> DatasetNetworkInputChannel.java:103)
>     >>         ... 10 more
>     >>
>     >> Driver stacktrace:
>     >>         at org.apache.spark.scheduler.DAGScheduler.org
>     >> $apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(
> DAGScheduler.scala:1431)
>     >>         at
>     >> org.apache.spark.scheduler.DAGScheduler$$anonfun$
> abortStage$1.apply(DAGScheduler.scala:1419)
>     >>         at
>     >> org.apache.spark.scheduler.DAGScheduler$$anonfun$
> abortStage$1.apply(DAGScheduler.scala:1418)
>     >>         at
>     >> scala.collection.mutable.ResizableArray$class.foreach(
> ResizableArray.scala:59)
>     >>         at
>     >> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
>     >>         at
>     >> org.apache.spark.scheduler.DAGScheduler.abortStage(
> DAGScheduler.scala:1418)
>     >>         at
>     >> org.apache.spark.scheduler.DAGScheduler$$anonfun$
> handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
>     >>         at
>     >> org.apache.spark.scheduler.DAGScheduler$$anonfun$
> handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
>     >>         at scala.Option.foreach(Option.scala:236)
>     >>         at
>     >> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(
> DAGScheduler.scala:799)
>     >>         at
>     >> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> doOnReceive(DAGScheduler.scala:1640)
>     >>         at
>     >> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> onReceive(DAGScheduler.scala:1599)
>     >>         at
>     >> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.
> onReceive(DAGScheduler.scala:1588)
>     >>         at org.apache.spark.util.EventLoop$$anon$1.run(
> EventLoop.scala:48)
>     >>         at
>     >> org.apache.spark.scheduler.DAGScheduler.runJob(
> DAGScheduler.scala:620)
>     >>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1832)
>     >>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1845)
>     >>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1858)
>     >>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:
> 1929)
>     >>         at
>     >> org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)
>     >>         at
>     >> org.apache.spark.rdd.RDDOperationScope$.withScope(
> RDDOperationScope.scala:150)
>     >>         at
>     >> org.apache.spark.rdd.RDDOperationScope$.withScope(
> RDDOperationScope.scala:111)
>     >>         at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
>     >>         at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
>     >>         at
>     >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.
> <init>(<console>:42)
>     >>         at
>     >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<
> init>(<console>:47)
>     >>         at
>     >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<
> console>:49)
>     >>         at
>     >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:51)
>     >>         at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$
> $iwC$$iwC.<init>(<console>:53)
>     >>         at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:55)
>     >>         at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:57)
>     >>         at $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:59)
>     >>         at $iwC$$iwC$$iwC$$iwC.<init>(<console>:61)
>     >>         at $iwC$$iwC$$iwC.<init>(<console>:63)
>     >>         at $iwC$$iwC.<init>(<console>:65)
>     >>         at $iwC.<init>(<console>:67)
>     >>         at <init>(<console>:69)
>     >>         at .<init>(<console>:73)
>     >>         at .<clinit>(<console>)
>     >>         at .<init>(<console>:7)
>     >>         at .<clinit>(<console>)
>     >>         at $print(<console>)
>     >>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
>     >>         at
>     >> sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:62)
>     >>         at
>     >> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>     >>         at java.lang.reflect.Method.invoke(Method.java:498)
>     >>         at
>     >> org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(
> SparkIMain.scala:1065)
>     >>         at
>     >> org.apache.spark.repl.SparkIMain$Request.loadAndRun(
> SparkIMain.scala:1346)
>     >>         at
>     >> org.apache.spark.repl.SparkIMain.loadAndRunReq$1(
> SparkIMain.scala:840)
>     >>         at
>     >> org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871)
>     >>         at
>     >> org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819)
>     >>         at
>     >> org.apache.spark.repl.SparkILoop.reallyInterpret$1(
> SparkILoop.scala:857)
>     >>         at
>     >> org.apache.spark.repl.SparkILoop.interpretStartingWith(
> SparkILoop.scala:902)
>     >>         at org.apache.spark.repl.SparkILoop.command(SparkILoop.
> scala:814)
>     >>         at
>     >> org.apache.spark.repl.SparkILoop.processLine$1(
> SparkILoop.scala:657)
>     >>         at
>     >> org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665)
>     >>         at org.apache.spark.repl.SparkILoop.org
>     >> $apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670)
>     >>         at
>     >> org.apache.spark.repl.SparkILoop$$anonfun$org$
> apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997)
>     >>         at
>     >> org.apache.spark.repl.SparkILoop$$anonfun$org$
> apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
>     >>         at
>     >> org.apache.spark.repl.SparkILoop$$anonfun$org$
> apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
>     >>         at
>     >> scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(
> ScalaClassLoader.scala:135)
>     >>         at org.apache.spark.repl.SparkILoop.org
>     >> $apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945)
>     >>         at org.apache.spark.repl.SparkILoop.process(SparkILoop.
> scala:1059)
>     >>         at org.apache.spark.repl.Main$.main(Main.scala:31)
>     >>         at org.apache.spark.repl.Main.main(Main.scala)
>     >>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
>     >>         at
>     >> sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:62)
>     >>         at
>     >> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>     >>         at java.lang.reflect.Method.invoke(Method.java:498)
>     >>         at
>     >> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
>     >>         at
>     >> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> SparkSubmit.scala:181)
>     >>         at
>     >> org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
>     >>         at
>     >> org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
>     >>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.
> scala)
>     >> Caused by: org.apache.hyracks.api.exceptions.HyracksDataException:
>     >> Connection failure
>     >>         at
>     >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open(
> DatasetNetworkInputChannel.java:105)
>     >>         at
>     >> org.apache.asterix.connector.result.AsterixResultReader.<
> init>(AsterixResultReader.scala:59)
>     >>         at
>     >> org.apache.asterix.connector.rdd.AsterixRDD.compute(
> AsterixRDD.scala:77)
>     >>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.
> scala:306)
>     >>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>     >>         at
>     >> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>     >>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>     >>         at
>     >> org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:213)
>     >>         at
>     >> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
>     >>         at
>     >> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
>     >>         at java.lang.Thread.run(Thread.java:745)
>     >> Caused by: org.apache.hyracks.net.exceptions.NetException:
> Connection
>     >> failure
>     >>         at
>     >> org.apache.hyracks.net.protocols.muxdemux.MultiplexedConnection.
> waitUntilConnected(MultiplexedConnection.java:119)
>     >>         at
>     >> org.apache.hyracks.net.protocols.muxdemux.MuxDemux.
> connect(MuxDemux.java:141)
>     >>         at
>     >> org.apache.hyracks.client.net.ClientNetworkManager.connect(
> ClientNetworkManager.java:53)
>     >>         at
>     >> org.apache.hyracks.comm.channels.DatasetNetworkInputChannel.open(
> DatasetNetworkInputChannel.java:103)
>     >>         ... 10 more
>     >>
>     >>
>     >>
>     >>
>     >> On 7/29/16, 12:31 PM, "Wail Alkowaileet" <wael.y.k@gmail.com>
> wrote:
>     >>
>     >>     Hi Illdar and Kevin,
>     >>
>     >>     Sorry for the late reply. Yes the servlet that provides the
> result
>     >>     locations is still not in the codebase.
>     >>     If you can apply the changes <
>     >> https://asterix-gerrit.ics.uci.edu/#/c/1003/>
>     >>     into your AsterixDB and try it again I would be thankful.
>     >>
>     >>     I'm still working on the Sonar comments. But it should still
> work
>     >> fine.
>     >>
>     >>     Thanks.
>     >>
>     >>     On Fri, Jul 29, 2016 at 1:21 AM, Ildar Absalyamov <
>     >>     ildar.absalyamov@gmail.com> wrote:
>     >>
>     >>     > I also got the same error with similar using connector inside
>     >> zeppelin.
>     >>     >
>     >>     > > On Jul 26, 2016, at 16:40, Coakley, Kevin <
> kcoakley@sdsc.edu>
>     >> wrote:
>     >>     > >
>     >>     > > Hi Wail,
>     >>     > >
>     >>     > > I am running the contents of
>     >>     >
>     >> https://github.com/Nullification/asterixdb-spark-
> connector/blob/master/zeppelin-notebook/asterixdb-spark-example/note.json
>     >>     > using spark-shell and I get a 404 error when trying to access
>     >>     > /query/result/location?handle=%7B%22handle%22%3A%5B13%2C0%
> 5D%7D
>     >> HTTP/1.1. I
>     >>     > don’t know if this is an error with what I am doing, an
> AsterixDB
>     >> error or
>     >>     > an error with asterixdb-spark-connector. Thank you for any
> help
>     >> that you
>     >>     > can provide.
>     >>     > >
>     >>     > > Below is the command that I used and the error (I removed
>     >> everything
>     >>     > else):
>     >>     > >
>     >>     > > /opt/spark/bin/spark-shell --packages
>     >>     > org.apache.asterix:asterixdb-spark-connector_2.10:1.6.0
> --conf
>     >>     > spark.asterix.connection.host=10.128.5.170 --conf
>     >>     > spark.asterix.connection.port=19002 --conf
>     >> spark.asterix.frame.size=131072
>     >>     > >
>     >>     > > …..
>     >>     > > scala> val df = sqlContext.aql(aqlQuery,infer = true,
>     >> printCaseClasses =
>     >>     > true)
>     >>     > > 16/07/26 23:27:11 INFO SparkContextFunctions:
>     >>     > spark.asterix.connection.host 10.128.5.170
>     >>     > > 16/07/26 23:27:11 INFO SparkContextFunctions:
>     >>     > spark.asterix.connection.port 19002
>     >>     > > 16/07/26 23:27:11 INFO SparkContextFunctions:
>     >> spark.asterix.frame.size
>     >>     > 131072
>     >>     > > 16/07/26 23:27:11 INFO SparkContextFunctions:
>     >> spark.asterix.frame.number
>     >>     > 1
>     >>     > > 16/07/26 23:27:11 INFO SparkContextFunctions:
>     >>     > spark.asterix.reader.number 2
>     >>     > > 16/07/26 23:27:11 INFO SparkContextFunctions:
>     >>     > spark.asterix.prefetch.threshold 2
>     >>     > > 16/07/26 23:27:11 DEBUG RequestAddCookies: CookieSpec
> selected:
>     >>     > best-match
>     >>     > > 16/07/26 23:27:11 DEBUG RequestAuthCache: Auth cache not
> set in
>     >> the
>     >>     > context
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > request: [route: {}->http://10.128.5.170:19002][total kept
> alive:
>     >> 0;
>     >>     > route allocated:
>     >>     > > 0 of 2; total allocated: 0 of 20]
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > leased: [id: 0][route: {}->http://10.128.5.170:19002][total
> kept
>     >> alive:
>     >>     > 0; route allo
>     >>     > > cated: 1 of 2; total allocated: 1 of 20]
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Opening connection
> {}->
>     >>     > http://10.128.5.170:19002
>     >>     > > 16/07/26 23:27:11 DEBUG HttpClientConnectionManager:
> Connecting
>     >> to /
>     >>     > 10.128.5.170:19002
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Executing request
> POST
>     >>     > /aql?mode=asynchronous&schema-inferencer=Spark HTTP/1.1
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Target auth state:
>     >> UNCHALLENGED
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Proxy auth state:
>     >> UNCHALLENGED
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >> POST
>     >>     > /aql?mode=asynchronous&schema-inferencer=Spark HTTP/1.1
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >>
>     >> Content-Length: 386
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >>
> Content-Type:
>     >>     > text/plain; charset=UTF-8
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >> Host:
>     >>     > 10.128.5.170:19002
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >>
> Connection:
>     >>     > Keep-Alive
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >>
> User-Agent:
>     >>     > Apache-HttpClient/4.3.2 (java 1.5)
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 >>
>     >> Accept-Encoding:
>     >>     > gzip,deflate
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "POST
>     >>     > /aql?mode=asynchronous&schema-inferencer=Spark
> HTTP/1.1[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >>
> "Content-Length:
>     >>     > 386[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >>
> "Content-Type:
>     >>     > text/plain; charset=UTF-8[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "Host:
>     >>     > 10.128.5.170:19002[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >>
> "Connection:
>     >>     > Keep-Alive[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >>
> "User-Agent:
>     >>     > Apache-HttpClient/4.3.2 (java 1.5)[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >>
> "Accept-Encoding:
>     >>     > gzip,deflate[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "[\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "
>   let
>     >>     > $exampleSet := [[\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "
>     >>  {"name" :
>     >>     > "Ann", "age" : 20, "salary" : 100000},[\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "
>     >>  {"name" :
>     >>     > "Bob", "age" : 30, "salary" : 200000},[\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "
>     >>  {"name" :
>     >>     > "Cat", "age" : 40, "salary" : 300000, "dependents" : [1, 2,
>     >> 3]},[\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "
>     >>  {"name" :
>     >>     > "Cat", "age" : 50, "salary" : 400000}[\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "
>     >> ][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "
>   for
>     >> $x in
>     >>     > $exampleSet[\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "
>     >> return
>     >>     > $x[\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 >> "         "
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << "HTTP/1.1
> 200
>     >>     > OK[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 <<
> "Content-Type:
>     >>     > application/json;charset=utf-8[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 <<
>     >> "Transfer-Encoding:
>     >>     > chunked[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << "Server:
>     >>     > Jetty(8.0.0.RC0)[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << "[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 <<
> "11[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 <<
>     >>     > "{"handle":[13,0]}[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << "0[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-0 << "[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 <<
> HTTP/1.1 200
>     >> OK
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 <<
> Content-Type:
>     >>     > application/json;charset=utf-8
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 <<
>     >> Transfer-Encoding:
>     >>     > chunked
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-0 << Server:
>     >>     > Jetty(8.0.0.RC0)
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Connection can be
> kept
>     >> alive
>     >>     > indefinitely
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > [id: 0][route: {}->http://10.128.5.170:19002] can be kept
> alive
>     >>     > indefinitely
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > released: [id: 0][route: {}->http://10.128.5.170:19002][total
> kept
>     >> alive:
>     >>     > 1; route al
>     >>     > > located: 1 of 2; total allocated: 1 of 20]
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > manager is shutting down
>     >>     > > 16/07/26 23:27:11 DEBUG DefaultManagedHttpClientConnection:
>     >>     > http-outgoing-0: Close connection
>     >>     > > 16/07/26 23:27:11 DEBUG DefaultManagedHttpClientConnection:
>     >>     > http-outgoing-0: Close connection
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > manager shut down
>     >>     > > 16/07/26 23:27:11 INFO AsterixHttpAPI: Response Handle:
>     >> {"handle":[13,0]}
>     >>     > > 16/07/26 23:27:11 INFO AsterixHttpAPI: Handle(JID:13,RSID:0)
>     >>     > > 16/07/26 23:27:11 DEBUG AsterixHttpAPI: Get status of:
>     >>     > %7B%22handle%22%3A%5B13%2C0%5D%7D
>     >>     > > 16/07/26 23:27:11 DEBUG RequestAddCookies: CookieSpec
> selected:
>     >>     > best-match
>     >>     > > 16/07/26 23:27:11 DEBUG RequestAuthCache: Auth cache not
> set in
>     >> the
>     >>     > context
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > request: [route: {}->http://10.128.5.170:19002][total kept
> alive:
>     >> 0;
>     >>     > route allocated:
>     >>     > > 0 of 2; total allocated: 0 of 20]
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > leased: [id: 1][route: {}->http://10.128.5.170:19002][total
> kept
>     >> alive:
>     >>     > 0; route allo
>     >>     > > cated: 1 of 2; total allocated: 1 of 20]
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Opening connection
> {}->
>     >>     > http://10.128.5.170:19002
>     >>     > > 16/07/26 23:27:11 DEBUG HttpClientConnectionManager:
> Connecting
>     >> to /
>     >>     > 10.128.5.170:19002
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Executing request
> GET
>     >>     > /query/status?handle=%7B%22handle%22%3A%5B13%2C0%5D%7D
> HTTP/1.1
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Target auth state:
>     >> UNCHALLENGED
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Proxy auth state:
>     >> UNCHALLENGED
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 >> GET
>     >>     > /query/status?handle=%7B%22handle%22%3A%5B13%2C0%5D%7D
> HTTP/1.1
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 >> Host:
>     >>     > 10.128.5.170:19002
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 >>
> Connection:
>     >>     > Keep-Alive
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 >>
> User-Agent:
>     >>     > Apache-HttpClient/4.3.2 (java 1.5)
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 >>
>     >> Accept-Encoding:
>     >>     > gzip,deflate
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >> "GET
>     >>     > /query/status?handle=%7B%22handle%22%3A%5B13%2C0%5D%7D
>     >> HTTP/1.1[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >> "Host:
>     >>     > 10.128.5.170:19002[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >>
> "Connection:
>     >>     > Keep-Alive[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >>
> "User-Agent:
>     >>     > Apache-HttpClient/4.3.2 (java 1.5)[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >>
> "Accept-Encoding:
>     >>     > gzip,deflate[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 >> "[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 << "HTTP/1.1
> 200
>     >>     > OK[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 <<
> "Content-Type:
>     >>     > text/html;charset=UTF-8[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 <<
> "Content-Length:
>     >>     > 20[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 << "Server:
>     >>     > Jetty(8.0.0.RC0)[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 << "[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-1 <<
>     >> "{"status":"SUCCESS"}"
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 <<
> HTTP/1.1 200
>     >> OK
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 <<
> Content-Type:
>     >>     > text/html;charset=UTF-8
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 <<
>     >> Content-Length: 20
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-1 << Server:
>     >>     > Jetty(8.0.0.RC0)
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Connection can be
> kept
>     >> alive
>     >>     > indefinitely
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > [id: 1][route: {}->http://10.128.5.170:19002] can be kept
> alive
>     >>     > indefinitely
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > released: [id: 1][route: {}->http://10.128.5.170:19002][total
> kept
>     >> alive:
>     >>     > 1; route al
>     >>     > > located: 1 of 2; total allocated: 1 of 20]
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > manager is shutting down
>     >>     > > 16/07/26 23:27:11 DEBUG DefaultManagedHttpClientConnection:
>     >>     > http-outgoing-1: Close connection
>     >>     > > 16/07/26 23:27:11 DEBUG DefaultManagedHttpClientConnection:
>     >>     > http-outgoing-1: Close connection
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > manager shut down
>     >>     > > 16/07/26 23:27:11 DEBUG AsterixHttpAPI: Get locations of:
>     >>     > %7B%22handle%22%3A%5B13%2C0%5D%7D
>     >>     > > 16/07/26 23:27:11 DEBUG RequestAddCookies: CookieSpec
> selected:
>     >>     > best-match
>     >>     > > 16/07/26 23:27:11 DEBUG RequestAuthCache: Auth cache not
> set in
>     >> the
>     >>     > context
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > request: [route: {}->http://10.128.5.170:19002][total kept
> alive:
>     >> 0;
>     >>     > route allocated:
>     >>     > > 0 of 2; total allocated: 0 of 20]
>     >>     > > 16/07/26 23:27:11 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > leased: [id: 2][route: {}->http://10.128.5.170:19002][total
> kept
>     >> alive:
>     >>     > 0; route allo
>     >>     > > cated: 1 of 2; total allocated: 1 of 20]
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Opening connection
> {}->
>     >>     > http://10.128.5.170:19002
>     >>     > > 16/07/26 23:27:11 DEBUG HttpClientConnectionManager:
> Connecting
>     >> to /
>     >>     > 10.128.5.170:19002
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Executing request
> GET
>     >>     > /query/result/location?handle=%7B%22handle%22%3A%5B13%2C0%
> 5D%7D
>     >> HTTP/1.1
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Target auth state:
>     >> UNCHALLENGED
>     >>     > > 16/07/26 23:27:11 DEBUG MainClientExec: Proxy auth state:
>     >> UNCHALLENGED
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-2 >> GET
>     >>     > /query/result/location?handle=%7B%22handle%22%3A%5B13%2C0%
> 5D%7D
>     >> HTTP/1.1
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-2 >> Host:
>     >>     > 10.128.5.170:19002
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-2 >>
> Connection:
>     >>     > Keep-Alive
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-2 >>
> User-Agent:
>     >>     > Apache-HttpClient/4.3.2 (java 1.5)
>     >>     > > 16/07/26 23:27:11 DEBUG headers: http-outgoing-2 >>
>     >> Accept-Encoding:
>     >>     > gzip,deflate
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >> "GET
>     >>     > /query/result/location?handle=%7B%22handle%22%3A%5B13%2C0%
> 5D%7D
>     >>     > HTTP/1.1[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >> "Host:
>     >>     > 10.128.5.170:19002[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >>
> "Connection:
>     >>     > Keep-Alive[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >>
> "User-Agent:
>     >>     > Apache-HttpClient/4.3.2 (java 1.5)[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >>
> "Accept-Encoding:
>     >>     > gzip,deflate[\r][\n]"
>     >>     > > 16/07/26 23:27:11 DEBUG wire: http-outgoing-2 >> "[\r][\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "HTTP/1.1
> 404 Not
>     >>     > Found[\r][\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 <<
> "Cache-Control:
>     >>     > must-revalidate,no-cache,no-store[\r][\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 <<
> "Content-Type:
>     >>     > text/html;charset=ISO-8859-1[\r][\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 <<
> "Content-Length:
>     >>     > 1288[\r][\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "Server:
>     >>     > Jetty(8.0.0.RC0)[\r][\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "[\r][\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 <<
> "<html>[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 <<
> "<head>[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "<meta
>     >>     > http-equiv="Content-Type"
>     >> content="text/html;charset=ISO-8859-1"/>[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 <<
> "<title>Error
>     >> 404 Not
>     >>     > Found</title>[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 <<
> "</head>[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 <<
> "<body>[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "<h2>HTTP
> ERROR:
>     >>     > 404</h2>[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "<p>Problem
>     >> accessing
>     >>     > /query/result/location. Reason:[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "<pre>
> Not
>     >>     > Found</pre></p>[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "<hr
>     >>     > /><i><small>Powered by Jetty://</small></i>[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 << "
>     >>     >                           [\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 <<
> "</body>[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG wire: http-outgoing-2 <<
> "</html>[\n]"
>     >>     > > 16/07/26 23:27:12 DEBUG headers: http-outgoing-2 <<
> HTTP/1.1 404
>     >> Not
>     >>     > Found
>     >>     > > 16/07/26 23:27:12 DEBUG headers: http-outgoing-2 <<
> Cache-Control:
>     >>     > must-revalidate,no-cache,no-store
>     >>     > > 16/07/26 23:27:12 DEBUG headers: http-outgoing-2 <<
> Content-Type:
>     >>     > text/html;charset=ISO-8859-1
>     >>     > > 16/07/26 23:27:12 DEBUG headers: http-outgoing-2 <<
>     >> Content-Length: 1288
>     >>     > > 16/07/26 23:27:12 DEBUG headers: http-outgoing-2 << Server:
>     >>     > Jetty(8.0.0.RC0)
>     >>     > > 16/07/26 23:27:12 DEBUG MainClientExec: Connection can be
> kept
>     >> alive
>     >>     > indefinitely
>     >>     > > 16/07/26 23:27:12 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > [id: 2][route: {}->http://10.128.5.170:19002] can be kept
> alive
>     >>     > indefinitely
>     >>     > > 16/07/26 23:27:12 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > released: [id: 2][route: {}->http://10.128.5.170:19002][total
> kept
>     >> alive:
>     >>     > 1; route al
>     >>     > > located: 1 of 2; total allocated: 1 of 20]
>     >>     > > 16/07/26 23:27:12 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > manager is shutting down
>     >>     > > 16/07/26 23:27:12 DEBUG DefaultManagedHttpClientConnection:
>     >>     > http-outgoing-2: Close connection
>     >>     > > 16/07/26 23:27:12 DEBUG DefaultManagedHttpClientConnection:
>     >>     > http-outgoing-2: Close connection
>     >>     > > 16/07/26 23:27:12 DEBUG PoolingHttpClientConnectionManager:
>     >> Connection
>     >>     > manager shut down
>     >>     > > 16/07/26 23:27:12 INFO AsterixHttpAPI: Result Locations:
> <html>
>     >>     > > <head>
>     >>     > > <meta http-equiv="Content-Type"
>     >> content="text/html;charset=ISO-8859-1"/>
>     >>     > > <title>Error 404 Not Found</title>
>     >>     > > </head>
>     >>     > > <body>
>     >>     > > <h2>HTTP ERROR: 404</h2>
>     >>     > > <p>Problem accessing /query/result/location. Reason:
>     >>     > > <pre>    Not Found</pre></p>
>     >>     > > <hr /><i><small>Powered by Jetty://</small></i>
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > > </body>
>     >>     > > </html>
>     >>     > >
>     >>     > > net.liftweb.json.JsonParser$ParseException: unknown token <
>     >>     > > Near: <h
>     >>     > >        at
>     >> net.liftweb.json.JsonParser$Parser.fail(JsonParser.scala:234)
>     >>     > >        at
>     >>     > net.liftweb.json.JsonParser$Parser.nextToken(JsonParser.
> scala:321)
>     >>     > >        at
>     >>     > net.liftweb.json.JsonParser$$anonfun$2.apply(JsonParser.
> scala:188)
>     >>     > >        at
>     >>     > net.liftweb.json.JsonParser$$anonfun$2.apply(JsonParser.
> scala:141)
>     >>     > >        at net.liftweb.json.JsonParser$.
> parse(JsonParser.scala:80)
>     >>     > >        at net.liftweb.json.JsonParser$.
> parse(JsonParser.scala:45)
>     >>     > >        at net.liftweb.json.package$.parse(package.scala:41)
>     >>     > >        at
>     >> net.liftweb.json.Serialization$.read(Serialization.scala:58)
>     >>     > >        at
>     >>     >
>     >> org.apache.asterix.connector.AsterixHttpAPI.getResultLocations(
> AsterixHttpAPI.scala:129)
>     >>     > >        at
>     >>     >
>     >> org.apache.asterix.connector.SparkContextFunctions.executeQuery(
> SparkContextFunctions.scala:103)
>     >>     > >        at
>     >>     >
>     >> org.apache.asterix.connector.SparkContextFunctions.aql(
> SparkContextFunctions.scala:80)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.sql.asterix.SQLContextFunctions.executeQuery(
> SQLContextFunctions.scala:103)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.sql.asterix.SQLContextFunctions.aql(
> SQLContextFunctions.scala:84)
>     >>     > >        at
>     >>     >
>     >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.
> <init>(<console>:35)
>     >>     > >        at
>     >>     >
>     >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<
> init>(<console>:40)
>     >>     > >        at
>     >>     >
>     >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<
> console>:42)
>     >>     > >        at
>     >>     > $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<
> console>:44)
>     >>     > >        at
>     >> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:46)
>     >>     > >        at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$
> $iwC.<init>(<console>:48)
>     >>     > >        at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.
> <init>(<console>:50)
>     >>     > >        at $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:52)
>     >>     > >        at $iwC$$iwC$$iwC$$iwC.<init>(<console>:54)
>     >>     > >        at $iwC$$iwC$$iwC.<init>(<console>:56)
>     >>     > >        at $iwC$$iwC.<init>(<console>:58)
>     >>     > >        at $iwC.<init>(<console>:60)
>     >>     > >        at <init>(<console>:62)
>     >>     > >        at .<init>(<console>:66)
>     >>     > >        at .<clinit>(<console>)
>     >>     > >        at .<init>(<console>:7)
>     >>     > >        at .<clinit>(<console>)
>     >>     > >        at $print(<console>)
>     >>     > >        at sun.reflect.NativeMethodAccessorImpl.
> invoke0(Native
>     >> Method)
>     >>     > >        at
>     >>     >
>     >> sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:62)
>     >>     > >        at
>     >>     >
>     >> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>     >>     > >        at java.lang.reflect.Method.invoke(Method.java:498)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(
> SparkIMain.scala:1065)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.repl.SparkIMain$Request.loadAndRun(
> SparkIMain.scala:1346)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.repl.SparkIMain.loadAndRunReq$1(
> SparkIMain.scala:840)
>     >>     > >        at
>     >>     > org.apache.spark.repl.SparkIMain.interpret(
> SparkIMain.scala:871)
>     >>     > >        at
>     >>     > org.apache.spark.repl.SparkIMain.interpret(
> SparkIMain.scala:819)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.repl.SparkILoop.reallyInterpret$1(
> SparkILoop.scala:857)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.repl.SparkILoop.interpretStartingWith(
> SparkILoop.scala:902)
>     >>     > >        at
>     >> org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814)
>     >>     > >        at
>     >>     > org.apache.spark.repl.SparkILoop.processLine$1(
> SparkILoop.scala:657)
>     >>     > >        at
>     >>     > org.apache.spark.repl.SparkILoop.innerLoop$1(
> SparkILoop.scala:665)
>     >>     > >        at org.apache.spark.repl.SparkILoop.org
>     >>     > $apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.repl.SparkILoop$$anonfun$org$
> apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.repl.SparkILoop$$anonfun$org$
> apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.repl.SparkILoop$$anonfun$org$
> apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
>     >>     > >        at
>     >>     >
>     >> scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(
> ScalaClassLoader.scala:135)
>     >>     > >        at org.apache.spark.repl.SparkILoop.org
>     >>     > $apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945)
>     >>     > >        at
>     >> org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059)
>     >>     > >        at org.apache.spark.repl.Main$.main(Main.scala:31)
>     >>     > >        at org.apache.spark.repl.Main.main(Main.scala)
>     >>     > >        at sun.reflect.NativeMethodAccessorImpl.
> invoke0(Native
>     >> Method)
>     >>     > >        at
>     >>     >
>     >> sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:62)
>     >>     > >        at
>     >>     >
>     >> sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>     >>     > >        at java.lang.reflect.Method.invoke(Method.java:498)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
>     >>     > >        at
>     >>     >
>     >> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> SparkSubmit.scala:181)
>     >>     > >        at
>     >>     > org.apache.spark.deploy.SparkSubmit$.submit(
> SparkSubmit.scala:206)
>     >>     > >        at
>     >>     > org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> scala:121)
>     >>     > >        at
>     >> org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > > From: Wail Alkowaileet <wael.y.k@gmail.com>
>     >>     > > Reply-To: "users@asterixdb.apache.org" <
>     >> users@asterixdb.apache.org>
>     >>     > > Date: Monday, July 18, 2016 at 5:26 AM
>     >>     > > To: "dev@asterixdb.apache.org" <dev@asterixdb.apache.org>,
> "
>     >>     > users@asterixdb.apache.org" <users@asterixdb.apache.org>
>     >>     > > Subject: Re: Trio: AsterixDB, Spark and Zeppelin.
>     >>     > >
>     >>     > > Sorry. Here's the link for the connecot:
>     >>     > > https://github.com/Nullification/asterixdb-spark-connector
>     >>     > >
>     >>     > > On Mon, Jul 18, 2016 at 2:34 PM, Wail Alkowaileet <
>     >> wael.y.k@gmail.com>
>     >>     > wrote:
>     >>     > > Dears,
>     >>     > >
>     >>     > > Finally I finished cleaning and documenting the
> AsterixDB-Spark
>     >>     > connector and finalize Zeppelin interpreter for AQL and SQL++.
>     >>     > >
>     >>     > > AsterixDB-Spark Connector:
>     >>     > > • Supports both AQL and SQL++ queries.
>     >>     > > • Much cleaner code now.
>     >>     > > • Please if you have ANY problem with it, create an issue
> in the
>     >> project
>     >>     > repo.
>     >>     > > • I'm working on a tutorial-video from the build to use it
> in
>     >> Zeppelin.
>     >>     > > • I recommend you to use Zeppelin. (you can import the
> connector
>     >> example
>     >>     > notebook)
>     >>     > > Source Code:
>     >> https://github.com/Nullification/astreixdb-spark-connector
>     >>     > >
>     >>     > > Apache Zeppelin with AsterixDB interpreter:
>     >>     > > • Supports JSON-flattening (which will allow zeppelin to
> visualize
>     >>     > results).
>     >>     > > • See attached screenshots.
>     >>     > > • Will try to initiate pull request to merge it to Zeppelin
>     >> master.
>     >>     > > Source Code: https://github.com/Nullification/zeppelin
>     >>     > >
>     >>     > > Finally, I just submitted Schema Inferencer. I have work on
> some
>     >> Sonar
>     >>     > comments and it should be ready soon.
>     >>     > >
>     >>     > > Thanks!
>     >>     > >
>     >>     > > --
>     >>     > >
>     >>     > > Regards,
>     >>     > > Wail Alkowaileet
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > >
>     >>     > > --
>     >>     > >
>     >>     > > Regards,
>     >>     > > Wail Alkowaileet
>     >>     > >
>     >>     > >
>     >>     >
>     >>     > Best regards,
>     >>     > Ildar
>     >>     >
>     >>     >
>     >>
>     >>
>     >>     --
>     >>
>     >>     *Regards,*
>     >>     Wail Alkowaileet
>     >>
>     >>
>     >>
>     >
>     >
>     > --
>     >
>     > *Regards,*
>     > Wail Alkowaileet
>     >
>
>
>
>     --
>
>     *Regards,*
>     Wail Alkowaileet
>
>
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message