[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]

2014-10-24 Thread Suhas Satish (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183290#comment-14183290
 ] 

Suhas Satish commented on HIVE-7916:


Not sure what solved it for you, but setting this seems to work for me on a Mac 
OS X -
export HADOOP_OPTS=-Dorg.xerial.snappy.tempdir=/tmp 
-Dorg.xerial.snappy.lib.name=libsnappyjava.jnilib $HADOOP_OPTS


 Snappy-java error when running hive query on spark [Spark Branch]
 -

 Key: HIVE-7916
 URL: https://issues.apache.org/jira/browse/HIVE-7916
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Xuefu Zhang
  Labels: Spark-M1

 Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. 
 While the new version addressed some lib conflicts (such as guava), I'm 
 afraid that it also introduced new problems. The following might be one, when 
 I set the master URL to be a spark standalone cluster:
 {code}
 hive set hive.execution.engine=spark;
 hive set spark.serializer=org.apache.spark.serializer.KryoSerializer;
 hive set spark.master=spark://xzdt:7077;
 hive select name, avg(value) from dec group by name;
 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as 
 values in memory (estimated size 333.0 KB, free 128.0 MB)
 java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at org.xerial.snappy.SnappyOutputStream.init(SnappyOutputStream.java:79)
 at 
 org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124)
 at 
 org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.init(TorrentBroadcast.scala:68)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29)
 at 
 org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62)
 at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809)
 at org.apache.spark.rdd.HadoopRDD.init(HadoopRDD.scala:116)
 at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541)
 at 
 org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156)
 at 
 org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198)
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408)
 at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
 Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path
 at 

[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]

2014-10-24 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183309#comment-14183309
 ] 

Xuefu Zhang commented on HIVE-7916:
---

Yeah. I'm on Ubuntu, and I didn't set up anything to HADOOP_OPTS. It just 
magically works now. :)

 Snappy-java error when running hive query on spark [Spark Branch]
 -

 Key: HIVE-7916
 URL: https://issues.apache.org/jira/browse/HIVE-7916
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Xuefu Zhang
  Labels: Spark-M1

 Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. 
 While the new version addressed some lib conflicts (such as guava), I'm 
 afraid that it also introduced new problems. The following might be one, when 
 I set the master URL to be a spark standalone cluster:
 {code}
 hive set hive.execution.engine=spark;
 hive set spark.serializer=org.apache.spark.serializer.KryoSerializer;
 hive set spark.master=spark://xzdt:7077;
 hive select name, avg(value) from dec group by name;
 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as 
 values in memory (estimated size 333.0 KB, free 128.0 MB)
 java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at org.xerial.snappy.SnappyOutputStream.init(SnappyOutputStream.java:79)
 at 
 org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124)
 at 
 org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.init(TorrentBroadcast.scala:68)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29)
 at 
 org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62)
 at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809)
 at org.apache.spark.rdd.HadoopRDD.init(HadoopRDD.scala:116)
 at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541)
 at 
 org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156)
 at 
 org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198)
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408)
 at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
 Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path
 at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1860)
 at java.lang.Runtime.loadLibrary0(Runtime.java:845)
 at 

[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]

2014-10-17 Thread Suhas Satish (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14175712#comment-14175712
 ] 

Suhas Satish commented on HIVE-7916:


I also hit the following snappy lib exceptions - I am using snappy 
snappy-java-1.0.5.jar. Let me try upgrading to snappy 1.1.1.3

2014-10-17 16:18:01,977 ERROR [Executor task launch worker-0]: 
executor.Executor (Logging.scala:logError(96)) - Exception in task 0.0 in stage 
0.0 (TID 0)
org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
at 
org.xerial.snappy.SnappyOutputStream.init(SnappyOutputStream.java:79)
at 
org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:125)
at 
org.apache.spark.storage.BlockManager.wrapForCompression(BlockManager.scala:1083)
at 
org.apache.spark.storage.BlockManager$$anonfun$7.apply(BlockManager.scala:579)
at 
org.apache.spark.storage.BlockManager$$anonfun$7.apply(BlockManager.scala:579)
at 
org.apache.spark.storage.DiskBlockObjectWriter.open(BlockObjectWriter.scala:126)
at 
org.apache.spark.storage.DiskBlockObjectWriter.write(BlockObjectWriter.scala:192)
at 
org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4$$anonfun$apply$2.apply(ExternalSorter.scala:732)
at 
org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4$$anonfun$apply$2.apply(ExternalSorter.scala:731)
at scala.collection.Iterator$class.foreach(Iterator.scala:727)
at 
org.apache.spark.util.collection.ExternalSorter$IteratorForPartition.foreach(ExternalSorter.scala:789)
at 
org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4.apply(ExternalSorter.scala:731)
at 
org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4.apply(ExternalSorter.scala:727)
at scala.collection.Iterator$class.foreach(Iterator.scala:727)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
at 
org.apache.spark.util.collection.ExternalSorter.writePartitionedFile(ExternalSorter.scala:727)
at 
org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:70)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
at org.apache.spark.scheduler.Task.run(Task.scala:56)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:181)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)

--
2014-10-17 16:18:02,021 INFO  [main]: scheduler.DAGScheduler 
(Logging.scala:logInfo(59)) - Job 0 failed: foreach at SparkPlan.java:80, took 
3.389683 s
2014-10-17 16:18:02,021 ERROR [main]: spark.SparkClient 
(SparkClient.java:execute(166)) - Error executing Spark Plan
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in 
stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 
0, localhost): org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] 
null
org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
org.xerial.snappy.Snappy.clinit(Snappy.java:44)
org.xerial.snappy.SnappyOutputStream.init(SnappyOutputStream.java:79)

org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:125)

org.apache.spark.storage.BlockManager.wrapForCompression(BlockManager.scala:1083)

org.apache.spark.storage.BlockManager$$anonfun$7.apply(BlockManager.scala:579)

org.apache.spark.storage.BlockManager$$anonfun$7.apply(BlockManager.scala:579)

org.apache.spark.storage.DiskBlockObjectWriter.open(BlockObjectWriter.scala:126)

org.apache.spark.storage.DiskBlockObjectWriter.write(BlockObjectWriter.scala:192)

org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4$$anonfun$apply$2.apply(ExternalSorter.scala:732)

org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4$$anonfun$apply$2.apply(ExternalSorter.scala:731)
scala.collection.Iterator$class.foreach(Iterator.scala:727)

org.apache.spark.util.collection.ExternalSorter$IteratorForPartition.foreach(ExternalSorter.scala:789)

org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4.apply(ExternalSorter.scala:731)

org.apache.spark.util.collection.ExternalSorter$$anonfun$writePartitionedFile$4.apply(ExternalSorter.scala:727)

[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]

2014-10-17 Thread Suhas Satish (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14175771#comment-14175771
 ] 

Suhas Satish commented on HIVE-7916:


Hitting the same problem with snappy 1.1.1.3 as well. Using hive tar ball as of 
today (fri, oct 17, 2014) with spark.master=local


 Snappy-java error when running hive query on spark [Spark Branch]
 -

 Key: HIVE-7916
 URL: https://issues.apache.org/jira/browse/HIVE-7916
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Xuefu Zhang
  Labels: Spark-M1

 Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. 
 While the new version addressed some lib conflicts (such as guava), I'm 
 afraid that it also introduced new problems. The following might be one, when 
 I set the master URL to be a spark standalone cluster:
 {code}
 hive set hive.execution.engine=spark;
 hive set spark.serializer=org.apache.spark.serializer.KryoSerializer;
 hive set spark.master=spark://xzdt:7077;
 hive select name, avg(value) from dec group by name;
 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as 
 values in memory (estimated size 333.0 KB, free 128.0 MB)
 java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at org.xerial.snappy.SnappyOutputStream.init(SnappyOutputStream.java:79)
 at 
 org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124)
 at 
 org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.init(TorrentBroadcast.scala:68)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29)
 at 
 org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62)
 at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809)
 at org.apache.spark.rdd.HadoopRDD.init(HadoopRDD.scala:116)
 at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541)
 at 
 org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156)
 at 
 org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198)
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408)
 at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
 Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path
 at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1860)
 at 

[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]

2014-09-02 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14119247#comment-14119247
 ] 

Xuefu Zhang commented on HIVE-7916:
---

[~lirui], I tried both branch 1.1 and master. Spark 1.1 branch doesn't have the 
guava lib fix, right? So, using that branch, I hit the guava lib conflict. 
Using master gives the snappy error.

I'm thinking if you can take a look at the guide at 
https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started#HiveonSpark:GettingStarted-HiveonSpark:GettingStarted
 and update it if necessary. Then, I can try to follow the guide to see if I 
still get the snappy problem or guava problem.

Thanks!

 Snappy-java error when running hive query on spark [Spark Branch]
 -

 Key: HIVE-7916
 URL: https://issues.apache.org/jira/browse/HIVE-7916
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Xuefu Zhang
  Labels: Spark-M1

 Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. 
 While the new version addressed some lib conflicts (such as guava), I'm 
 afraid that it also introduced new problems. The following might be one, when 
 I set the master URL to be a spark standalone cluster:
 {code}
 hive set hive.execution.engine=spark;
 hive set spark.serializer=org.apache.spark.serializer.KryoSerializer;
 hive set spark.master=spark://xzdt:7077;
 hive select name, avg(value) from dec group by name;
 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as 
 values in memory (estimated size 333.0 KB, free 128.0 MB)
 java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at org.xerial.snappy.SnappyOutputStream.init(SnappyOutputStream.java:79)
 at 
 org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124)
 at 
 org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.init(TorrentBroadcast.scala:68)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29)
 at 
 org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62)
 at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809)
 at org.apache.spark.rdd.HadoopRDD.init(HadoopRDD.scala:116)
 at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541)
 at 
 org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156)
 at 
 org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198)
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408)
 at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]

2014-09-01 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14117872#comment-14117872
 ] 

Xuefu Zhang commented on HIVE-7916:
---

[~lirui] Thanks for looking into this. I had noticed the version difference, 
and thought for a moment it was the cause. However, I don't see any 
/tmp/snappy*.so file in my machine. I also built spark-assembly myself, but 
with slightly different params:
{code}
mvn -Dhadoop.version=2.3.0-cdh5.0.1 -Phadoop-2.3 -DskipTests clean install.
{code}

Aslo, I'm not sure if my machine has ever run snappy before. Is there anything 
that needs to be done in order for it to work?

I can try again to reproduce the problem.




 Snappy-java error when running hive query on spark [Spark Branch]
 -

 Key: HIVE-7916
 URL: https://issues.apache.org/jira/browse/HIVE-7916
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Xuefu Zhang
  Labels: Spark-M1

 Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. 
 While the new version addressed some lib conflicts (such as guava), I'm 
 afraid that it also introduced new problems. The following might be one, when 
 I set the master URL to be a spark standalone cluster:
 {code}
 hive set hive.execution.engine=spark;
 hive set spark.serializer=org.apache.spark.serializer.KryoSerializer;
 hive set spark.master=spark://xzdt:7077;
 hive select name, avg(value) from dec group by name;
 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as 
 values in memory (estimated size 333.0 KB, free 128.0 MB)
 java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at org.xerial.snappy.SnappyOutputStream.init(SnappyOutputStream.java:79)
 at 
 org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124)
 at 
 org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.init(TorrentBroadcast.scala:68)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29)
 at 
 org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62)
 at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809)
 at org.apache.spark.rdd.HadoopRDD.init(HadoopRDD.scala:116)
 at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541)
 at 
 org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156)
 at 
 org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198)
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408)
 at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
  

[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]

2014-09-01 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14117911#comment-14117911
 ] 

Rui Li commented on HIVE-7916:
--

Hi [~xuefuz], do you use the latest code of spark 1.1 branch? SPARK-2881 is 
resolved with [#1999|https://github.com/apache/spark/pull/1999] for branch-1.1. 
You can check the pom file of your spark to verify that, latest code uses 
snappy-java-1.0.5.3.

 Snappy-java error when running hive query on spark [Spark Branch]
 -

 Key: HIVE-7916
 URL: https://issues.apache.org/jira/browse/HIVE-7916
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Xuefu Zhang
  Labels: Spark-M1

 Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. 
 While the new version addressed some lib conflicts (such as guava), I'm 
 afraid that it also introduced new problems. The following might be one, when 
 I set the master URL to be a spark standalone cluster:
 {code}
 hive set hive.execution.engine=spark;
 hive set spark.serializer=org.apache.spark.serializer.KryoSerializer;
 hive set spark.master=spark://xzdt:7077;
 hive select name, avg(value) from dec group by name;
 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as 
 values in memory (estimated size 333.0 KB, free 128.0 MB)
 java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at org.xerial.snappy.SnappyOutputStream.init(SnappyOutputStream.java:79)
 at 
 org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124)
 at 
 org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.init(TorrentBroadcast.scala:68)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29)
 at 
 org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62)
 at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809)
 at org.apache.spark.rdd.HadoopRDD.init(HadoopRDD.scala:116)
 at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541)
 at 
 org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156)
 at 
 org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198)
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408)
 at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
 Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in 

[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]

2014-08-31 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14117050#comment-14117050
 ] 

Rui Li commented on HIVE-7916:
--

Hi [~xuefuz], I tried on my cluster but cannot reproduce the problem. I removed 
the spark jars from local maven repo before building hive, so that the jars are 
downloaded from the AWS server we maintain. After hive is built, I linked the 
spark-assembly jar to {{lib}} of the hive home directory. The spark-assembly 
jar is built with {{mvn -Pyarn -Phadoop-2.4 -DskipTests clean package}} of the 
spark 1.1 branch.
Could you provide more info about your environment, e.g. the spark jars you 
used or if the table is snappy compressed?

 Snappy-java error when running hive query on spark [Spark Branch]
 -

 Key: HIVE-7916
 URL: https://issues.apache.org/jira/browse/HIVE-7916
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Xuefu Zhang
  Labels: Spark-M1

 Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. 
 While the new version addressed some lib conflicts (such as guava), I'm 
 afraid that it also introduced new problems. The following might be one, when 
 I set the master URL to be a spark standalone cluster:
 {code}
 hive set hive.execution.engine=spark;
 hive set spark.serializer=org.apache.spark.serializer.KryoSerializer;
 hive set spark.master=spark://xzdt:7077;
 hive select name, avg(value) from dec group by name;
 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as 
 values in memory (estimated size 333.0 KB, free 128.0 MB)
 java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at org.xerial.snappy.SnappyOutputStream.init(SnappyOutputStream.java:79)
 at 
 org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124)
 at 
 org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.init(TorrentBroadcast.scala:68)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29)
 at 
 org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62)
 at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809)
 at org.apache.spark.rdd.HadoopRDD.init(HadoopRDD.scala:116)
 at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541)
 at 
 org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156)
 at 
 org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198)
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408)
 at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 

[jira] [Commented] (HIVE-7916) Snappy-java error when running hive query on spark [Spark Branch]

2014-08-31 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14117057#comment-14117057
 ] 

Rui Li commented on HIVE-7916:
--

I noted this may be related to SPARK-2881. Snappy-java is bumped to 1.0.5.3 in 
the 1.1 branch and to 1.1.1.3 in the master branch. Hadoop-2.4.0 seems to use 
snappy-java-1.0.4.1.
While the snappy-java version is different, I don't see any conflicts on my 
side.
[~xuefuz], I found the following in the description of SPARK-2881:
{quote}
The issue was that someone else had run with snappy and it created 
/tmp/snappy-*.so but it had restrictive permissions so I was not able to use it 
or remove it. This caused my spark job to not start.
{quote}
Could you check if this is the case in your environment?

 Snappy-java error when running hive query on spark [Spark Branch]
 -

 Key: HIVE-7916
 URL: https://issues.apache.org/jira/browse/HIVE-7916
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Xuefu Zhang
  Labels: Spark-M1

 Recently spark branch upgraded its dependency on Spark to 1.1.0-SNAPSHOT. 
 While the new version addressed some lib conflicts (such as guava), I'm 
 afraid that it also introduced new problems. The following might be one, when 
 I set the master URL to be a spark standalone cluster:
 {code}
 hive set hive.execution.engine=spark;
 hive set spark.serializer=org.apache.spark.serializer.KryoSerializer;
 hive set spark.master=spark://xzdt:7077;
 hive select name, avg(value) from dec group by name;
 14/08/28 16:41:52 INFO storage.MemoryStore: Block broadcast_0 stored as 
 values in memory (estimated size 333.0 KB, free 128.0 MB)
 java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
 at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
 at org.xerial.snappy.Snappy.clinit(Snappy.java:44)
 at org.xerial.snappy.SnappyOutputStream.init(SnappyOutputStream.java:79)
 at 
 org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:124)
 at 
 org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:83)
 at 
 org.apache.spark.broadcast.TorrentBroadcast.init(TorrentBroadcast.scala:68)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:36)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:29)
 at 
 org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62)
 at org.apache.spark.SparkContext.broadcast(SparkContext.scala:809)
 at org.apache.spark.rdd.HadoopRDD.init(HadoopRDD.scala:116)
 at org.apache.spark.SparkContext.hadoopRDD(SparkContext.scala:541)
 at 
 org.apache.spark.api.java.JavaSparkContext.hadoopRDD(JavaSparkContext.scala:318)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generateRDD(SparkPlanGenerator.java:160)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator.generate(SparkPlanGenerator.java:88)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkClient.execute(SparkClient.java:156)
 at 
 org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.submit(SparkSessionImpl.java:52)
 at 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:77)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:161)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1537)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1304)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1116)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:940)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:930)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:246)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:198)
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:408)
 at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:781)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:614)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at