Hi Lee,
Thanks for the info that really helped. I set the compression codec in the 
Spark side -i.e. inside the SPARK_HOME and now the problem resolved. However, I 
was wondering if it's possible to set the same from the Zeppelin notebook.
I tried in the following way: 

%spark conf.set("spark.io.compression.codec", "lz4") 

But getting an error. Please suggest. 



    On Thursday, February 16, 2017 7:40 AM, Jongyoul Lee <jongy...@gmail.com> 
wrote:
 

 Hi, Can you check if the script passes in spark-shell or not? AFAIK, you have 
to add compression codec by yourself in Spark side.
On Wed, Feb 15, 2017 at 1:10 AM, Muhammad Rezaul Karim <reza_cse...@yahoo.com> 
wrote:

Hi All,
I am receiving the following exception while executing SQL queries: 
 java.lang. NoSuchMethodException: org.apache.spark.io. 
LZ4CompressionCodec.<init>( org.apache.spark.SparkConf)
    at java.lang.Class. getConstructor0(Class.java: 3082)
    at java.lang.Class. getConstructor(Class.java: 1825)
    at org.apache.spark.io. CompressionCodec$.createCodec( 
CompressionCodec.scala:71)
    at org.apache.spark.io. CompressionCodec$.createCodec( 
CompressionCodec.scala:65)
    at org.apache.spark.sql. execution.SparkPlan.org$ 
apache$spark$sql$execution$ SparkPlan$$decodeUnsafeRows( SparkPlan.scala:250)
    at org.apache.spark.sql. execution.SparkPlan$$anonfun$ 
executeCollect$1.apply( SparkPlan.scala:276)
    at org.apache.spark.sql. execution.SparkPlan$$anonfun$ 
executeCollect$1.apply( SparkPlan.scala:275)
    at scala.collection. IndexedSeqOptimized$class. 
foreach(IndexedSeqOptimized. scala:33)
    at scala.collection.mutable. ArrayOps$ofRef.foreach( ArrayOps.scala:186)
    at org.apache.spark.sql. execution.SparkPlan. executeCollect(SparkPlan. 
scala:275)
    at org.apache.spark.sql. execution.exchange. BroadcastExchangeExec$$ 
anonfun$relationFuture$1$$ anonfun$apply$1.apply( BroadcastExchangeExec.scala: 
78)
    at org.apache.spark.sql. execution.exchange. BroadcastExchangeExec$$ 
anonfun$relationFuture$1$$ anonfun$apply$1.apply( BroadcastExchangeExec.scala: 
75)
    at org.apache.spark.sql. execution.SQLExecution$. 
withExecutionId(SQLExecution. scala:94)
    at org.apache.spark.sql. execution.exchange. BroadcastExchangeExec$$ 
anonfun$relationFuture$1. apply(BroadcastExchangeExec. scala:74)
    at org.apache.spark.sql. execution.exchange. BroadcastExchangeExec$$ 
anonfun$relationFuture$1. apply(BroadcastExchangeExec. scala:74)
    at scala.concurrent.impl.Future$ PromiseCompletingRunnable. 
liftedTree1$1(Future.scala:24)
    at scala.concurrent.impl.Future$ PromiseCompletingRunnable.run( 
Future.scala:24)
    at java.util.concurrent. ThreadPoolExecutor.runWorker( 
ThreadPoolExecutor.java:1142)
    at java.util.concurrent. ThreadPoolExecutor$Worker.run( 
ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread. java:745)

My SQL query is: 
%sql  select * from land where Price >= 10000 AND CLUSTER = 2 

I am experiencing the above exception in the 1st run always but when I 
re-execute the same query for the 2nd or 3rd time, I don't get this error. 

Am I doing something wrong? Someone, please help me out. 





Kinds regards,Reza




-- 
이종열, Jongyoul Lee, 李宗烈
http://madeng.net


   

Reply via email to