Any news on this problem? I did some more research and I found a similar
issue logged into zeppelin issue tracker at
https://issues.apache.org/jira/browse/ZEPPELIN-194

The exception is not exactly the same but it may be due to a different
version of com.databricks:spark-csv_2.10, which is 1.1.0 vs. 1.2.0 that I'm
using. Thanks in advance.

- michele


On Mon, Oct 19, 2015 at 4:18 PM, michele crudele <mic.crud...@gmail.com>
wrote:

> Thanks, I just did it. Some progresses, but not yet there.
> Now in the paragraph:
>
> --3--
> %sql
> select * from sms
>
> I'm getting the following exception. I googled other people getting
> similar exception but did not find root cause. Any idea of what's going on?
> Note that my env is very simple, I just built zeppelin as per your
> suggestion, started it, and tried my supersimple notebook.
>
> java.lang.ClassNotFoundException:
> com.databricks.spark.csv.CsvRelation$$anonfun$tokenRdd$1$$anonfun$1 at
> java.net.URLClassLoader$1.run(URLClassLoader.java:366) at
> java.net.URLClassLoader$1.run(URLClassLoader.java:355) at
> java.security.AccessController.doPrivileged(Native Method) at
> java.net.URLClassLoader.findClass(URLClassLoader.java:354) at
> java.lang.ClassLoader.loadClass(ClassLoader.java:425) at
> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) at
> java.lang.ClassLoader.loadClass(ClassLoader.java:358) at
> java.lang.Class.forName0(Native Method) at
> java.lang.Class.forName(Class.java:274) at
> org.apache.spark.util.InnerClosureFinder$$anon$4.visitMethodInsn(ClosureCleaner.scala:455)
> at
> com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.accept(Unknown
> Source) at
> com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.accept(Unknown
> Source) at
> org.apache.spark.util.ClosureCleaner$.getInnerClosureClasses(ClosureCleaner.scala:101)
> at
> org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:197)
> at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:132) at
> org.apache.spark.SparkContext.clean(SparkContext.scala:1893) at
> org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1.apply(RDD.scala:683) at
> org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1.apply(RDD.scala:682) at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108)
> at org.apache.spark.rdd.RDD.withScope(RDD.scala:286) at
> org.apache.spark.rdd.RDD.mapPartitions(RDD.scala:682) at
> com.databricks.spark.csv.CsvRelation.tokenRdd(CsvRelation.scala:90) at
> com.databricks.spark.csv.CsvRelation.buildScan(CsvRelation.scala:105) at
> org.apache.spark.sql.sources.DataSourceStrategy$.apply(DataSourceStrategy.scala:101)
> at
> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
> at
> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
> at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) at
> org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
> at
> org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54)
> at
> org.apache.spark.sql.execution.SparkStrategies$BasicOperators$.apply(SparkStrategies.scala:314)
> at
> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
> at
> org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
> at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) at
> org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
> at
> org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:943)
> at
> org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:941)
> at
> org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:947)
> at
> org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:947)
> at org.apache.spark.sql.DataFrame.collect(DataFrame.scala:1269) at
> org.apache.spark.sql.DataFrame.head(DataFrame.scala:1203) at
> org.apache.spark.sql.DataFrame.take(DataFrame.scala:1262) at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606) at
> org.apache.zeppelin.spark.ZeppelinContext.showDF(ZeppelinContext.java:300)
> at
> org.apache.zeppelin.spark.SparkSqlInterpreter.interpret(SparkSqlInterpreter.java:142)
> at
> org.apache.zeppelin.interpreter.ClassloaderInterpreter.interpret(ClassloaderInterpreter.java:57)
> at
> org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93)
> at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:276)
> at org.apache.zeppelin.scheduler.Job.run(Job.java:170) at
> org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:118)
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
> at java.util.concurrent.FutureTask.run(FutureTask.java:262) at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
> at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
>
>
> On Mon, Oct 19, 2015 at 1:59 PM, moon soo Lee <m...@apache.org> wrote:
>
>> You'll need to build your self.
>> Please let me know if you have any problem on building Zeppelin.
>>
>> Thanks,
>> moon
>>
>>
>> On Mon, Oct 19, 2015 at 8:32 PM michele crudele <mic.crud...@gmail.com>
>> wrote:
>>
>>> Thanks moon,
>>>
>>> is there a repo where I can download the 0.6.0-SNAPHOT or do I have to
>>> build zeppelin myself ? Thanks again for your help.
>>>
>>> - michele
>>>
>>>
>>> On Mon, Oct 19, 2015 at 7:41 AM, moon soo Lee <m...@apache.org> wrote:
>>>
>>>> Hi Michele,
>>>>
>>>> Thanks for sharing the problem.
>>>> I have tested your code on both 0.5.0 and 0.6.0-SNAPSHOT.
>>>> I have the same problem on 0.5.0 but 0.6.0-SNAPSHOT runs smoothly.
>>>>
>>>> So, you can try 0.6.0-SNAPSHOT until next release is out.
>>>> Or if you want to see what's going on with 0.5.0, you'll need to apply
>>>> this commit
>>>> https://github.com/apache/incubator-zeppelin/commit/d0a30435414726e7fa6d8b8e106e4b6ddb46da67
>>>>  to
>>>> see exception in your notebook.
>>>>
>>>> Best,
>>>> moon
>>>>
>>>> On Sat, Oct 17, 2015 at 7:18 AM michele crudele <mic.crud...@gmail.com>
>>>> wrote:
>>>>
>>>>> I've created a supersimple notebook:
>>>>>
>>>>> -- 1 --
>>>>> %dep
>>>>>
>>>>> z.reset
>>>>> z.load("com.databricks:spark-csv_2.10:1.2.0")
>>>>>
>>>>> --2--
>>>>> %spark
>>>>>
>>>>> val smsFile = "/home/barabba/data/SMSSpamCollection.csv"
>>>>> sqlContext.load("com.databricks.spark.csv", Map("path" -> smsFile,
>>>>> "header" -> "true", "delimiter" -> "|")).registerTempTable("sms")
>>>>>
>>>>> --3--
>>>>> %sql
>>>>> select * from sms
>>>>>
>>>>> 1 and 2 runs fine, while 3 displays the
>>>>> java.lang.reflect.InvocationTargetException. I cannot find anything useful
>>>>> in the logs.
>>>>> If I add the paragraph
>>>>>
>>>>> --4--
>>>>> %spark
>>>>> sqlContext.sql("select * from sms").show
>>>>>
>>>>> it works correctly, showing the top rows of the table.
>>>>> +--+----+--------------------+
>>>>> |id|type|                text|
>>>>> +--+----+--------------------+
>>>>> | 0| ham|Go until jurong p...|
>>>>> | 1| ham|Ok lar... Joking ...|
>>>>> | 2|spam|Free entry in 2 a...|
>>>>> | 3| ham|U dun say so earl...|
>>>>> | 4| ham|Nah I don't think...|
>>>>> | 5|spam|FreeMsg Hey there...|
>>>>> | 6| ham|Even my brother i...|
>>>>> | 7| ham|As per your reque...|
>>>>> ...
>>>>>
>>>>> I'm using zeppelin-0.5.0-incubating-bin-spark-1.3.1_hadoop-2.3.
>>>>>
>>>>>
>>>>> Any idea of what's going on? Thanks
>>>>> - michele
>>>>>
>>>>>
>>>
>

Reply via email to