Actually found the culprit. The JavaSerializerInstance.deserialize is
called with a classloader (of type MutableURLClassLoader) which has access
to all the hive classes. But internally it triggers a call to loadClass but
with the default classloader. Below is the stacktrace (line numbers in the
JavaSerialization.scala will be a bit off due to my debugging statements).

I will try out the SPARK_CLASSPATH setting. But I was wondering if this has
something to do with the way spark-project.hive jars are created v/s the
way open source apache-hive jars are created. Is this documented somewhere?
The only info I see is Patrick Wendell's comment in
https://github.com/apache/spark/pull/2241 (grep for "published a modified
version").

15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: Uncaught
exception in thread Thread[Executor task launch worker-3,5,main]
java.lang.NoClassDefFoundError: org/apache/hadoop/hive/conf/HiveConf

15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.net.URLClassLoader$1.run(URLClassLoader.java:366)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.net.URLClassLoader$1.run(URLClassLoader.java:355)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.security.AccessController.doPrivileged(Native Method)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.net.URLClassLoader.findClass(URLClassLoader.java:354)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.lang.ClassLoader.loadClass(ClassLoader.java:425)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.lang.ClassLoader.loadClass(ClassLoader.java:358)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.lang.Class.getDeclaredFields0(Native Method)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.lang.Class.privateGetDeclaredFields(Class.java:2436)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.lang.Class.getDeclaredField(Class.java:1946)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectStreamClass.getDeclaredSUID(ObjectStreamClass.java:1659)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectStreamClass.access$700(ObjectStreamClass.java:72)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:480)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:468)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.security.AccessController.doPrivileged(Native Method)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectStreamClass.<init>(ObjectStreamClass.java:468)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:365)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:602)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
scala.collection.immutable.$colon$colon.readObject(List.scala:362)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.lang.reflect.Method.invoke(Method.java:606)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:67)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:103)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:57)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
org.apache.spark.scheduler.Task.run(Task.scala:64)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:206)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler:
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)


On 22 April 2015 at 13:07, Akhil Das <ak...@sigmoidanalytics.com> wrote:

> I see, now try a bit tricky approach, Add the hive jar to the
> SPARK_CLASSPATH (in conf/spark-env.sh file on all machines) and make sure
> that jar is available on all the machines in the cluster in the same path.
>
> Thanks
> Best Regards
>
> On Wed, Apr 22, 2015 at 11:24 AM, Manku Timma <manku.tim...@gmail.com>
> wrote:
>
>> Akhil, Thanks for the suggestions.
>> I tried out sc.addJar, --jars, --conf spark.executor.extraClassPath and
>> none of them helped. I added stuff into compute-classpath.sh. That did not
>> change anything. I checked the classpath of the running executor and made
>> sure that the hive jars are in that dir. For me the most confusing thing is
>> that the executor can actually create HiveConf objects but when it cannot
>> find that when the task deserializer is at work.
>>
>> On 20 April 2015 at 14:18, Akhil Das <ak...@sigmoidanalytics.com> wrote:
>>
>>> Can you try sc.addJar("/path/to/your/hive/jar"), i think it will resolve
>>> it.
>>>
>>> Thanks
>>> Best Regards
>>>
>>> On Mon, Apr 20, 2015 at 12:26 PM, Manku Timma <manku.tim...@gmail.com>
>>> wrote:
>>>
>>>> Akhil,
>>>> But the first case of creating HiveConf on the executor works fine (map
>>>> case). Only the second case fails. I was suspecting some foul play with
>>>> classloaders.
>>>>
>>>> On 20 April 2015 at 12:20, Akhil Das <ak...@sigmoidanalytics.com>
>>>> wrote:
>>>>
>>>>> Looks like a missing jar, try to print the classpath and make sure the
>>>>> hive jar is present.
>>>>>
>>>>> Thanks
>>>>> Best Regards
>>>>>
>>>>> On Mon, Apr 20, 2015 at 11:52 AM, Manku Timma <manku.tim...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> I am using spark-1.3 with hadoop-provided and hive-provided and
>>>>>> hive-0.13.1 profiles. I am running a simple spark job on a yarn cluster 
>>>>>> by
>>>>>> adding all hadoop2 and hive13 jars to the spark classpaths.
>>>>>>
>>>>>> If I remove the hive-provided while building spark, I dont face any
>>>>>> issue. But with hive-provided I am getting a
>>>>>> "java.lang.NoClassDefFoundError: org/apache/hadoop/hive/conf/HiveConf" in
>>>>>> the yarn executor.
>>>>>>
>>>>>> Code is below:
>>>>>> import org.apache.spark._
>>>>>> import org.apache.spark.sql._
>>>>>> import org.apache.hadoop.hive.conf.HiveConf
>>>>>>
>>>>>> object Simple {
>>>>>>   def main(args: Array[String]) = {
>>>>>>     val sc = new SparkContext(new SparkConf())
>>>>>>     val sqlC = new  org.apache.spark.sql.hive.HiveContext(sc)
>>>>>>
>>>>>>     val x = sc.parallelize(1 to 2).map(x =>
>>>>>>       { val h = new HiveConf; h.getBoolean("hive.test", false) })
>>>>>>     x.collect.foreach(x => println(s"-----------------  $x
>>>>>> ----------------"))
>>>>>>
>>>>>>     val result = sqlC.sql("""
>>>>>>       select * from products_avro order by month, name, price
>>>>>>       """)
>>>>>>     result.collect.foreach(println)
>>>>>>   }
>>>>>> }
>>>>>>
>>>>>> The first job (involving map) runs fine. HiveConf is instantiated and
>>>>>> the conf variable is looked up etc. But the second job (involving the
>>>>>> select * query) throws the class not found exception.
>>>>>>
>>>>>> The task deserializer is the one throwing the exception. It is unable
>>>>>> to find the class in its classpath. Not sure what is different from the
>>>>>> first job which also involved HiveConf.
>>>>>>
>>>>>> 157573 [task-result-getter-3] 2015/04/20 11:01:48:287 WARN
>>>>>> TaskSetManager: Lost task 0.2 in stage 2.0 (TID 4, localhost):
>>>>>> java.lang.NoClassDefFoundError: org/apache/hadoop/hive/conf/HiveConf
>>>>>>     at java.lang.Class.getDeclaredFields0(Native Method)
>>>>>>     at java.lang.Class.privateGetDeclaredFields(Class.java:2436)
>>>>>>     at java.lang.Class.getDeclaredField(Class.java:1946)
>>>>>>     at
>>>>>> java.io.ObjectStreamClass.getDeclaredSUID(ObjectStreamClass.java:1659)
>>>>>>     at java.io.ObjectStreamClass.access$700(ObjectStreamClass.java:72)
>>>>>>     at java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:480)
>>>>>>     at java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:468)
>>>>>>     at java.security.AccessController.doPrivileged(Native Method)
>>>>>>     at java.io.ObjectStreamClass.<init>(ObjectStreamClass.java:468)
>>>>>>     at java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:365)
>>>>>>     at
>>>>>> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:602)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
>>>>>>     at
>>>>>> scala.collection.immutable.$colon$colon.readObject(List.scala:362)
>>>>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>>     at
>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>>>>     at
>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>>>     at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>>     at
>>>>>> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>>     at
>>>>>> java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
>>>>>>     at
>>>>>> org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:68)
>>>>>>     at
>>>>>> org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:94)
>>>>>>     at
>>>>>> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:57)
>>>>>>     at org.apache.spark.scheduler.Task.run(Task.scala:64)
>>>>>>     at
>>>>>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203)
>>>>>>     at
>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>>>     at
>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>>>>     at java.lang.Thread.run(Thread.java:745)
>>>>>> Caused by: java.lang.ClassNotFoundException:
>>>>>> org.apache.hadoop.hive.conf.HiveConf
>>>>>>     at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
>>>>>>     at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>>>>>>     at java.security.AccessController.doPrivileged(Native Method)
>>>>>>     at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>>>>>>     at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>>>>>>     at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>>>>>>     at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
>>>>>>     ... 70 more
>>>>>>
>>>>>> -- Thanks
>>>>>>
>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to