Actually found the culprit. The JavaSerializerInstance.deserialize is called with a classloader (of type MutableURLClassLoader) which has access to all the hive classes. But internally it triggers a call to loadClass but with the default classloader. Below is the stacktrace (line numbers in the JavaSerialization.scala will be a bit off due to my debugging statements).
I will try out the SPARK_CLASSPATH setting. But I was wondering if this has something to do with the way spark-project.hive jars are created v/s the way open source apache-hive jars are created. Is this documented somewhere? The only info I see is Patrick Wendell's comment in https://github.com/apache/spark/pull/2241 (grep for "published a modified version"). 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: Uncaught exception in thread Thread[Executor task launch worker-3,5,main] java.lang.NoClassDefFoundError: org/apache/hadoop/hive/conf/HiveConf 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.net.URLClassLoader$1.run(URLClassLoader.java:366) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.net.URLClassLoader$1.run(URLClassLoader.java:355) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.security.AccessController.doPrivileged(Native Method) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.net.URLClassLoader.findClass(URLClassLoader.java:354) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.lang.ClassLoader.loadClass(ClassLoader.java:425) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.lang.ClassLoader.loadClass(ClassLoader.java:358) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.lang.Class.getDeclaredFields0(Native Method) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.lang.Class.privateGetDeclaredFields(Class.java:2436) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.lang.Class.getDeclaredField(Class.java:1946) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectStreamClass.getDeclaredSUID(ObjectStreamClass.java:1659) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectStreamClass.access$700(ObjectStreamClass.java:72) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:480) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:468) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.security.AccessController.doPrivileged(Native Method) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectStreamClass.<init>(ObjectStreamClass.java:468) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:365) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:602) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: scala.collection.immutable.$colon$colon.readObject(List.scala:362) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.lang.reflect.Method.invoke(Method.java:606) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:67) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:103) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:57) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: org.apache.spark.scheduler.Task.run(Task.scala:64) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:206) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) 15/04/25 01:41:04 ERROR util.SparkUncaughtExceptionHandler: java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) On 22 April 2015 at 13:07, Akhil Das <ak...@sigmoidanalytics.com> wrote: > I see, now try a bit tricky approach, Add the hive jar to the > SPARK_CLASSPATH (in conf/spark-env.sh file on all machines) and make sure > that jar is available on all the machines in the cluster in the same path. > > Thanks > Best Regards > > On Wed, Apr 22, 2015 at 11:24 AM, Manku Timma <manku.tim...@gmail.com> > wrote: > >> Akhil, Thanks for the suggestions. >> I tried out sc.addJar, --jars, --conf spark.executor.extraClassPath and >> none of them helped. I added stuff into compute-classpath.sh. That did not >> change anything. I checked the classpath of the running executor and made >> sure that the hive jars are in that dir. For me the most confusing thing is >> that the executor can actually create HiveConf objects but when it cannot >> find that when the task deserializer is at work. >> >> On 20 April 2015 at 14:18, Akhil Das <ak...@sigmoidanalytics.com> wrote: >> >>> Can you try sc.addJar("/path/to/your/hive/jar"), i think it will resolve >>> it. >>> >>> Thanks >>> Best Regards >>> >>> On Mon, Apr 20, 2015 at 12:26 PM, Manku Timma <manku.tim...@gmail.com> >>> wrote: >>> >>>> Akhil, >>>> But the first case of creating HiveConf on the executor works fine (map >>>> case). Only the second case fails. I was suspecting some foul play with >>>> classloaders. >>>> >>>> On 20 April 2015 at 12:20, Akhil Das <ak...@sigmoidanalytics.com> >>>> wrote: >>>> >>>>> Looks like a missing jar, try to print the classpath and make sure the >>>>> hive jar is present. >>>>> >>>>> Thanks >>>>> Best Regards >>>>> >>>>> On Mon, Apr 20, 2015 at 11:52 AM, Manku Timma <manku.tim...@gmail.com> >>>>> wrote: >>>>> >>>>>> I am using spark-1.3 with hadoop-provided and hive-provided and >>>>>> hive-0.13.1 profiles. I am running a simple spark job on a yarn cluster >>>>>> by >>>>>> adding all hadoop2 and hive13 jars to the spark classpaths. >>>>>> >>>>>> If I remove the hive-provided while building spark, I dont face any >>>>>> issue. But with hive-provided I am getting a >>>>>> "java.lang.NoClassDefFoundError: org/apache/hadoop/hive/conf/HiveConf" in >>>>>> the yarn executor. >>>>>> >>>>>> Code is below: >>>>>> import org.apache.spark._ >>>>>> import org.apache.spark.sql._ >>>>>> import org.apache.hadoop.hive.conf.HiveConf >>>>>> >>>>>> object Simple { >>>>>> def main(args: Array[String]) = { >>>>>> val sc = new SparkContext(new SparkConf()) >>>>>> val sqlC = new org.apache.spark.sql.hive.HiveContext(sc) >>>>>> >>>>>> val x = sc.parallelize(1 to 2).map(x => >>>>>> { val h = new HiveConf; h.getBoolean("hive.test", false) }) >>>>>> x.collect.foreach(x => println(s"----------------- $x >>>>>> ----------------")) >>>>>> >>>>>> val result = sqlC.sql(""" >>>>>> select * from products_avro order by month, name, price >>>>>> """) >>>>>> result.collect.foreach(println) >>>>>> } >>>>>> } >>>>>> >>>>>> The first job (involving map) runs fine. HiveConf is instantiated and >>>>>> the conf variable is looked up etc. But the second job (involving the >>>>>> select * query) throws the class not found exception. >>>>>> >>>>>> The task deserializer is the one throwing the exception. It is unable >>>>>> to find the class in its classpath. Not sure what is different from the >>>>>> first job which also involved HiveConf. >>>>>> >>>>>> 157573 [task-result-getter-3] 2015/04/20 11:01:48:287 WARN >>>>>> TaskSetManager: Lost task 0.2 in stage 2.0 (TID 4, localhost): >>>>>> java.lang.NoClassDefFoundError: org/apache/hadoop/hive/conf/HiveConf >>>>>> at java.lang.Class.getDeclaredFields0(Native Method) >>>>>> at java.lang.Class.privateGetDeclaredFields(Class.java:2436) >>>>>> at java.lang.Class.getDeclaredField(Class.java:1946) >>>>>> at >>>>>> java.io.ObjectStreamClass.getDeclaredSUID(ObjectStreamClass.java:1659) >>>>>> at java.io.ObjectStreamClass.access$700(ObjectStreamClass.java:72) >>>>>> at java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:480) >>>>>> at java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:468) >>>>>> at java.security.AccessController.doPrivileged(Native Method) >>>>>> at java.io.ObjectStreamClass.<init>(ObjectStreamClass.java:468) >>>>>> at java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:365) >>>>>> at >>>>>> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:602) >>>>>> at >>>>>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622) >>>>>> at >>>>>> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) >>>>>> at >>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) >>>>>> at >>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) >>>>>> at >>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) >>>>>> at >>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) >>>>>> at >>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) >>>>>> at >>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) >>>>>> at >>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) >>>>>> at >>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) >>>>>> at >>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) >>>>>> at >>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) >>>>>> at >>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) >>>>>> at >>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) >>>>>> at >>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) >>>>>> at >>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) >>>>>> at >>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) >>>>>> at >>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) >>>>>> at >>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) >>>>>> at >>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) >>>>>> at >>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) >>>>>> at >>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) >>>>>> at >>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) >>>>>> at >>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) >>>>>> at >>>>>> scala.collection.immutable.$colon$colon.readObject(List.scala:362) >>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>>> at >>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>>> at >>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>>>> at >>>>>> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017) >>>>>> at >>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893) >>>>>> at >>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) >>>>>> at >>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) >>>>>> at >>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) >>>>>> at >>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) >>>>>> at >>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) >>>>>> at >>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) >>>>>> at >>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) >>>>>> at >>>>>> java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) >>>>>> at >>>>>> org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:68) >>>>>> at >>>>>> org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:94) >>>>>> at >>>>>> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:57) >>>>>> at org.apache.spark.scheduler.Task.run(Task.scala:64) >>>>>> at >>>>>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203) >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>> Caused by: java.lang.ClassNotFoundException: >>>>>> org.apache.hadoop.hive.conf.HiveConf >>>>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:366) >>>>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:355) >>>>>> at java.security.AccessController.doPrivileged(Native Method) >>>>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:354) >>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:425) >>>>>> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:358) >>>>>> ... 70 more >>>>>> >>>>>> -- Thanks >>>>>> >>>>> >>>>> >>>> >>> >> >