[ https://issues.apache.org/jira/browse/SPARK-7029?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Sean Owen resolved SPARK-7029. ------------------------------ Resolution: Invalid This is more of a question than valid issue report at this stage. It is likely your environment; you don't have Hive classes imported, perhaps. > Unable to use hive built-in functions in sparkSQL > ------------------------------------------------- > > Key: SPARK-7029 > URL: https://issues.apache.org/jira/browse/SPARK-7029 > Project: Spark > Issue Type: Test > Components: SQL > Affects Versions: 1.2.0 > Reporter: Aditya Parmar > > Trying to use hive built-in functions in spark sql. facing classnotfound > exception. > Caused by: java.lang.ClassNotFoundException: > org.apache.hadoop.hive.ql.udf.generic.GenericUDF > at java.net.URLClassLoader$1.run(URLClassLoader.java:366) > at java.net.URLClassLoader$1.run(URLClassLoader.java:355) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:354) > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > ... 160 more > 2015-04-20 18:00:45,663 INFO [sparkDriver-akka.actor.default-dispatcher-15] > scheduler.TaskSetManager (Logging.scala:logInfo(59)) - Starting task 0.1 in > stage 1.0 (TID 3, , PROCESS_LOCAL, 1107 bytes) > 2015-04-20 18:00:45,988 ERROR [sparkDriver-akka.actor.default-dispatcher-4] > scheduler.TaskSchedulerImpl (Logging.scala:logError(75)) - Lost executor 0 on > : remote Akka client disassociated > 2015-04-20 18:00:45,994 INFO [sparkDriver-akka.actor.default-dispatcher-4] > scheduler.TaskSetManager (Logging.scala:logInfo(59)) - Re-queueing tasks for > 0 from TaskSet 1.0 > 2015-04-20 18:00:45,995 INFO [sparkDriver-akka.actor.default-dispatcher-7] > client.AppClient$ClientActor (Logging.scala:logInfo(59)) - Executor updated: > app-20150420180040-0003/0 is now EXITED (Command exited with code 50) > 2015-04-20 18:00:45,998 INFO [sparkDriver-akka.actor.default-dispatcher-7] > cluster.SparkDeploySchedulerBackend (Logging.scala:logInfo(59)) - Executor > app-20150420180040-0003/0 removed: Command exited with code 50 > 2015-04-20 18:00:46,003 WARN [sparkDriver-akka.actor.default-dispatcher-4] > scheduler.TaskSetManager (Logging.scala:logWarning(71)) - Lost task 0.1 in > stage 1.0 (TID 3, ): ExecutorLostFailure (executor 0 lost) > 2015-04-20 18:00:46,005 ERROR [sparkDriver-akka.actor.default-dispatcher-4] > cluster.SparkDeploySchedulerBackend (Logging.scala:logError(75)) - Asked to > remove non-existent executor 0 > 2015-04-20 18:00:46,005 ERROR [sparkDriver-akka.actor.default-dispatcher-4] > cluster.SparkDeploySchedulerBackend (Logging.scala:logError(75)) - Asked to > remove non-existent executor 0 > 2015-04-20 18:00:46,005 INFO [sparkDriver-akka.actor.default-dispatcher-19] > scheduler.DAGScheduler (Logging.scala:logInfo(59)) - Executor lost: 0 (epoch > 1) > 2015-04-20 18:00:46,005 ERROR [sparkDriver-akka.actor.default-dispatcher-4] > cluster.SparkDeploySchedulerBackend (Logging.scala:logError(75)) - Asked to > remove non-existent executor 0 > 2015-04-20 18:00:46,005 INFO [sparkDriver-akka.actor.default-dispatcher-7] > client.AppClient$ClientActor (Logging.scala:logInfo(59)) - Executor added: > app-20150420180040-0003/1 on worker-20150420155831- -7078 ( :7078) with 4 > cores > 2015-04-20 18:00:46,006 INFO [sparkDriver-akka.actor.default-dispatcher-7] > cluster.SparkDeploySchedulerBackend (Logging.scala:logInfo(59)) - Granted > executor ID app-20150420180040-0003/1 on hostPort :7078 with 4 cores, 512.0 > MB RAM > 2015-04-20 18:00:46,006 INFO [sparkDriver-akka.actor.default-dispatcher-17] > storage.BlockManagerMasterActor (Logging.scala:logInfo(59)) - Trying to > remove executor 0 from BlockManagerMaster. > 2015-04-20 18:00:46,006 INFO [sparkDriver-akka.actor.default-dispatcher-7] > client.AppClient$ClientActor (Logging.scala:logInfo(59)) - Executor updated: > app-20150420180040-0003/1 is now LOADING > 2015-04-20 18:00:46,007 INFO [sparkDriver-akka.actor.default-dispatcher-17] > storage.BlockManagerMasterActor (Logging.scala:logInfo(59)) - Removing block > manager BlockManagerId(0, , 40573) > 2015-04-20 18:00:46,008 ERROR [sparkDriver-akka.actor.default-dispatcher-5] > cluster.SparkDeploySchedulerBackend (Logging.scala:logError(75)) - Asked to > remove non-existent executor 0 > 2015-04-20 18:00:46,008 INFO [sparkDriver-akka.actor.default-dispatcher-19] > storage.BlockManagerMaster (Logging.scala:logInfo(59)) - Removed 0 > successfully in removeExecutor > 2015-04-20 18:00:46,010 INFO [sparkDriver-akka.actor.default-dispatcher-19] > scheduler.Stage (Logging.scala:logInfo(59)) - Stage 0 is now unavailable on > executor 0 (0/2, false) > 2015-04-20 18:00:46,011 ERROR [sparkDriver-akka.actor.default-dispatcher-6] > cluster.SparkDeploySchedulerBackend (Logging.scala:logError(75)) - Asked to > remove non-existent executor 0 > 2015-04-20 18:00:46,012 INFO [sparkDriver-akka.actor.default-dispatcher-6] > client.AppClient$ClientActor (Logging.scala:logInfo(59)) - Executor updated: > app-20150420180040-0003/1 is now RUNNING > 2015-04-20 18:00:47,215 INFO [sparkDriver-akka.actor.default-dispatcher-6] > cluster.SparkDeploySchedulerBackend (Logging.scala:logInfo(59)) - Registered > executor: Actor[akka.tcp://sparkExecutor@ :38892/user/Executor#-1691085362] > with ID 1 > 2015-04-20 18:00:47,216 INFO [sparkDriver-akka.actor.default-dispatcher-6] > scheduler.TaskSetManager (Logging.scala:logInfo(59)) - Starting task 0.2 in > stage 1.0 (TID 4, , PROCESS_LOCAL, 1107 bytes) > 2015-04-20 18:00:47,480 INFO [sparkDriver-akka.actor.default-dispatcher-19] > storage.BlockManagerMasterActor (Logging.scala:logInfo(59)) - Registering > block manager :58200 with 265.4 MB RAM, BlockManagerId(1, , 58200) > 2015-04-20 18:00:47,688 INFO [sparkDriver-akka.actor.default-dispatcher-4] > storage.BlockManagerInfo (Logging.scala:logInfo(59)) - Added > broadcast_2_piece0 in memory on :58200 (size: 45.8 KB, free: 265.4 MB) > 2015-04-20 18:00:47,916 INFO [task-result-getter-3] scheduler.TaskSetManager > (Logging.scala:logInfo(59)) - Lost task 0.2 in stage 1.0 (TID 4) on executor > : java.lang.NoClassDefFoundError > (Lorg/apache/hadoop/hive/ql/udf/generic/GenericUDF;) [duplicate 1] > 2015-04-20 18:00:47,917 INFO [sparkDriver-akka.actor.default-dispatcher-4] > scheduler.TaskSetManager (Logging.scala:logInfo(59)) - Starting task 0.3 in > stage 1.0 (TID 5, , PROCESS_LOCAL, 1107 bytes) > 2015-04-20 18:00:48,246 ERROR [sparkDriver-akka.actor.default-dispatcher-19] > scheduler.TaskSchedulerImpl (Logging.scala:logError(75)) - Lost executor 1 on > : remote Akka client disassociated > 2015-04-20 18:00:48,246 INFO [sparkDriver-akka.actor.default-dispatcher-19] > scheduler.TaskSetManager (Logging.scala:logInfo(59)) - Re-queueing tasks for > 1 from TaskSet 1.0 > 2015-04-20 18:00:48,246 WARN [sparkDriver-akka.actor.default-dispatcher-19] > scheduler.TaskSetManager (Logging.scala:logWarning(71)) - Lost task 0.3 in > stage 1.0 (TID 5, ): ExecutorLostFailure (executor 1 lost) > 2015-04-20 18:00:48,251 ERROR [sparkDriver-akka.actor.default-dispatcher-19] > scheduler.TaskSetManager (Logging.scala:logError(75)) - Task 0 in stage 1.0 > failed 4 times; aborting job > 2015-04-20 18:00:48,254 INFO [sparkDriver-akka.actor.default-dispatcher-5] > client.AppClient$ClientActor (Logging.scala:logInfo(59)) - Executor updated: > app-20150420180040-0003/1 is now EXITED (Command exited with code 50) > 2015-04-20 18:00:48,254 INFO [sparkDriver-akka.actor.default-dispatcher-5] > cluster.SparkDeploySchedulerBackend (Logging.scala:logInfo(59)) - Executor > app-20150420180040-0003/1 removed: Command exited with code 50 > 2015-04-20 18:00:48,258 INFO [sparkDriver-akka.actor.default-dispatcher-19] > scheduler.TaskSchedulerImpl (Logging.scala:logInfo(59)) - Removed TaskSet > 1.0, whose tasks have all completed, from pool > 2015-04-20 18:00:48,258 ERROR [sparkDriver-akka.actor.default-dispatcher-19] > cluster.SparkDeploySchedulerBackend (Logging.scala:logError(75)) - Asked to > remove non-existent executor 1 > 2015-04-20 18:00:48,260 ERROR [sparkDriver-akka.actor.default-dispatcher-19] > cluster.SparkDeploySchedulerBackend (Logging.scala:logError(75)) - Asked to > remove non-existent executor 1 > 2015-04-20 18:00:48,260 ERROR [sparkDriver-akka.actor.default-dispatcher-19] > cluster.SparkDeploySchedulerBackend (Logging.scala:logError(75)) - Asked to > remove non-existent executor 1 > 2015-04-20 18:00:48,260 INFO [sparkDriver-akka.actor.default-dispatcher-4] > scheduler.TaskSchedulerImpl (Logging.scala:logInfo(59)) - Cancelling stage 1 > 2015-04-20 18:00:48,260 INFO [sparkDriver-akka.actor.default-dispatcher-5] > client.AppClient$ClientActor (Logging.scala:logInfo(59)) - Executor added: > app-20 0-0003/2 on worker-20150420155831- -7078 ( :7078) with 4 > cores > 2015-04-20 18:00:48,260 INFO [sparkDriver-akka.actor.default-dispatcher-5] > cluster.SparkDeploySchedulerBackend (Logging.scala:logInfo(59)) - Granted > executo 150420180040-0003/2 on hostPort :7078 with 4 cores, 512.0 > MB RAM > 2015-04-20 18:00:48,261 INFO [sparkDriver-akka.actor.default-dispatcher-5] > client.AppClient$ClientActor (Logging.scala:logInfo(59)) - Executor updated: > app- 040-0003/2 is now LOADING > 2015-04-20 18:00:48,265 ERROR [sparkDriver-akka.actor.default-dispatcher-19] > cluster.SparkDeploySchedulerBackend (Logging.scala:logError(75)) - Asked to > remo tent executor 1 > 2015-04-20 18:00:48,267 INFO [main] scheduler.DAGScheduler > (Logging.scala:logInfo(59)) - Job 0 failed: saveAsTextFile at > JavaSchemaRDD.scala:42, took 4.8417 > Exception in thread "main" 2015-04-20 18:00:48,267 INFO > [sparkDriver-akka.actor.default-dispatcher-16] client.AppClient$ClientActor > (Logging.scala:logInfo(5 tor updated: app-20150420180040-0003/2 is > now RUNNING > 2015-04-20 18:00:48,268 ERROR [sparkDriver-akka.actor.default-dispatcher-20] > cluster.SparkDeploySchedulerBackend (Logging.scala:logError(75)) - Asked to > remo tent executor 1 > org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in > stage 1.0 failed 4 times, most recent failure: Lost task 0.3 in stage 1.0 > (TID 5 6.igatecorp.com): ExecutorLostFailure (executor 1 lost) > Driver stacktrace: > at > org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1214) > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1203) > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1202) > at > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) > at > org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1202) > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:696) > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:696) > at scala.Option.foreach(Option.scala:236) > at > org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:696) > at > org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1420) > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498) > at akka.actor.ActorCell.invoke(ActorCell.scala:456) > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237) > at akka.dispatch.Mailbox.run(Mailbox.scala:219) > at > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386) > at > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > at > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > 2015-04-20 18:00:48,272 INFO [sparkDriver-akka.actor.default-dispatcher-4] > scheduler.DAGScheduler (Logging.scala:logInfo(59)) - Executor lost: 1 (epoch > 2) > 2015-04-20 18:00:48,272 INFO [sparkDriver-akka.actor.default-dispatcher-3] > storage.BlockManagerMasterActor (Logging.scala:logInfo(59)) - Trying to > remove ex om BlockManagerMaster. > 2015-04-20 18:00:48,272 INFO [sparkDriver-akka.actor.default-dispatcher-3] > storage.BlockManagerMasterActor (Logging.scala:logInfo(59)) - Removing block > mana nagerId(1, , 58200) > 2015-04-20 18:00:48,273 INFO [sparkDriver-akka.actor.default-dispatcher-4] > storage.BlockManagerMaster (Logging.scala:logInfo(59)) - Removed 1 > successfully i cutor -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org