This issue had been fixed. On Mon, Aug 21, 2017 at 4:04 PM, Lu Cao <whuca...@gmail.com> wrote:
> Hi dev, > > I'm trying to insert data from a hive table to carbon table: > > cc.sql("insert into carbon_test select * from target_table where pt = > '20170101'") > > > Any one knows how to fix this error? > > [Stage 8:> (0 + 4) > / 156]17/08/21 15:59:01 WARN scheduler.TaskSetManager: Lost task 1.0 in > stage 8.0 (TID 48, , executor 16): java.lang.ClassNotFoundException: > org.apache.carbondata.spark.rdd.CarbonBlockDistinctValuesCombineRDD > > at org.apache.spark.repl.ExecutorClassLoader.findClass( > ExecutorClassLoader.scala:82) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > > at java.lang.Class.forName0(Native Method) > > at java.lang.Class.forName(Class.java:270) > > at org.apache.spark.serializer.JavaDeserializationStream$$ > anon$1.resolveClass(JavaSerializer.scala:67) > > at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1612) > > at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) > > at java.io.ObjectInputStream.readOrdinaryObject( > ObjectInputStream.java:1771) > > at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > > at java.io.ObjectInputStream.defaultReadFields( > ObjectInputStream.java:1990) > > at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > > at java.io.ObjectInputStream.readOrdinaryObject( > ObjectInputStream.java:1798) > > at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > > at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) > > at org.apache.spark.serializer.JavaDeserializationStream. > readObject(JavaSerializer.scala:75) > > at org.apache.spark.serializer.JavaSerializerInstance. > deserialize(JavaSerializer.scala:114) > > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > ShuffleMapTask.scala:85) > > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > ShuffleMapTask.scala:53) > > at org.apache.spark.scheduler.Task.run(Task.scala:99) > > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:322) > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1145) > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:615) > > at java.lang.Thread.run(Thread.java:745) > > Caused by: java.lang.ClassNotFoundException: org.apache.carbondata.spark. > rdd.CarbonBlockDistinctValuesCombineRDD > > at java.lang.ClassLoader.findClass(ClassLoader.java:531) > > at org.apache.spark.util.ParentClassLoader.findClass( > ParentClassLoader.scala:26) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > > at org.apache.spark.util.ParentClassLoader.loadClass( > ParentClassLoader.scala:34) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > > at org.apache.spark.util.ParentClassLoader.loadClass( > ParentClassLoader.scala:30) > > at org.apache.spark.repl.ExecutorClassLoader.findClass( > ExecutorClassLoader.scala:77) > > ... 23 more > > > [Stage 8:> (0 + 4) > / 156]17/08/21 15:59:02 ERROR scheduler.TaskSetManager: Task 1 in stage 8.0 > failed 4 times; aborting job > > 17/08/21 15:59:02 ERROR util.GlobalDictionaryUtil$: main generate global > dictionary failed > > org.apache.spark.SparkException: Job aborted due to stage failure: Task 1 > in stage 8.0 failed 4 times, most recent failure: Lost task 1.3 in stage > 8.0 (TID 61, scsp00382.saicdt.com, executor 16): > java.lang.ClassNotFoundException: > org.apache.carbondata.spark.rdd.CarbonBlockDistinctValuesCombineRDD > > at org.apache.spark.repl.ExecutorClassLoader.findClass( > ExecutorClassLoader.scala:82) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > > at java.lang.Class.forName0(Native Method) > > at java.lang.Class.forName(Class.java:270) > > at org.apache.spark.serializer.JavaDeserializationStream$$ > anon$1.resolveClass(JavaSerializer.scala:67) > > at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1612) > > at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) > > at java.io.ObjectInputStream.readOrdinaryObject( > ObjectInputStream.java:1771) > > at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > > at java.io.ObjectInputStream.defaultReadFields( > ObjectInputStream.java:1990) > > at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > > at java.io.ObjectInputStream.readOrdinaryObject( > ObjectInputStream.java:1798) > > at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > > at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) > > at org.apache.spark.serializer.JavaDeserializationStream. > readObject(JavaSerializer.scala:75) > > at org.apache.spark.serializer.JavaSerializerInstance. > deserialize(JavaSerializer.scala:114) > > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > ShuffleMapTask.scala:85) > > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > ShuffleMapTask.scala:53) > > at org.apache.spark.scheduler.Task.run(Task.scala:99) > > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:322) > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1145) > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:615) > > at java.lang.Thread.run(Thread.java:745) > > Caused by: java.lang.ClassNotFoundException: org.apache.carbondata.spark. > rdd.CarbonBlockDistinctValuesCombineRDD > > at java.lang.ClassLoader.findClass(ClassLoader.java:531) > > at org.apache.spark.util.ParentClassLoader.findClass( > ParentClassLoader.scala:26) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > > at org.apache.spark.util.ParentClassLoader.loadClass( > ParentClassLoader.scala:34) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > > at org.apache.spark.util.ParentClassLoader.loadClass( > ParentClassLoader.scala:30) > > at org.apache.spark.repl.ExecutorClassLoader.findClass( > ExecutorClassLoader.scala:77) > > ... 23 more > > > Driver stacktrace: > > at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$ > scheduler$DAGScheduler$$failJobAndIndependentStages( > DAGScheduler.scala:1435) > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply( > DAGScheduler.scala:1423) > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply( > DAGScheduler.scala:1422) > > at scala.collection.mutable.ResizableArray$class.foreach( > ResizableArray.scala:59) > > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > > at org.apache.spark.scheduler.DAGScheduler.abortStage( > DAGScheduler.scala:1422) > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$ > handleTaskSetFailed$1.apply(DAGScheduler.scala:802) > > at org.apache.spark.scheduler.DAGScheduler$$anonfun$ > handleTaskSetFailed$1.apply(DAGScheduler.scala:802) > > at scala.Option.foreach(Option.scala:257) > > at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed( > DAGScheduler.scala:802) > > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop. > doOnReceive(DAGScheduler.scala:1650) > > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop. > onReceive(DAGScheduler.scala:1605) > > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop. > onReceive(DAGScheduler.scala:1594) > > at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) > > at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:628) > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1918) > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1931) > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1944) > > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1958) > > at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:935) > > at org.apache.spark.rdd.RDDOperationScope$.withScope( > RDDOperationScope.scala:151) > > at org.apache.spark.rdd.RDDOperationScope$.withScope( > RDDOperationScope.scala:112) > > at org.apache.spark.rdd.RDD.withScope(RDD.scala:362) > > at org.apache.spark.rdd.RDD.collect(RDD.scala:934) > > at org.apache.carbondata.spark.util.GlobalDictionaryUtil$. > generateGlobalDictionary(GlobalDictionaryUtil.scala:741) > > at org.apache.spark.sql.execution.command.LoadTable. > run(carbonTableSchema.scala:703) > > at org.apache.spark.sql.execution.command.LoadTableByInsert.run( > carbonTableSchema.scala:338) > > at org.apache.spark.sql.execution.command.ExecutedCommandExec. > sideEffectResult$lzycompute(commands.scala:58) > > at org.apache.spark.sql.execution.command.ExecutedCommandExec. > sideEffectResult(commands.scala:56) > > at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute( > commands.scala:74) > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$ > execute$1.apply(SparkPlan.scala:114) > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$ > execute$1.apply(SparkPlan.scala:114) > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply( > SparkPlan.scala:135) > > at org.apache.spark.rdd.RDDOperationScope$.withScope( > RDDOperationScope.scala:151) > > at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala: > 132) > > at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113) > > at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute( > QueryExecution.scala:87) > > at org.apache.spark.sql.execution.QueryExecution. > toRdd(QueryExecution.scala:87) > > at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185) > > at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64) > > at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init> > (<console>:33) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:38) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:40) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:42) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:44) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw.<init>(<console>:46) > > at $line30.$read$$iw$$iw$$iw$$iw.<init>(<console>:48) > > at $line30.$read$$iw$$iw$$iw.<init>(<console>:50) > > at $line30.$read$$iw$$iw.<init>(<console>:52) > > at $line30.$read$$iw.<init>(<console>:54) > > at $line30.$read.<init>(<console>:56) > > at $line30.$read$.<init>(<console>:60) > > at $line30.$read$.<clinit>(<console>) > > at $line30.$eval$.$print$lzycompute(<console>:7) > > at $line30.$eval$.$print(<console>:6) > > at $line30.$eval.$print(<console>) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:57) > > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:606) > > at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:786) > > at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1047) > > at scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$ > loadAndRunReq$1.apply(IMain.scala:638) > > at scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$ > loadAndRunReq$1.apply(IMain.scala:637) > > at scala.reflect.internal.util.ScalaClassLoader$class. > asContext(ScalaClassLoader.scala:31) > > at scala.reflect.internal.util.AbstractFileClassLoader.asContext( > AbstractFileClassLoader.scala:19) > > at scala.tools.nsc.interpreter.IMain$WrappedRequest. > loadAndRunReq(IMain.scala:637) > > at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:569) > > at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:565) > > at scala.tools.nsc.interpreter.ILoop.interpretStartingWith( > ILoop.scala:807) > > at scala.tools.nsc.interpreter.ILoop.command(ILoop.scala:681) > > at scala.tools.nsc.interpreter.ILoop.processLine(ILoop.scala:395) > > at scala.tools.nsc.interpreter.ILoop.loop(ILoop.scala:415) > > at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1. > apply$mcZ$sp(ILoop.scala:923) > > at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1. > apply(ILoop.scala:909) > > at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1. > apply(ILoop.scala:909) > > at scala.reflect.internal.util.ScalaClassLoader$.savingContextLoader( > ScalaClassLoader.scala:97) > > at scala.tools.nsc.interpreter.ILoop.process(ILoop.scala:909) > > at org.apache.spark.repl.Main$.doMain(Main.scala:69) > > at org.apache.spark.repl.Main$.main(Main.scala:52) > > at org.apache.spark.repl.Main.main(Main.scala) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:57) > > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:606) > > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > deploy$SparkSubmit$$runMain(SparkSubmit.scala:738) > > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187) > > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212) > > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126) > > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > Caused by: java.lang.ClassNotFoundException: org.apache.carbondata.spark. > rdd.CarbonBlockDistinctValuesCombineRDD > > at org.apache.spark.repl.ExecutorClassLoader.findClass( > ExecutorClassLoader.scala:82) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > > at java.lang.Class.forName0(Native Method) > > at java.lang.Class.forName(Class.java:270) > > at org.apache.spark.serializer.JavaDeserializationStream$$ > anon$1.resolveClass(JavaSerializer.scala:67) > > at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1612) > > at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) > > at java.io.ObjectInputStream.readOrdinaryObject( > ObjectInputStream.java:1771) > > at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > > at java.io.ObjectInputStream.defaultReadFields( > ObjectInputStream.java:1990) > > at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > > at java.io.ObjectInputStream.readOrdinaryObject( > ObjectInputStream.java:1798) > > at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > > at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) > > at org.apache.spark.serializer.JavaDeserializationStream. > readObject(JavaSerializer.scala:75) > > at org.apache.spark.serializer.JavaSerializerInstance. > deserialize(JavaSerializer.scala:114) > > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > ShuffleMapTask.scala:85) > > at org.apache.spark.scheduler.ShuffleMapTask.runTask( > ShuffleMapTask.scala:53) > > at org.apache.spark.scheduler.Task.run(Task.scala:99) > > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:322) > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1145) > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:615) > > at java.lang.Thread.run(Thread.java:745) > > Caused by: java.lang.ClassNotFoundException: org.apache.carbondata.spark. > rdd.CarbonBlockDistinctValuesCombineRDD > > at java.lang.ClassLoader.findClass(ClassLoader.java:531) > > at org.apache.spark.util.ParentClassLoader.findClass( > ParentClassLoader.scala:26) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > > at org.apache.spark.util.ParentClassLoader.loadClass( > ParentClassLoader.scala:34) > > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > > at org.apache.spark.util.ParentClassLoader.loadClass( > ParentClassLoader.scala:30) > > at org.apache.spark.repl.ExecutorClassLoader.findClass( > ExecutorClassLoader.scala:77) > > ... 23 more > > 17/08/21 15:59:02 ERROR command.LoadTable: main > > java.lang.Exception: generate global dictionary failed, null > > at org.apache.carbondata.spark.util.GlobalDictionaryUtil$. > generateGlobalDictionary(GlobalDictionaryUtil.scala:761) > > at org.apache.spark.sql.execution.command.LoadTable. > run(carbonTableSchema.scala:703) > > at org.apache.spark.sql.execution.command.LoadTableByInsert.run( > carbonTableSchema.scala:338) > > at org.apache.spark.sql.execution.command.ExecutedCommandExec. > sideEffectResult$lzycompute(commands.scala:58) > > at org.apache.spark.sql.execution.command.ExecutedCommandExec. > sideEffectResult(commands.scala:56) > > at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute( > commands.scala:74) > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$ > execute$1.apply(SparkPlan.scala:114) > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$ > execute$1.apply(SparkPlan.scala:114) > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply( > SparkPlan.scala:135) > > at org.apache.spark.rdd.RDDOperationScope$.withScope( > RDDOperationScope.scala:151) > > at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala: > 132) > > at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113) > > at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute( > QueryExecution.scala:87) > > at org.apache.spark.sql.execution.QueryExecution. > toRdd(QueryExecution.scala:87) > > at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185) > > at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64) > > at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init> > (<console>:33) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:38) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:40) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:42) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:44) > > at $line30.$read$$iw$$iw$$iw$$iw$$iw.<init>(<console>:46) > > at $line30.$read$$iw$$iw$$iw$$iw.<init>(<console>:48) > > at $line30.$read$$iw$$iw$$iw.<init>(<console>:50) > > at $line30.$read$$iw$$iw.<init>(<console>:52) > > at $line30.$read$$iw.<init>(<console>:54) > > at $line30.$read.<init>(<console>:56) > > at $line30.$read$.<init>(<console>:60) > > at $line30.$read$.<clinit>(<console>) > > at $line30.$eval$.$print$lzycompute(<console>:7) > > at $line30.$eval$.$print(<console>:6) > > at $line30.$eval.$print(<console>) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:57) > > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:606) > > at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:786) > > at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1047) > > at scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$ > loadAndRunReq$1.apply(IMain.scala:638) > > at scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$ > loadAndRunReq$1.apply(IMain.scala:637) > > at scala.reflect.internal.util.ScalaClassLoader$class. > asContext(ScalaClassLoader.scala:31) > > at scala.reflect.internal.util.AbstractFileClassLoader.asContext( > AbstractFileClassLoader.scala:19) > > at scala.tools.nsc.interpreter.IMain$WrappedRequest. > loadAndRunReq(IMain.scala:637) > > at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:569) > > at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:565) > > at scala.tools.nsc.interpreter.ILoop.interpretStartingWith( > ILoop.scala:807) > > at scala.tools.nsc.interpreter.ILoop.command(ILoop.scala:681) > > at scala.tools.nsc.interpreter.ILoop.processLine(ILoop.scala:395) > > at scala.tools.nsc.interpreter.ILoop.loop(ILoop.scala:415) > > at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1. > apply$mcZ$sp(ILoop.scala:923) > > at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1. > apply(ILoop.scala:909) > > at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1. > apply(ILoop.scala:909) > > at scala.reflect.internal.util.ScalaClassLoader$.savingContextLoader( > ScalaClassLoader.scala:97) > > at scala.tools.nsc.interpreter.ILoop.process(ILoop.scala:909) > > at org.apache.spark.repl.Main$.doMain(Main.scala:69) > > at org.apache.spark.repl.Main$.main(Main.scala:52) > > at org.apache.spark.repl.Main.main(Main.scala) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:57) > > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:606) > > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > deploy$SparkSubmit$$runMain(SparkSubmit.scala:738) > > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187) > > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212) > > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126) > > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > 17/08/21 15:59:02 AUDIT command.LoadTable: > [scsp00267.saicdt.com][carbon2][Thread-1]Dataload > failure for default.crx5_tbox_normal_test. Please check the logs > > 17/08/21 15:59:02 WARN spark.ExecutorAllocationManager: No stages are > running, but numRunningTasks != 0 > > java.lang.Exception: generate global dictionary failed, null > > at org.apache.carbondata.spark.util.GlobalDictionaryUtil$. > generateGlobalDictionary(GlobalDictionaryUtil.scala:761) > > at org.apache.spark.sql.execution.command.LoadTable. > run(carbonTableSchema.scala:703) > > at org.apache.spark.sql.execution.command.LoadTableByInsert.run( > carbonTableSchema.scala:338) > > at org.apache.spark.sql.execution.command.ExecutedCommandExec. > sideEffectResult$lzycompute(commands.scala:58) > > at org.apache.spark.sql.execution.command.ExecutedCommandExec. > sideEffectResult(commands.scala:56) > > at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute( > commands.scala:74) > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$ > execute$1.apply(SparkPlan.scala:114) > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$ > execute$1.apply(SparkPlan.scala:114) > > at org.apache.spark.sql.execution.SparkPlan$$anonfun$ > executeQuery$1.apply(SparkPlan.scala:135) > > at org.apache.spark.rdd.RDDOperationScope$.withScope( > RDDOperationScope.scala:151) > > at org.apache.spark.sql.execution.SparkPlan. > executeQuery(SparkPlan.scala:132) > > at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113) > > at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute( > QueryExecution.scala:87) > > at org.apache.spark.sql.execution.QueryExecution. > toRdd(QueryExecution.scala:87) > > at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185) > > at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64) > > at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592) > > ... 50 elided > > > > Thanks, > > CaoLu >