[ 
https://issues.apache.org/jira/browse/CARBONDATA-82?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15385431#comment-15385431
 ] 

Ravindra Pesala commented on CARBONDATA-82:
-------------------------------------------

Can you provide more details on this issue. liike 'create sql command', 'sample 
csv files'. And is this issue happens only on big file or even for small file 
also happens? 

> Dataload Failure then the csv file is big
> -----------------------------------------
>
>                 Key: CARBONDATA-82
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-82
>             Project: CarbonData
>          Issue Type: Improvement
>          Components: carbon-spark
>    Affects Versions: Apache CarbonData 0.2.0-incubating
>            Reporter: Shoujie Zhuo
>             Fix For: Apache CarbonData 0.2.0-incubating
>
>
> > LOAD DATA  inpath 
> > 'hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales' INTO table 
> > store_sales;
> INFO  20-07 13:43:39,249 - main Query [LOAD DATA  INPATH 
> 'HDFS://HOLODESK01/USER/CARBON-SPARK-SQL/TPCDS/2/STORE_SALES' INTO TABLE 
> STORE_SALES]
> INFO  20-07 13:43:39,307 - Successfully able to get the table metadata file 
> lock
> INFO  20-07 13:43:39,324 - main Initiating Direct Load for the Table : 
> (tpcds_carbon_2.store_sales)
> INFO  20-07 13:43:39,331 - [Block Distribution]
> INFO  20-07 13:43:39,332 - totalInputSpaceConsumed : 778266079 , 
> defaultParallelism : 24
> INFO  20-07 13:43:39,332 - mapreduce.input.fileinputformat.split.maxsize : 
> 32427753
> INFO  20-07 13:43:39,392 - Block broadcast_8 stored as values in memory 
> (estimated size 264.0 KB, free 573.6 KB)
> INFO  20-07 13:43:39,465 - Block broadcast_8_piece0 stored as bytes in memory 
> (estimated size 23.9 KB, free 597.4 KB)
> INFO  20-07 13:43:39,467 - Added broadcast_8_piece0 in memory on 
> localhost:50762 (size: 23.9 KB, free: 511.4 MB)
> INFO  20-07 13:43:39,468 - Created broadcast 8 from NewHadoopRDD at 
> CarbonTextFile.scala:45
> INFO  20-07 13:43:39,478 - Total input paths to process : 1
> INFO  20-07 13:43:39,493 - Starting job: take at CarbonCsvRelation.scala:175
> INFO  20-07 13:43:39,494 - Got job 5 (take at CarbonCsvRelation.scala:175) 
> with 1 output partitions
> INFO  20-07 13:43:39,494 - Final stage: ResultStage 6 (take at 
> CarbonCsvRelation.scala:175)
> INFO  20-07 13:43:39,494 - Parents of final stage: List()
> INFO  20-07 13:43:39,495 - Missing parents: List()
> INFO  20-07 13:43:39,496 - Submitting ResultStage 6 (MapPartitionsRDD[23] at 
> map at CarbonTextFile.scala:55), which has no missing parents
> INFO  20-07 13:43:39,499 - Block broadcast_9 stored as values in memory 
> (estimated size 2.6 KB, free 600.0 KB)
> INFO  20-07 13:43:39,511 - Block broadcast_9_piece0 stored as bytes in memory 
> (estimated size 1600.0 B, free 601.5 KB)
> INFO  20-07 13:43:39,512 - Added broadcast_9_piece0 in memory on 
> localhost:50762 (size: 1600.0 B, free: 511.4 MB)
> INFO  20-07 13:43:39,513 - Created broadcast 9 from broadcast at 
> DAGScheduler.scala:1006
> INFO  20-07 13:43:39,514 - Submitting 1 missing tasks from ResultStage 6 
> (MapPartitionsRDD[23] at map at CarbonTextFile.scala:55)
> INFO  20-07 13:43:39,514 - Adding task set 6.0 with 1 tasks
> INFO  20-07 13:43:39,517 - Starting task 0.0 in stage 6.0 (TID 9, localhost, 
> partition 0,ANY, 2302 bytes)
> INFO  20-07 13:43:39,518 - Running task 0.0 in stage 6.0 (TID 9)
> INFO  20-07 13:43:39,523 - Input split: 
> hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales/data-m-00001.csv:0+32427753
> INFO  20-07 13:43:39,545 - Finished task 0.0 in stage 6.0 (TID 9). 3580 bytes 
> result sent to driver
> INFO  20-07 13:43:39,558 - Finished task 0.0 in stage 6.0 (TID 9) in 42 ms on 
> localhost (1/1)
> INFO  20-07 13:43:39,558 - ResultStage 6 (take at 
> CarbonCsvRelation.scala:175) finished in 0.042 s
> INFO  20-07 13:43:39,558 - Removed TaskSet 6.0, whose tasks have all 
> completed, from pool 
> INFO  20-07 13:43:39,558 - Job 5 finished: take at 
> CarbonCsvRelation.scala:175, took 0.065209 s
> INFO  20-07 13:43:39,558 - Finished stage: 
> org.apache.spark.scheduler.StageInfo@6c7379d3
> INFO  20-07 13:43:39,561 - task runtime:(count: 1, mean: 42.000000, stdev: 
> 0.000000, max: 42.000000, min: 42.000000)
> INFO  20-07 13:43:39,561 -    0%      5%      10%     25%     50%     75%     
> 90%     95%     100%
> INFO  20-07 13:43:39,561 -    42.0 ms 42.0 ms 42.0 ms 42.0 ms 42.0 ms 42.0 ms 
> 42.0 ms 42.0 ms 42.0 ms
> INFO  20-07 13:43:39,563 - task result size:(count: 1, mean: 3580.000000, 
> stdev: 0.000000, max: 3580.000000, min: 3580.000000)
> INFO  20-07 13:43:39,563 -    0%      5%      10%     25%     50%     75%     
> 90%     95%     100%
> INFO  20-07 13:43:39,563 -    3.5 KB  3.5 KB  3.5 KB  3.5 KB  3.5 KB  3.5 KB  
> 3.5 KB  3.5 KB  3.5 KB
> INFO  20-07 13:43:39,564 - have no column need to generate global dictionary
> AUDIT 20-07 13:43:39,564 - [holodesk01][hdfs][Thread-1]Data load request has 
> been received for table tpcds_carbon_2.store_sales
> INFO  20-07 13:43:39,565 - executor (non-fetch) time pct: (count: 1, mean: 
> 26.190476, stdev: 0.000000, max: 26.190476, min: 26.190476)
> INFO  20-07 13:43:39,565 -    0%      5%      10%     25%     50%     75%     
> 90%     95%     100%
> INFO  20-07 13:43:39,565 -    26 %    26 %    26 %    26 %    26 %    26 %    
> 26 %    26 %    26 %
> INFO  20-07 13:43:39,567 - other time pct: (count: 1, mean: 73.809524, stdev: 
> 0.000000, max: 73.809524, min: 73.809524)
> INFO  20-07 13:43:39,567 -    0%      5%      10%     25%     50%     75%     
> 90%     95%     100%
> INFO  20-07 13:43:39,568 -    74 %    74 %    74 %    74 %    74 %    74 %    
> 74 %    74 %    74 %
> INFO  20-07 13:43:39,582 - main compaction need status is false
> INFO  20-07 13:43:39,583 - [Block Distribution]
> INFO  20-07 13:43:39,584 - totalInputSpaceConsumed : 778266079 , 
> defaultParallelism : 24
> INFO  20-07 13:43:39,584 - mapreduce.input.fileinputformat.split.maxsize : 
> 32427753
> INFO  20-07 13:43:39,586 - Total input paths to process : 1
> INFO  20-07 13:43:39,599 - Total no of blocks : 24, No.of Nodes : 4
> INFO  20-07 13:43:39,599 - #Node: holodesk02 no.of.blocks: 6
> #Node: holodesk01 no.of.blocks: 6
> #Node: holodesk04 no.of.blocks: 6
> #Node: holodesk03 no.of.blocks: 6
> INFO  20-07 13:43:40,605 - Starting job: collect at 
> CarbonDataRDDFactory.scala:717
> INFO  20-07 13:43:40,606 - Got job 6 (collect at 
> CarbonDataRDDFactory.scala:717) with 4 output partitions
> INFO  20-07 13:43:40,606 - Final stage: ResultStage 7 (collect at 
> CarbonDataRDDFactory.scala:717)
> INFO  20-07 13:43:40,607 - Parents of final stage: List()
> INFO  20-07 13:43:40,607 - Missing parents: List()
> INFO  20-07 13:43:40,607 - Submitting ResultStage 7 (CarbonDataLoadRDD[24] at 
> RDD at CarbonDataLoadRDD.scala:94), which has no missing parents
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk02
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk01
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk04
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk03
> INFO  20-07 13:43:40,613 - Block broadcast_10 stored as values in memory 
> (estimated size 15.8 KB, free 617.3 KB)
> INFO  20-07 13:43:40,625 - Block broadcast_10_piece0 stored as bytes in 
> memory (estimated size 5.9 KB, free 623.2 KB)
> INFO  20-07 13:43:40,627 - Added broadcast_10_piece0 in memory on 
> localhost:50762 (size: 5.9 KB, free: 511.4 MB)
> INFO  20-07 13:43:40,627 - Created broadcast 10 from broadcast at 
> DAGScheduler.scala:1006
> INFO  20-07 13:43:40,628 - Submitting 4 missing tasks from ResultStage 7 
> (CarbonDataLoadRDD[24] at RDD at CarbonDataLoadRDD.scala:94)
> INFO  20-07 13:43:40,628 - Adding task set 7.0 with 4 tasks
> INFO  20-07 13:43:40,631 - Starting task 0.0 in stage 7.0 (TID 10, localhost, 
> partition 0,ANY, 2892 bytes)
> INFO  20-07 13:43:40,632 - Starting task 1.0 in stage 7.0 (TID 11, localhost, 
> partition 1,ANY, 2892 bytes)
> INFO  20-07 13:43:40,633 - Starting task 2.0 in stage 7.0 (TID 12, localhost, 
> partition 2,ANY, 2892 bytes)
> INFO  20-07 13:43:40,634 - Starting task 3.0 in stage 7.0 (TID 13, localhost, 
> partition 3,ANY, 2892 bytes)
> INFO  20-07 13:43:40,634 - Running task 0.0 in stage 7.0 (TID 10)
> INFO  20-07 13:43:40,635 - Running task 1.0 in stage 7.0 (TID 11)
> INFO  20-07 13:43:40,635 - Running task 2.0 in stage 7.0 (TID 12)
> INFO  20-07 13:43:40,635 - Running task 3.0 in stage 7.0 (TID 13)
> INFO  20-07 13:43:40,648 - Input split: holodesk04
> INFO  20-07 13:43:40,648 - The Block Count in this node :6
> INFO  20-07 13:43:40,649 - Input split: holodesk01
> INFO  20-07 13:43:40,649 - The Block Count in this node :6
> INFO  20-07 13:43:40,649 - [Executor task launch 
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
>  ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,649 - [Executor task launch 
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
>  ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,649 - Input split: holodesk03
> INFO  20-07 13:43:40,650 - The Block Count in this node :6
> INFO  20-07 13:43:40,650 - [Executor task launch 
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
>  ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,649 - Input split: holodesk02
> INFO  20-07 13:43:40,651 - The Block Count in this node :6
> INFO  20-07 13:43:40,651 - [Executor task launch 
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
>  ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,701 - [Executor task launch 
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
>  Kettle environment initialized
> INFO  20-07 13:43:40,706 - [Executor task launch 
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
>  Kettle environment initialized
> INFO  20-07 13:43:40,707 - [Executor task launch 
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
>  Kettle environment initialized
> INFO  20-07 13:43:40,713 - [Executor task launch 
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
>  Kettle environment initialized
> INFO  20-07 13:43:40,751 - [Executor task launch 
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
>  ** Using csv file **
> INFO  20-07 13:43:40,756 - [Executor task launch 
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
>  ** Using csv file **
> INFO  20-07 13:43:40,764 - store_sales: Graph - CSV Input 
> *****************Started all csv reading***********
> INFO  20-07 13:43:40,774 - 
> [pool-40-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-1] 
> *****************started csv reading by thread***********
> INFO  20-07 13:43:40,788 - 
> [pool-40-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-2] 
> *****************started csv reading by thread***********
> INFO  20-07 13:43:40,795 - [Executor task launch 
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
>  Graph execution is started 
> /mnt/disk1/spark/438978154880668/3/etl/tpcds_carbon_2/store_sales/0/3/store_sales.ktr
> INFO  20-07 13:43:40,798 - store_sales: Graph - CSV Input 
> *****************Started all csv reading***********
> INFO  20-07 13:43:40,809 - [Executor task launch 
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
>  Graph execution is started 
> /mnt/disk1/spark/438978153902729/1/etl/tpcds_carbon_2/store_sales/0/1/store_sales.ktr
> INFO  20-07 13:43:40,813 - 
> [pool-41-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-1] 
> *****************started csv reading by thread***********
> INFO  20-07 13:43:40,814 - 
> [pool-41-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-2] 
> *****************started csv reading by thread***********
> ERROR 20-07 13:43:40,819 - [store_sales: Graph - Carbon Surrogate Key 
> Generator][partitionID:0] 
> java.lang.NullPointerException
>       at 
> org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
>       at 
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
>       at 
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
>       at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
>       at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: 
> Write: 0
> INFO  20-07 13:43:40,820 - [store_sales: Graph - Carbon Slice 
> Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> INFO  20-07 13:43:40,820 - [store_sales: Graph - Carbon Slice 
> Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: 
> Read: 0: Write: 0
> INFO  20-07 13:43:40,820 - [Executor task launch 
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
>  ** Using csv file **
> ERROR 20-07 13:43:40,821 - [store_sales: Graph - 
> MDKeyGenstore_sales][partitionID:0] Local data load folder location does not 
> exist: 
> /mnt/disk1/spark/438978154880668/3/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/3
> INFO  20-07 13:43:40,841 - [Executor task launch 
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
>  ** Using csv file **
> INFO  20-07 13:43:40,854 - [Executor task launch 
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
>  Graph execution is started 
> /mnt/disk2/spark/438978155737218/0/etl/tpcds_carbon_2/store_sales/0/0/store_sales.ktr
> ERROR 20-07 13:43:40,854 - [store_sales: Graph - Carbon Surrogate Key 
> Generator][partitionID:0] 
> java.lang.NullPointerException
>       at 
> org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
>       at 
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
>       at 
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
>       at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
>       at java.lang.Thread.run(Thread.java:745)
> ERROR 20-07 13:43:40,855 - [store_sales: Graph - 
> MDKeyGenstore_sales][partitionID:0] Local data load folder location does not 
> exist: 
> /mnt/disk1/spark/438978153902729/1/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/1
> INFO  20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: 
> Write: 0
> INFO  20-07 13:43:40,856 - store_sales: Graph - CSV Input 
> *****************Started all csv reading***********
> INFO  20-07 13:43:40,857 - [store_sales: Graph - Carbon Slice 
> Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> INFO  20-07 13:43:40,857 - [store_sales: Graph - Carbon Slice 
> Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: 
> Read: 0: Write: 0
> INFO  20-07 13:43:40,867 - 
> [pool-42-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-2] 
> *****************started csv reading by thread***********
> INFO  20-07 13:43:40,869 - 
> [pool-42-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-1] 
> *****************started csv reading by thread***********
> INFO  20-07 13:43:40,872 - store_sales: Graph - CSV Input 
> *****************Started all csv reading***********
> INFO  20-07 13:43:40,878 - 
> [pool-43-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-1] 
> *****************started csv reading by thread***********
> INFO  20-07 13:43:40,881 - 
> [pool-43-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-2] 
> *****************started csv reading by thread***********
> INFO  20-07 13:43:40,886 - [Executor task launch 
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
>  Graph execution is started 
> /mnt/disk1/spark/438978153678637/2/etl/tpcds_carbon_2/store_sales/0/2/store_sales.ktr
> ERROR 20-07 13:43:40,898 - [store_sales: Graph - Carbon Surrogate Key 
> Generator][partitionID:0] 
> java.lang.NullPointerException
>       at 
> org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
>       at 
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
>       at 
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
>       at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
>       at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Carbon Slice 
> Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> ERROR 20-07 13:43:40,899 - [store_sales: Graph - 
> MDKeyGenstore_sales][partitionID:0] Local data load folder location does not 
> exist: 
> /mnt/disk2/spark/438978155737218/0/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/0
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Carbon Slice 
> Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: 
> Read: 0: Write: 0
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,900 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: 
> Write: 0
> ERROR 20-07 13:43:40,904 - [store_sales: Graph - Carbon Surrogate Key 
> Generator][partitionID:0] 
> java.lang.NullPointerException
>       at 
> org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
>       at 
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
>       at 
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
>       at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
>       at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:40,906 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,906 - [store_sales: Graph - Carbon Slice 
> Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> ERROR 20-07 13:43:40,907 - [store_sales: Graph - 
> MDKeyGenstore_sales][partitionID:0] Local data load folder location does not 
> exist: 
> /mnt/disk1/spark/438978153678637/2/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/2
> INFO  20-07 13:43:40,907 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,907 - [store_sales: Graph - Carbon Slice 
> Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: 
> Read: 0: Write: 0
> INFO  20-07 13:43:40,907 - [store_sales: Graph - Sort Key: Sort 
> keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: 
> Write: 0
> INFO  20-07 13:43:41,464 - Cleaned accumulator 18
> INFO  20-07 13:43:41,492 - Removed broadcast_8_piece0 on localhost:50762 in 
> memory (size: 23.9 KB, free: 511.5 MB)
> INFO  20-07 13:43:41,497 - Removed broadcast_7_piece0 on localhost:50762 in 
> memory (size: 23.9 KB, free: 511.5 MB)
> INFO  20-07 13:43:41,499 - Removed broadcast_9_piece0 on localhost:50762 in 
> memory (size: 1600.0 B, free: 511.5 MB)
> INFO  20-07 13:43:49,599 - 
> [pool-41-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-2] 
> *****************Completed csv reading by thread***********
> INFO  20-07 13:43:49,855 - 
> [pool-41-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-1] 
> *****************Completed csv reading by thread***********
> INFO  20-07 13:43:49,957 - store_sales: Graph - CSV Input 
> *****************Completed all csv reading***********
> INFO  20-07 13:43:49,957 - [Executor task launch 
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
>  Graph execution is finished.
> ERROR 20-07 13:43:49,957 - [Executor task launch 
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
>  Graph Execution had errors
> ERROR 20-07 13:43:49,957 - [Executor task launch 
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
>  
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
>       at 
> org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
>       at 
> org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
>       at 
> org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
>       at 
> org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
>       at 
> org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
>       at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>       at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>       at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>       at org.apache.spark.scheduler.Task.run(Task.scala:89)
>       at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:49,958 - DataLoad failure
> INFO  20-07 13:43:49,969 - Finished task 1.0 in stage 7.0 (TID 11). 952 bytes 
> result sent to driver
> INFO  20-07 13:43:49,982 - Finished task 1.0 in stage 7.0 (TID 11) in 9350 ms 
> on localhost (1/4)
> INFO  20-07 13:43:50,482 - 
> [pool-40-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-2] 
> *****************Completed csv reading by thread***********
> INFO  20-07 13:43:50,943 - 
> [pool-42-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-2] 
> *****************Completed csv reading by thread***********
> INFO  20-07 13:43:51,270 - 
> [pool-40-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-1] 
> *****************Completed csv reading by thread***********
> INFO  20-07 13:43:51,408 - store_sales: Graph - CSV Input 
> *****************Completed all csv reading***********
> INFO  20-07 13:43:51,408 - [Executor task launch 
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
>  Graph execution is finished.
> ERROR 20-07 13:43:51,409 - [Executor task launch 
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
>  Graph Execution had errors
> ERROR 20-07 13:43:51,409 - [Executor task launch 
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
>  
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
>       at 
> org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
>       at 
> org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
>       at 
> org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
>       at 
> org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
>       at 
> org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
>       at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>       at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>       at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>       at org.apache.spark.scheduler.Task.run(Task.scala:89)
>       at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:51,409 - DataLoad failure
> INFO  20-07 13:43:51,420 - Finished task 3.0 in stage 7.0 (TID 13). 952 bytes 
> result sent to driver
> INFO  20-07 13:43:51,434 - Finished task 3.0 in stage 7.0 (TID 13) in 10800 
> ms on localhost (2/4)
> INFO  20-07 13:43:51,435 - 
> [pool-43-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-2] 
> *****************Completed csv reading by thread***********
> INFO  20-07 13:43:52,466 - 
> [pool-42-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-1] 
> *****************Completed csv reading by thread***********
> INFO  20-07 13:43:52,588 - store_sales: Graph - CSV Input 
> *****************Completed all csv reading***********
> INFO  20-07 13:43:52,590 - [Executor task launch 
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
>  Graph execution is finished.
> ERROR 20-07 13:43:52,590 - [Executor task launch 
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
>  Graph Execution had errors
> ERROR 20-07 13:43:52,590 - [Executor task launch 
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
>  
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
>       at 
> org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
>       at 
> org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
>       at 
> org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
>       at 
> org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
>       at 
> org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
>       at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>       at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>       at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>       at org.apache.spark.scheduler.Task.run(Task.scala:89)
>       at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:52,591 - DataLoad failure
> INFO  20-07 13:43:52,603 - Finished task 0.0 in stage 7.0 (TID 10). 952 bytes 
> result sent to driver
> INFO  20-07 13:43:52,614 - Finished task 0.0 in stage 7.0 (TID 10) in 11984 
> ms on localhost (3/4)
> INFO  20-07 13:43:52,638 - 
> [pool-43-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-1] 
> *****************Completed csv reading by thread***********
> INFO  20-07 13:43:52,824 - store_sales: Graph - CSV Input 
> *****************Completed all csv reading***********
> INFO  20-07 13:43:52,824 - [Executor task launch 
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
>  Graph execution is finished.
> ERROR 20-07 13:43:52,825 - [Executor task launch 
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
>  Graph Execution had errors
> ERROR 20-07 13:43:52,825 - [Executor task launch 
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
>  
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
>       at 
> org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
>       at 
> org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
>       at 
> org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
>       at 
> org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
>       at 
> org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
>       at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
>       at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
>       at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>       at org.apache.spark.scheduler.Task.run(Task.scala:89)
>       at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:52,825 - DataLoad failure
> INFO  20-07 13:43:52,837 - Finished task 2.0 in stage 7.0 (TID 12). 952 bytes 
> result sent to driver
> INFO  20-07 13:43:52,849 - Finished task 2.0 in stage 7.0 (TID 12) in 12216 
> ms on localhost (4/4)
> INFO  20-07 13:43:52,849 - ResultStage 7 (collect at 
> CarbonDataRDDFactory.scala:717) finished in 12.219 s
> INFO  20-07 13:43:52,849 - Removed TaskSet 7.0, whose tasks have all 
> completed, from pool 
> INFO  20-07 13:43:52,849 - Finished stage: 
> org.apache.spark.scheduler.StageInfo@46ffcf8b
> INFO  20-07 13:43:52,849 - Job 6 finished: collect at 
> CarbonDataRDDFactory.scala:717, took 12.244086 s
> INFO  20-07 13:43:52,850 - ********starting clean up**********
> INFO  20-07 13:43:52,851 - task runtime:(count: 4, mean: 11087.500000, stdev: 
> 1137.847419, max: 12216.000000, min: 9350.000000)
> INFO  20-07 13:43:52,851 -    0%      5%      10%     25%     50%     75%     
> 90%     95%     100%
> INFO  20-07 13:43:52,851 -    9.4 s   9.4 s   9.4 s   10.8 s  12.0 s  12.2 s  
> 12.2 s  12.2 s  12.2 s
> INFO  20-07 13:43:52,853 - task result size:(count: 4, mean: 952.000000, 
> stdev: 0.000000, max: 952.000000, min: 952.000000)
> INFO  20-07 13:43:52,853 -    0%      5%      10%     25%     50%     75%     
> 90%     95%     100%
> INFO  20-07 13:43:52,853 -    952.0 B 952.0 B 952.0 B 952.0 B 952.0 B 952.0 B 
> 952.0 B 952.0 B 952.0 B
> INFO  20-07 13:43:52,855 - executor (non-fetch) time pct: (count: 4, mean: 
> 99.639701, stdev: 0.042276, max: 99.688933, min: 99.572193)
> INFO  20-07 13:43:52,855 -    0%      5%      10%     25%     50%     75%     
> 90%     95%     100%
> INFO  20-07 13:43:52,855 -    100 %   100 %   100 %   100 %   100 %   100 %   
> 100 %   100 %   100 %
> INFO  20-07 13:43:52,857 - other time pct: (count: 4, mean: 0.360299, stdev: 
> 0.042276, max: 0.427807, min: 0.311067)
> INFO  20-07 13:43:52,857 -    0%      5%      10%     25%     50%     75%     
> 90%     95%     100%
> INFO  20-07 13:43:52,857 -     0 %     0 %     0 %     0 %     0 %     0 %    
>  0 %     0 %     0 %
> INFO  20-07 13:43:53,079 - ********clean up done**********
> AUDIT 20-07 13:43:53,079 - [holodesk01][hdfs][Thread-1]Data load is failed 
> for tpcds_carbon_2.store_sales
> WARN  20-07 13:43:53,080 - Unable to write load metadata file
> ERROR 20-07 13:43:53,080 - main 
> java.lang.Exception: Dataload failure
>       at 
> org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
>       at 
> org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
>       at 
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
>       at 
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
>       at 
> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
>       at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
>       at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
>       at 
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
>       at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
>       at 
> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
>       at 
> org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
>       at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
>       at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
>       at 
> org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
>       at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
>       at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
>       at 
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
>       at 
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:606)
>       at 
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
>       at 
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
>       at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
>       at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
>       at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> AUDIT 20-07 13:43:53,081 - [holodesk01][hdfs][Thread-1]Dataload failure for 
> tpcds_carbon_2.store_sales. Please check the logs
> INFO  20-07 13:43:53,083 - Table MetaData Unlocked Successfully after data 
> load
> ERROR 20-07 13:43:53,083 - Failed in [LOAD DATA  inpath 
> 'hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales' INTO table 
> store_sales]
> java.lang.Exception: Dataload failure
>       at 
> org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
>       at 
> org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
>       at 
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
>       at 
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
>       at 
> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
>       at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
>       at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
>       at 
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
>       at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
>       at 
> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
>       at 
> org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
>       at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
>       at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
>       at 
> org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
>       at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
>       at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
>       at 
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
>       at 
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:606)
>       at 
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
>       at 
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
>       at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
>       at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
>       at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> java.lang.Exception: Dataload failure
>       at 
> org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
>       at 
> org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
>       at 
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
>       at 
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
>       at 
> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
>       at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
>       at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
>       at 
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
>       at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
>       at 
> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
>       at 
> org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
>       at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
>       at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
>       at 
> org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
>       at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
>       at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
>       at 
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
>       at 
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
>       at 
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:606)
>       at 
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
>       at 
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
>       at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
>       at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
>       at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> ```



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to