[
https://issues.apache.org/jira/browse/CARBONDATA-82?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15388939#comment-15388939
]
Ravindra Pesala commented on CARBONDATA-82:
-------------------------------------------
[~xiaoyesoso] Actually you used numeric datatypes for all columns, so
carbondata treats all columns as measure but no dimension. That is why it is
thrown error. It is an issue as it supposed throw error at the time of
creating the table it self.
Carbon requires atleast one dimension to work, so for this type scenario please
use either DICTIONARY_INCLUDE or DICTIONARY_EXCLUDE for atleast one column to
include that column as dimension.
You can use as follows.
{code}
create table if not exists store_sales
(
ss_sold_date_sk int,
ss_sold_time_sk int,
ss_item_sk int,
ss_customer_sk int,
ss_cdemo_sk int,
ss_hdemo_sk int,
ss_addr_sk int,
ss_store_sk int,
ss_promo_sk int,
ss_ticket_number int,
ss_quantity int,
ss_wholesale_cost double,
ss_list_price double,
ss_sales_price double,
ss_ext_discount_amt double,
ss_ext_sales_price double,
ss_ext_wholesale_cost double,
ss_ext_list_price double,
ss_ext_tax double,
ss_coupon_amt double,
ss_net_paid double,
ss_net_paid_inc_tax double,
ss_net_profit double
)
STORED BY 'org.apache.carbondata.format'
TBLPROPERTIES ('DICTIONARY_INCLUDE'='ss_sold_date_sk')
{code}
> NullPointerException by
> ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> ---------------------------------------------------------------------------------------------
>
> Key: CARBONDATA-82
> URL: https://issues.apache.org/jira/browse/CARBONDATA-82
> Project: CarbonData
> Issue Type: Bug
> Components: carbon-spark
> Affects Versions: Apache CarbonData 0.2.0-incubating
> Reporter: Shoujie Zhuo
> Fix For: Apache CarbonData 0.2.0-incubating
>
>
> csv file:
> ss_sold_date_sk,ss_sold_time_sk,ss_item_sk,ss_customer_sk,ss_cdemo_sk,ss_hdemo_sk,ss_addr_sk,ss_store_sk,ss_promo_sk,ss_ticket_number,ss_quantity,ss_wholesale_cost,ss_list_price,ss_sales_price,ss_ext_discount_amt,ss_ext_sales_price,ss_ext_wholesale_cost,ss_ext_list_price,ss_ext_tax,ss_coupon_amt,ss_net_paid,ss_net_paid_inc_tax,ss_net_profit
> 2451813,65495,7649,79006,591617,3428,44839,10,5,1,79,11.41,18.71,2.80,99.54,221.20,901.39,1478.09,6.08,99.54,121.66,127.74,-779.73
> DDL:
> create table if not exists store_sales
> (
> ss_sold_date_sk int,
> ss_sold_time_sk int,
> ss_item_sk int,
> ss_customer_sk int,
> ss_cdemo_sk int,
> ss_hdemo_sk int,
> ss_addr_sk int,
> ss_store_sk int,
> ss_promo_sk int,
> ss_ticket_number int,
> ss_quantity int,
> ss_wholesale_cost double,
> ss_list_price double,
> ss_sales_price double,
> ss_ext_discount_amt double,
> ss_ext_sales_price double,
> ss_ext_wholesale_cost double,
> ss_ext_list_price double,
> ss_ext_tax double,
> ss_coupon_amt double,
> ss_net_paid double,
> ss_net_paid_inc_tax double,
> ss_net_profit double
> )
> STORED BY 'org.apache.carbondata.format';
> Log:
> > LOAD DATA inpath
> > 'hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales' INTO table
> > store_sales;
> INFO 20-07 13:43:39,249 - main Query [LOAD DATA INPATH
> 'HDFS://HOLODESK01/USER/CARBON-SPARK-SQL/TPCDS/2/STORE_SALES' INTO TABLE
> STORE_SALES]
> INFO 20-07 13:43:39,307 - Successfully able to get the table metadata file
> lock
> INFO 20-07 13:43:39,324 - main Initiating Direct Load for the Table :
> (tpcds_carbon_2.store_sales)
> INFO 20-07 13:43:39,331 - [Block Distribution]
> INFO 20-07 13:43:39,332 - totalInputSpaceConsumed : 778266079 ,
> defaultParallelism : 24
> INFO 20-07 13:43:39,332 - mapreduce.input.fileinputformat.split.maxsize :
> 32427753
> INFO 20-07 13:43:39,392 - Block broadcast_8 stored as values in memory
> (estimated size 264.0 KB, free 573.6 KB)
> INFO 20-07 13:43:39,465 - Block broadcast_8_piece0 stored as bytes in memory
> (estimated size 23.9 KB, free 597.4 KB)
> INFO 20-07 13:43:39,467 - Added broadcast_8_piece0 in memory on
> localhost:50762 (size: 23.9 KB, free: 511.4 MB)
> INFO 20-07 13:43:39,468 - Created broadcast 8 from NewHadoopRDD at
> CarbonTextFile.scala:45
> INFO 20-07 13:43:39,478 - Total input paths to process : 1
> INFO 20-07 13:43:39,493 - Starting job: take at CarbonCsvRelation.scala:175
> INFO 20-07 13:43:39,494 - Got job 5 (take at CarbonCsvRelation.scala:175)
> with 1 output partitions
> INFO 20-07 13:43:39,494 - Final stage: ResultStage 6 (take at
> CarbonCsvRelation.scala:175)
> INFO 20-07 13:43:39,494 - Parents of final stage: List()
> INFO 20-07 13:43:39,495 - Missing parents: List()
> INFO 20-07 13:43:39,496 - Submitting ResultStage 6 (MapPartitionsRDD[23] at
> map at CarbonTextFile.scala:55), which has no missing parents
> INFO 20-07 13:43:39,499 - Block broadcast_9 stored as values in memory
> (estimated size 2.6 KB, free 600.0 KB)
> INFO 20-07 13:43:39,511 - Block broadcast_9_piece0 stored as bytes in memory
> (estimated size 1600.0 B, free 601.5 KB)
> INFO 20-07 13:43:39,512 - Added broadcast_9_piece0 in memory on
> localhost:50762 (size: 1600.0 B, free: 511.4 MB)
> INFO 20-07 13:43:39,513 - Created broadcast 9 from broadcast at
> DAGScheduler.scala:1006
> INFO 20-07 13:43:39,514 - Submitting 1 missing tasks from ResultStage 6
> (MapPartitionsRDD[23] at map at CarbonTextFile.scala:55)
> INFO 20-07 13:43:39,514 - Adding task set 6.0 with 1 tasks
> INFO 20-07 13:43:39,517 - Starting task 0.0 in stage 6.0 (TID 9, localhost,
> partition 0,ANY, 2302 bytes)
> INFO 20-07 13:43:39,518 - Running task 0.0 in stage 6.0 (TID 9)
> INFO 20-07 13:43:39,523 - Input split:
> hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales/data-m-00001.csv:0+32427753
> INFO 20-07 13:43:39,545 - Finished task 0.0 in stage 6.0 (TID 9). 3580 bytes
> result sent to driver
> INFO 20-07 13:43:39,558 - Finished task 0.0 in stage 6.0 (TID 9) in 42 ms on
> localhost (1/1)
> INFO 20-07 13:43:39,558 - ResultStage 6 (take at
> CarbonCsvRelation.scala:175) finished in 0.042 s
> INFO 20-07 13:43:39,558 - Removed TaskSet 6.0, whose tasks have all
> completed, from pool
> INFO 20-07 13:43:39,558 - Job 5 finished: take at
> CarbonCsvRelation.scala:175, took 0.065209 s
> INFO 20-07 13:43:39,558 - Finished stage:
> org.apache.spark.scheduler.StageInfo@6c7379d3
> INFO 20-07 13:43:39,561 - task runtime:(count: 1, mean: 42.000000, stdev:
> 0.000000, max: 42.000000, min: 42.000000)
> INFO 20-07 13:43:39,561 - 0% 5% 10% 25% 50% 75%
> 90% 95% 100%
> INFO 20-07 13:43:39,561 - 42.0 ms 42.0 ms 42.0 ms 42.0 ms 42.0 ms 42.0 ms
> 42.0 ms 42.0 ms 42.0 ms
> INFO 20-07 13:43:39,563 - task result size:(count: 1, mean: 3580.000000,
> stdev: 0.000000, max: 3580.000000, min: 3580.000000)
> INFO 20-07 13:43:39,563 - 0% 5% 10% 25% 50% 75%
> 90% 95% 100%
> INFO 20-07 13:43:39,563 - 3.5 KB 3.5 KB 3.5 KB 3.5 KB 3.5 KB 3.5 KB
> 3.5 KB 3.5 KB 3.5 KB
> INFO 20-07 13:43:39,564 - have no column need to generate global dictionary
> AUDIT 20-07 13:43:39,564 - [holodesk01][hdfs][Thread-1]Data load request has
> been received for table tpcds_carbon_2.store_sales
> INFO 20-07 13:43:39,565 - executor (non-fetch) time pct: (count: 1, mean:
> 26.190476, stdev: 0.000000, max: 26.190476, min: 26.190476)
> INFO 20-07 13:43:39,565 - 0% 5% 10% 25% 50% 75%
> 90% 95% 100%
> INFO 20-07 13:43:39,565 - 26 % 26 % 26 % 26 % 26 % 26 %
> 26 % 26 % 26 %
> INFO 20-07 13:43:39,567 - other time pct: (count: 1, mean: 73.809524, stdev:
> 0.000000, max: 73.809524, min: 73.809524)
> INFO 20-07 13:43:39,567 - 0% 5% 10% 25% 50% 75%
> 90% 95% 100%
> INFO 20-07 13:43:39,568 - 74 % 74 % 74 % 74 % 74 % 74 %
> 74 % 74 % 74 %
> INFO 20-07 13:43:39,582 - main compaction need status is false
> INFO 20-07 13:43:39,583 - [Block Distribution]
> INFO 20-07 13:43:39,584 - totalInputSpaceConsumed : 778266079 ,
> defaultParallelism : 24
> INFO 20-07 13:43:39,584 - mapreduce.input.fileinputformat.split.maxsize :
> 32427753
> INFO 20-07 13:43:39,586 - Total input paths to process : 1
> INFO 20-07 13:43:39,599 - Total no of blocks : 24, No.of Nodes : 4
> INFO 20-07 13:43:39,599 - #Node: holodesk02 no.of.blocks: 6
> #Node: holodesk01 no.of.blocks: 6
> #Node: holodesk04 no.of.blocks: 6
> #Node: holodesk03 no.of.blocks: 6
> INFO 20-07 13:43:40,605 - Starting job: collect at
> CarbonDataRDDFactory.scala:717
> INFO 20-07 13:43:40,606 - Got job 6 (collect at
> CarbonDataRDDFactory.scala:717) with 4 output partitions
> INFO 20-07 13:43:40,606 - Final stage: ResultStage 7 (collect at
> CarbonDataRDDFactory.scala:717)
> INFO 20-07 13:43:40,607 - Parents of final stage: List()
> INFO 20-07 13:43:40,607 - Missing parents: List()
> INFO 20-07 13:43:40,607 - Submitting ResultStage 7 (CarbonDataLoadRDD[24] at
> RDD at CarbonDataLoadRDD.scala:94), which has no missing parents
> INFO 20-07 13:43:40,608 - Prefered Location for split : holodesk02
> INFO 20-07 13:43:40,608 - Prefered Location for split : holodesk01
> INFO 20-07 13:43:40,608 - Prefered Location for split : holodesk04
> INFO 20-07 13:43:40,608 - Prefered Location for split : holodesk03
> INFO 20-07 13:43:40,613 - Block broadcast_10 stored as values in memory
> (estimated size 15.8 KB, free 617.3 KB)
> INFO 20-07 13:43:40,625 - Block broadcast_10_piece0 stored as bytes in
> memory (estimated size 5.9 KB, free 623.2 KB)
> INFO 20-07 13:43:40,627 - Added broadcast_10_piece0 in memory on
> localhost:50762 (size: 5.9 KB, free: 511.4 MB)
> INFO 20-07 13:43:40,627 - Created broadcast 10 from broadcast at
> DAGScheduler.scala:1006
> INFO 20-07 13:43:40,628 - Submitting 4 missing tasks from ResultStage 7
> (CarbonDataLoadRDD[24] at RDD at CarbonDataLoadRDD.scala:94)
> INFO 20-07 13:43:40,628 - Adding task set 7.0 with 4 tasks
> INFO 20-07 13:43:40,631 - Starting task 0.0 in stage 7.0 (TID 10, localhost,
> partition 0,ANY, 2892 bytes)
> INFO 20-07 13:43:40,632 - Starting task 1.0 in stage 7.0 (TID 11, localhost,
> partition 1,ANY, 2892 bytes)
> INFO 20-07 13:43:40,633 - Starting task 2.0 in stage 7.0 (TID 12, localhost,
> partition 2,ANY, 2892 bytes)
> INFO 20-07 13:43:40,634 - Starting task 3.0 in stage 7.0 (TID 13, localhost,
> partition 3,ANY, 2892 bytes)
> INFO 20-07 13:43:40,634 - Running task 0.0 in stage 7.0 (TID 10)
> INFO 20-07 13:43:40,635 - Running task 1.0 in stage 7.0 (TID 11)
> INFO 20-07 13:43:40,635 - Running task 2.0 in stage 7.0 (TID 12)
> INFO 20-07 13:43:40,635 - Running task 3.0 in stage 7.0 (TID 13)
> INFO 20-07 13:43:40,648 - Input split: holodesk04
> INFO 20-07 13:43:40,648 - The Block Count in this node :6
> INFO 20-07 13:43:40,649 - Input split: holodesk01
> INFO 20-07 13:43:40,649 - The Block Count in this node :6
> INFO 20-07 13:43:40,649 - [Executor task launch
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
> ************* Is Columnar Storagetrue
> INFO 20-07 13:43:40,649 - [Executor task launch
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
> ************* Is Columnar Storagetrue
> INFO 20-07 13:43:40,649 - Input split: holodesk03
> INFO 20-07 13:43:40,650 - The Block Count in this node :6
> INFO 20-07 13:43:40,650 - [Executor task launch
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
> ************* Is Columnar Storagetrue
> INFO 20-07 13:43:40,649 - Input split: holodesk02
> INFO 20-07 13:43:40,651 - The Block Count in this node :6
> INFO 20-07 13:43:40,651 - [Executor task launch
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
> ************* Is Columnar Storagetrue
> INFO 20-07 13:43:40,701 - [Executor task launch
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
> Kettle environment initialized
> INFO 20-07 13:43:40,706 - [Executor task launch
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
> Kettle environment initialized
> INFO 20-07 13:43:40,707 - [Executor task launch
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
> Kettle environment initialized
> INFO 20-07 13:43:40,713 - [Executor task launch
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
> Kettle environment initialized
> INFO 20-07 13:43:40,751 - [Executor task launch
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
> ** Using csv file **
> INFO 20-07 13:43:40,756 - [Executor task launch
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
> ** Using csv file **
> INFO 20-07 13:43:40,764 - store_sales: Graph - CSV Input
> *****************Started all csv reading***********
> INFO 20-07 13:43:40,774 -
> [pool-40-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-1]
> *****************started csv reading by thread***********
> INFO 20-07 13:43:40,788 -
> [pool-40-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-2]
> *****************started csv reading by thread***********
> INFO 20-07 13:43:40,795 - [Executor task launch
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
> Graph execution is started
> /mnt/disk1/spark/438978154880668/3/etl/tpcds_carbon_2/store_sales/0/3/store_sales.ktr
> INFO 20-07 13:43:40,798 - store_sales: Graph - CSV Input
> *****************Started all csv reading***********
> INFO 20-07 13:43:40,809 - [Executor task launch
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
> Graph execution is started
> /mnt/disk1/spark/438978153902729/1/etl/tpcds_carbon_2/store_sales/0/1/store_sales.ktr
> INFO 20-07 13:43:40,813 -
> [pool-41-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-1]
> *****************started csv reading by thread***********
> INFO 20-07 13:43:40,814 -
> [pool-41-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-2]
> *****************started csv reading by thread***********
> ERROR 20-07 13:43:40,819 - [store_sales: Graph - Carbon Surrogate Key
> Generator][partitionID:0]
> java.lang.NullPointerException
> at
> org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> at
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> at
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> at java.lang.Thread.run(Thread.java:745)
> INFO 20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO 20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Number of Records was Zero
> INFO 20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0:
> Write: 0
> INFO 20-07 13:43:40,820 - [store_sales: Graph - Carbon Slice
> Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> INFO 20-07 13:43:40,820 - [store_sales: Graph - Carbon Slice
> Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step:
> Read: 0: Write: 0
> INFO 20-07 13:43:40,820 - [Executor task launch
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
> ** Using csv file **
> ERROR 20-07 13:43:40,821 - [store_sales: Graph -
> MDKeyGenstore_sales][partitionID:0] Local data load folder location does not
> exist:
> /mnt/disk1/spark/438978154880668/3/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/3
> INFO 20-07 13:43:40,841 - [Executor task launch
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
> ** Using csv file **
> INFO 20-07 13:43:40,854 - [Executor task launch
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
> Graph execution is started
> /mnt/disk2/spark/438978155737218/0/etl/tpcds_carbon_2/store_sales/0/0/store_sales.ktr
> ERROR 20-07 13:43:40,854 - [store_sales: Graph - Carbon Surrogate Key
> Generator][partitionID:0]
> java.lang.NullPointerException
> at
> org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> at
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> at
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> at java.lang.Thread.run(Thread.java:745)
> ERROR 20-07 13:43:40,855 - [store_sales: Graph -
> MDKeyGenstore_sales][partitionID:0] Local data load folder location does not
> exist:
> /mnt/disk1/spark/438978153902729/1/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/1
> INFO 20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO 20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Number of Records was Zero
> INFO 20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0:
> Write: 0
> INFO 20-07 13:43:40,856 - store_sales: Graph - CSV Input
> *****************Started all csv reading***********
> INFO 20-07 13:43:40,857 - [store_sales: Graph - Carbon Slice
> Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> INFO 20-07 13:43:40,857 - [store_sales: Graph - Carbon Slice
> Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step:
> Read: 0: Write: 0
> INFO 20-07 13:43:40,867 -
> [pool-42-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-2]
> *****************started csv reading by thread***********
> INFO 20-07 13:43:40,869 -
> [pool-42-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-1]
> *****************started csv reading by thread***********
> INFO 20-07 13:43:40,872 - store_sales: Graph - CSV Input
> *****************Started all csv reading***********
> INFO 20-07 13:43:40,878 -
> [pool-43-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-1]
> *****************started csv reading by thread***********
> INFO 20-07 13:43:40,881 -
> [pool-43-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-2]
> *****************started csv reading by thread***********
> INFO 20-07 13:43:40,886 - [Executor task launch
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
> Graph execution is started
> /mnt/disk1/spark/438978153678637/2/etl/tpcds_carbon_2/store_sales/0/2/store_sales.ktr
> ERROR 20-07 13:43:40,898 - [store_sales: Graph - Carbon Surrogate Key
> Generator][partitionID:0]
> java.lang.NullPointerException
> at
> org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> at
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> at
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> at java.lang.Thread.run(Thread.java:745)
> INFO 20-07 13:43:40,899 - [store_sales: Graph - Carbon Slice
> Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> ERROR 20-07 13:43:40,899 - [store_sales: Graph -
> MDKeyGenstore_sales][partitionID:0] Local data load folder location does not
> exist:
> /mnt/disk2/spark/438978155737218/0/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/0
> INFO 20-07 13:43:40,899 - [store_sales: Graph - Carbon Slice
> Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step:
> Read: 0: Write: 0
> INFO 20-07 13:43:40,899 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO 20-07 13:43:40,899 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Number of Records was Zero
> INFO 20-07 13:43:40,900 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0:
> Write: 0
> ERROR 20-07 13:43:40,904 - [store_sales: Graph - Carbon Surrogate Key
> Generator][partitionID:0]
> java.lang.NullPointerException
> at
> org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> at
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> at
> org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> at java.lang.Thread.run(Thread.java:745)
> INFO 20-07 13:43:40,906 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO 20-07 13:43:40,906 - [store_sales: Graph - Carbon Slice
> Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> ERROR 20-07 13:43:40,907 - [store_sales: Graph -
> MDKeyGenstore_sales][partitionID:0] Local data load folder location does not
> exist:
> /mnt/disk1/spark/438978153678637/2/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/2
> INFO 20-07 13:43:40,907 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Number of Records was Zero
> INFO 20-07 13:43:40,907 - [store_sales: Graph - Carbon Slice
> Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step:
> Read: 0: Write: 0
> INFO 20-07 13:43:40,907 - [store_sales: Graph - Sort Key: Sort
> keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0:
> Write: 0
> INFO 20-07 13:43:41,464 - Cleaned accumulator 18
> INFO 20-07 13:43:41,492 - Removed broadcast_8_piece0 on localhost:50762 in
> memory (size: 23.9 KB, free: 511.5 MB)
> INFO 20-07 13:43:41,497 - Removed broadcast_7_piece0 on localhost:50762 in
> memory (size: 23.9 KB, free: 511.5 MB)
> INFO 20-07 13:43:41,499 - Removed broadcast_9_piece0 on localhost:50762 in
> memory (size: 1600.0 B, free: 511.5 MB)
> INFO 20-07 13:43:49,599 -
> [pool-41-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-2]
> *****************Completed csv reading by thread***********
> INFO 20-07 13:43:49,855 -
> [pool-41-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-1]
> *****************Completed csv reading by thread***********
> INFO 20-07 13:43:49,957 - store_sales: Graph - CSV Input
> *****************Completed all csv reading***********
> INFO 20-07 13:43:49,957 - [Executor task launch
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
> Graph execution is finished.
> ERROR 20-07 13:43:49,957 - [Executor task launch
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
> Graph Execution had errors
> ERROR 20-07 13:43:49,957 - [Executor task launch
> worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c]
>
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> at
> org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> at
> org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> at
> org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> at
> org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> at
> org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> INFO 20-07 13:43:49,958 - DataLoad failure
> INFO 20-07 13:43:49,969 - Finished task 1.0 in stage 7.0 (TID 11). 952 bytes
> result sent to driver
> INFO 20-07 13:43:49,982 - Finished task 1.0 in stage 7.0 (TID 11) in 9350 ms
> on localhost (1/4)
> INFO 20-07 13:43:50,482 -
> [pool-40-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-2]
> *****************Completed csv reading by thread***********
> INFO 20-07 13:43:50,943 -
> [pool-42-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-2]
> *****************Completed csv reading by thread***********
> INFO 20-07 13:43:51,270 -
> [pool-40-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-1]
> *****************Completed csv reading by thread***********
> INFO 20-07 13:43:51,408 - store_sales: Graph - CSV Input
> *****************Completed all csv reading***********
> INFO 20-07 13:43:51,408 - [Executor task launch
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
> Graph execution is finished.
> ERROR 20-07 13:43:51,409 - [Executor task launch
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
> Graph Execution had errors
> ERROR 20-07 13:43:51,409 - [Executor task launch
> worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21]
>
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> at
> org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> at
> org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> at
> org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> at
> org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> at
> org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> INFO 20-07 13:43:51,409 - DataLoad failure
> INFO 20-07 13:43:51,420 - Finished task 3.0 in stage 7.0 (TID 13). 952 bytes
> result sent to driver
> INFO 20-07 13:43:51,434 - Finished task 3.0 in stage 7.0 (TID 13) in 10800
> ms on localhost (2/4)
> INFO 20-07 13:43:51,435 -
> [pool-43-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-2]
> *****************Completed csv reading by thread***********
> INFO 20-07 13:43:52,466 -
> [pool-42-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-1]
> *****************Completed csv reading by thread***********
> INFO 20-07 13:43:52,588 - store_sales: Graph - CSV Input
> *****************Completed all csv reading***********
> INFO 20-07 13:43:52,590 - [Executor task launch
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
> Graph execution is finished.
> ERROR 20-07 13:43:52,590 - [Executor task launch
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
> Graph Execution had errors
> ERROR 20-07 13:43:52,590 - [Executor task launch
> worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605]
>
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> at
> org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> at
> org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> at
> org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> at
> org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> at
> org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> INFO 20-07 13:43:52,591 - DataLoad failure
> INFO 20-07 13:43:52,603 - Finished task 0.0 in stage 7.0 (TID 10). 952 bytes
> result sent to driver
> INFO 20-07 13:43:52,614 - Finished task 0.0 in stage 7.0 (TID 10) in 11984
> ms on localhost (3/4)
> INFO 20-07 13:43:52,638 -
> [pool-43-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-1]
> *****************Completed csv reading by thread***********
> INFO 20-07 13:43:52,824 - store_sales: Graph - CSV Input
> *****************Completed all csv reading***********
> INFO 20-07 13:43:52,824 - [Executor task launch
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
> Graph execution is finished.
> ERROR 20-07 13:43:52,825 - [Executor task launch
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
> Graph Execution had errors
> ERROR 20-07 13:43:52,825 - [Executor task launch
> worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3]
>
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> at
> org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> at
> org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> at
> org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> at
> org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> at
> org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:89)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> INFO 20-07 13:43:52,825 - DataLoad failure
> INFO 20-07 13:43:52,837 - Finished task 2.0 in stage 7.0 (TID 12). 952 bytes
> result sent to driver
> INFO 20-07 13:43:52,849 - Finished task 2.0 in stage 7.0 (TID 12) in 12216
> ms on localhost (4/4)
> INFO 20-07 13:43:52,849 - ResultStage 7 (collect at
> CarbonDataRDDFactory.scala:717) finished in 12.219 s
> INFO 20-07 13:43:52,849 - Removed TaskSet 7.0, whose tasks have all
> completed, from pool
> INFO 20-07 13:43:52,849 - Finished stage:
> org.apache.spark.scheduler.StageInfo@46ffcf8b
> INFO 20-07 13:43:52,849 - Job 6 finished: collect at
> CarbonDataRDDFactory.scala:717, took 12.244086 s
> INFO 20-07 13:43:52,850 - ********starting clean up**********
> INFO 20-07 13:43:52,851 - task runtime:(count: 4, mean: 11087.500000, stdev:
> 1137.847419, max: 12216.000000, min: 9350.000000)
> INFO 20-07 13:43:52,851 - 0% 5% 10% 25% 50% 75%
> 90% 95% 100%
> INFO 20-07 13:43:52,851 - 9.4 s 9.4 s 9.4 s 10.8 s 12.0 s 12.2 s
> 12.2 s 12.2 s 12.2 s
> INFO 20-07 13:43:52,853 - task result size:(count: 4, mean: 952.000000,
> stdev: 0.000000, max: 952.000000, min: 952.000000)
> INFO 20-07 13:43:52,853 - 0% 5% 10% 25% 50% 75%
> 90% 95% 100%
> INFO 20-07 13:43:52,853 - 952.0 B 952.0 B 952.0 B 952.0 B 952.0 B 952.0 B
> 952.0 B 952.0 B 952.0 B
> INFO 20-07 13:43:52,855 - executor (non-fetch) time pct: (count: 4, mean:
> 99.639701, stdev: 0.042276, max: 99.688933, min: 99.572193)
> INFO 20-07 13:43:52,855 - 0% 5% 10% 25% 50% 75%
> 90% 95% 100%
> INFO 20-07 13:43:52,855 - 100 % 100 % 100 % 100 % 100 % 100 %
> 100 % 100 % 100 %
> INFO 20-07 13:43:52,857 - other time pct: (count: 4, mean: 0.360299, stdev:
> 0.042276, max: 0.427807, min: 0.311067)
> INFO 20-07 13:43:52,857 - 0% 5% 10% 25% 50% 75%
> 90% 95% 100%
> INFO 20-07 13:43:52,857 - 0 % 0 % 0 % 0 % 0 % 0 %
> 0 % 0 % 0 %
> INFO 20-07 13:43:53,079 - ********clean up done**********
> AUDIT 20-07 13:43:53,079 - [holodesk01][hdfs][Thread-1]Data load is failed
> for tpcds_carbon_2.store_sales
> WARN 20-07 13:43:53,080 - Unable to write load metadata file
> ERROR 20-07 13:43:53,080 - main
> java.lang.Exception: Dataload failure
> at
> org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
> at
> org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
> at
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
> at
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
> at
> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
> at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
> at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
> at
> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
> at
> org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> at
> org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
> at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
> at
> org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
> at
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
> at
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
> at
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
> at
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> at
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
> at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> AUDIT 20-07 13:43:53,081 - [holodesk01][hdfs][Thread-1]Dataload failure for
> tpcds_carbon_2.store_sales. Please check the logs
> INFO 20-07 13:43:53,083 - Table MetaData Unlocked Successfully after data
> load
> ERROR 20-07 13:43:53,083 - Failed in [LOAD DATA inpath
> 'hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales' INTO table
> store_sales]
> java.lang.Exception: Dataload failure
> at
> org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
> at
> org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
> at
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
> at
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
> at
> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
> at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
> at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
> at
> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
> at
> org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> at
> org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
> at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
> at
> org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
> at
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
> at
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
> at
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
> at
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> at
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
> at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> java.lang.Exception: Dataload failure
> at
> org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
> at
> org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
> at
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
> at
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
> at
> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
> at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
> at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
> at
> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
> at
> org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> at
> org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
> at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
> at
> org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
> at
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
> at
> org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
> at
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
> at
> org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> at
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
> at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> ```
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)