[
https://issues.apache.org/jira/browse/CARBONDATA-3853?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Chetan Bhat updated CARBONDATA-3853:
------------------------------------
Description:
Steps and Issue
0: jdbc:hive2://10.20.255.171:23040/> create table if not exists
all_data_types1(bool_1 boolean,bool_2 boolean,chinese string,Number
int,smallNumber smallint,BigNumber bigint,LargeDecimal double,smalldecimal
float,customdecimal decimal(38,15),words string,smallwords char(8),varwords
varchar(20),time timestamp,day date,emptyNumber int,emptysmallNumber
smallint,emptyBigNumber bigint,emptyLargeDecimal double,emptysmalldecimal
float,emptycustomdecimal decimal(38,38),emptywords string,emptysmallwords
char(8),emptyvarwords varchar(20)) stored as carbondata TBLPROPERTIES
('BUCKET_NUMBER'='2', *'BUCKET_COLUMNS'='day'*);
+----------+-+
|Result|
+----------+-+
+----------+-+
No rows selected (0.494 seconds)
0: jdbc:hive2://10.20.255.171:23040/> LOAD DATA INPATH
'hdfs://hacluster/chetan/datafile_0.csv' into table all_data_types1
OPTIONS('DELIMITER'=',' ,
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='bool_1 ,bool_2
,chinese ,Number ,smallNumber ,BigNumber ,LargeDecimal ,smalldecimal
,customdecimal,words ,smallwords ,varwords ,time ,day ,emptyNumber
,emptysmallNumber ,emptyBigNumber ,emptyLargeDecimal
,emptysmalldecimal,emptycustomdecimal ,emptywords ,emptysmallwords
,emptyvarwords');
*Error: java.lang.Exception: DataLoad failure (state=,code=0)*
*Log-*
java.lang.Exception: DataLoad failure
at
org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:560)
at
org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:207)
at
org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:168)
at
org.apache.spark.sql.execution.command.AtomicRunnableCommand$$anonfun$run$3.apply(package.scala:148)
at
org.apache.spark.sql.execution.command.AtomicRunnableCommand$$anonfun$run$3.apply(package.scala:145)
at
org.apache.spark.sql.execution.command.Auditable$class.runWithAudit(package.scala:104)
at
org.apache.spark.sql.execution.command.AtomicRunnableCommand.runWithAudit(package.scala:141)
at
org.apache.spark.sql.execution.command.AtomicRunnableCommand.run(package.scala:145)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79)
at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190)
at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190)
at org.apache.spark.sql.Dataset$$anonfun$52.apply(Dataset.scala:3259)
at
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:77)
at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3258)
at org.apache.spark.sql.Dataset.<init>(Dataset.scala:190)
at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:75)
at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:694)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:232)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
2020-06-11 23:47:24,973 | ERROR | [HiveServer2-Background-Pool: Thread-104] |
Error running hive query: |
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:179)
org.apache.hive.service.cli.HiveSQLException: java.lang.Exception: DataLoad
failure
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:269)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
was:
Steps and Issue
0: jdbc:hive2://10.20.255.171:23040/> create table if not exists
all_data_types1(bool_1 boolean,bool_2 boolean,chinese string,Number
int,smallNumber smallint,BigNumber bigint,LargeDecimal double,smalldecimal
float,customdecimal decimal(38,15),words string,smallwords char(8),varwords
varchar(20),time timestamp,day date,emptyNumber int,emptysmallNumber
smallint,emptyBigNumber bigint,emptyLargeDecimal double,emptysmalldecimal
float,emptycustomdecimal decimal(38,38),emptywords string,emptysmallwords
char(8),emptyvarwords varchar(20)) stored as carbondata TBLPROPERTIES
('BUCKET_NUMBER'='2', *'BUCKET_COLUMNS'='day'*);
+---------+--+
| Result |
+---------+--+
+---------+--+
No rows selected (0.494 seconds)
0: jdbc:hive2://10.20.255.171:23040/> LOAD DATA INPATH
'hdfs://hacluster/chetan/datafile_0.csv' into table all_data_types1
OPTIONS('DELIMITER'=',' ,
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='bool_1 ,bool_2
,chinese ,Number ,smallNumber ,BigNumber ,LargeDecimal ,smalldecimal
,customdecimal,words ,smallwords ,varwords ,time ,day ,emptyNumber
,emptysmallNumber ,emptyBigNumber ,emptyLargeDecimal
,emptysmalldecimal,emptycustomdecimal ,emptywords ,emptysmallwords
,emptyvarwords');
*Error: java.lang.Exception: DataLoad failure (state=,code=0)*
*Log-*
> Dataload fails for date column configured as BUCKET_COLUMNS
> -----------------------------------------------------------
>
> Key: CARBONDATA-3853
> URL: https://issues.apache.org/jira/browse/CARBONDATA-3853
> Project: CarbonData
> Issue Type: Bug
> Components: data-load
> Affects Versions: 2.0.0
> Reporter: Chetan Bhat
> Priority: Major
>
> Steps and Issue
> 0: jdbc:hive2://10.20.255.171:23040/> create table if not exists
> all_data_types1(bool_1 boolean,bool_2 boolean,chinese string,Number
> int,smallNumber smallint,BigNumber bigint,LargeDecimal double,smalldecimal
> float,customdecimal decimal(38,15),words string,smallwords char(8),varwords
> varchar(20),time timestamp,day date,emptyNumber int,emptysmallNumber
> smallint,emptyBigNumber bigint,emptyLargeDecimal double,emptysmalldecimal
> float,emptycustomdecimal decimal(38,38),emptywords string,emptysmallwords
> char(8),emptyvarwords varchar(20)) stored as carbondata TBLPROPERTIES
> ('BUCKET_NUMBER'='2', *'BUCKET_COLUMNS'='day'*);
> +----------+-+
> |Result|
> +----------+-+
> +----------+-+
> No rows selected (0.494 seconds)
> 0: jdbc:hive2://10.20.255.171:23040/> LOAD DATA INPATH
> 'hdfs://hacluster/chetan/datafile_0.csv' into table all_data_types1
> OPTIONS('DELIMITER'=',' ,
> 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='bool_1 ,bool_2
> ,chinese ,Number ,smallNumber ,BigNumber ,LargeDecimal ,smalldecimal
> ,customdecimal,words ,smallwords ,varwords ,time ,day ,emptyNumber
> ,emptysmallNumber ,emptyBigNumber ,emptyLargeDecimal
> ,emptysmalldecimal,emptycustomdecimal ,emptywords ,emptysmallwords
> ,emptyvarwords');
> *Error: java.lang.Exception: DataLoad failure (state=,code=0)*
>
> *Log-*
> java.lang.Exception: DataLoad failure
> at
> org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:560)
> at
> org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:207)
> at
> org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:168)
> at
> org.apache.spark.sql.execution.command.AtomicRunnableCommand$$anonfun$run$3.apply(package.scala:148)
> at
> org.apache.spark.sql.execution.command.AtomicRunnableCommand$$anonfun$run$3.apply(package.scala:145)
> at
> org.apache.spark.sql.execution.command.Auditable$class.runWithAudit(package.scala:104)
> at
> org.apache.spark.sql.execution.command.AtomicRunnableCommand.runWithAudit(package.scala:141)
> at
> org.apache.spark.sql.execution.command.AtomicRunnableCommand.run(package.scala:145)
> at
> org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
> at
> org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
> at
> org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79)
> at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190)
> at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190)
> at org.apache.spark.sql.Dataset$$anonfun$52.apply(Dataset.scala:3259)
> at
> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:77)
> at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3258)
> at org.apache.spark.sql.Dataset.<init>(Dataset.scala:190)
> at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:75)
> at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
> at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:694)
> at
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:232)
> at
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175)
> at
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
> at
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185)
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> 2020-06-11 23:47:24,973 | ERROR | [HiveServer2-Background-Pool: Thread-104] |
> Error running hive query: |
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:179)
> org.apache.hive.service.cli.HiveSQLException: java.lang.Exception: DataLoad
> failure
> at
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:269)
> at
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175)
> at
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
> at
> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185)
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
--
This message was sent by Atlassian Jira
(v8.3.4#803005)