Chetan Bhat created CARBONDATA-3846:
---------------------------------------

             Summary: Dataload fails for boolean column configured as 
BUCKET_COLUMNS
                 Key: CARBONDATA-3846
                 URL: https://issues.apache.org/jira/browse/CARBONDATA-3846
             Project: CarbonData
          Issue Type: Bug
          Components: data-query
    Affects Versions: 2.0.0
         Environment: Spark 2.3.2, Spark 2.4.5
            Reporter: Chetan Bhat


*Steps-*

0: jdbc:hive2://10.20.255.171:23040/default> create table if not exists 
all_data_types1(*bool_1 boolean*,bool_2 boolean,chinese string,Number 
int,smallNumber smallint,BigNumber bigint,LargeDecimal double,smalldecimal 
float,customdecimal decimal(38,15),words string,smallwords char(8),varwords 
varchar(20),time timestamp,day date,emptyNumber int,emptysmallNumber 
smallint,emptyBigNumber bigint,emptyLargeDecimal double,emptysmalldecimal 
float,emptycustomdecimal decimal(38,38),emptywords string,emptysmallwords 
char(8),emptyvarwords varchar(20)) stored as carbondata TBLPROPERTIES 
('BUCKET_NUMBER'='1', '*BUCKET_COLUMNS'='bool_1*');
+---------+--+
| Result |
+---------+--+
+---------+--+
No rows selected (0.939 seconds)
0: jdbc:hive2://10.20.255.171:23040/default> LOAD DATA INPATH 
'hdfs://hacluster/chetan/datafile_0.csv' into table all_data_types1 
OPTIONS('DELIMITER'=',' , 
'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='bool_1 ,bool_2 
,chinese ,Number ,smallNumber ,BigNumber ,LargeDecimal ,smalldecimal 
,customdecimal,words ,smallwords ,varwords ,time ,day ,emptyNumber 
,emptysmallNumber ,emptyBigNumber ,emptyLargeDecimal 
,emptysmalldecimal,emptycustomdecimal ,emptywords ,emptysmallwords 
,emptyvarwords');
*Error: java.lang.Exception: DataLoad failure: (state=,code=0)*

 

*Log-*

java.lang.Exception: DataLoad failure: 
 at 
org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:560)
 at 
org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.loadData(CarbonLoadDataCommand.scala:207)
 at 
org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:168)
 at 
org.apache.spark.sql.execution.command.AtomicRunnableCommand$$anonfun$run$3.apply(package.scala:148)
 at 
org.apache.spark.sql.execution.command.AtomicRunnableCommand$$anonfun$run$3.apply(package.scala:145)
 at 
org.apache.spark.sql.execution.command.Auditable$class.runWithAudit(package.scala:104)
 at 
org.apache.spark.sql.execution.command.AtomicRunnableCommand.runWithAudit(package.scala:141)
 at 
org.apache.spark.sql.execution.command.AtomicRunnableCommand.run(package.scala:145)
 at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
 at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
 at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79)
 at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190)
 at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190)
 at org.apache.spark.sql.Dataset$$anonfun$52.apply(Dataset.scala:3259)
 at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:77)
 at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3258)
 at org.apache.spark.sql.Dataset.<init>(Dataset.scala:190)
 at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:75)
 at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
 at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:694)
 at 
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:232)
 at 
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175)
 at 
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:422)
 at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
 at 
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185)
 at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
 at java.util.concurrent.FutureTask.run(FutureTask.java:266)
 at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
 at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
 at java.lang.Thread.run(Thread.java:745)
2020-06-05 02:05:56,789 | ERROR | [HiveServer2-Background-Pool: Thread-138] | 
Error running hive query: | 
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:179)
org.apache.hive.service.cli.HiveSQLException: java.lang.Exception: DataLoad 
failure: 
 at 
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:269)
 at 
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175)
 at 
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:422)
 at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
 at 
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185)
 at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
 at java.util.concurrent.FutureTask.run(FutureTask.java:266)
 at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
 at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
 at java.lang.Thread.run(Thread.java:745)
2020-06-05 02:05:56,792 | INFO | [HiveServer2-Handler-Pool: Thread-97] | Asked 
to cancel job group 072ea11a-9b63-48c6-bed2-2f396ad82b36 | 
org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to