bhushanamk opened a new issue #2258:
URL: https://github.com/apache/hudi/issues/2258


   **_Tips before filing an issue_**
   
   - Have you gone through our 
[FAQs](https://cwiki.apache.org/confluence/display/HUDI/FAQ)?Yes
   
   - Join the mailing list to engage in conversations and get faster support at 
[email protected].
   
   - If you have triaged this as a bug, then file an 
[issue](https://issues.apache.org/jira/projects/HUDI/issues) directly.
   
   **Describe the problem you faced**
   
   presto:smart_api> select * from cities;Query 20201117_173741_00013_27sh4, 
FAILED, 2 nodesSplits: 59 total, 0 done (0.00%)0:03 [0 rows, 0B] [0 rows/s, 
0B/s]Query 20201117_173741_00013_27sh4 failed: 
org.apache.hadoop.io.LongWritable cannot be cast to 
org.apache.hadoop.hive.serde2.io.HiveDecimalWritable
   A clear and concise description of the problem.
   
   **To Reproduce**
   
   Steps to reproduce the behavior:
   
   1. create an emr
   2. run deltasreamer 
   spark-submit --class 
org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer --conf 
spark.sql.hive.convertMetastoreParquet=false --packages 
org.apache.spark:spark-avro_2.11:2.4.6 --master yarn --deploy-mode client 
s3://mrTeam-4/dev/hudi-utilities-bundle_2.11-0.6.0.jar --table-type 
COPY_ON_WRITE   --source-ordering-field cdc_ts   --source-class 
org.apache.hudi.utilities.sources.ParquetDFSSource   --target-base-path 
s3://nxgen-dataplatform-4/prod/mysql/default/cities --target-table cities  
--enable-hive-sync  --transformer-class 
org.apache.hudi.utilities.transform.AWSDmsTransformer   --payload-class 
org.apache.hudi.payload.AWSDmsAvroPayload   --hoodie-conf 
hoodie.datasource.write.recordkey.field=id --hoodie-conf 
hoodie.datasource.write.partitionpath.field=state_id --hoodie-conf 
hoodie.datasource.hive_sync.partition_fields=state_id --hoodie-conf 
hoodie.deltastreamer.source.dfs.root=s3a://nxgen-dataplatform-4-rawzone/mysql/default/cities
 --hoodie-conf hoodie.datasource.hive_sync.e
 nable=true --hoodie-conf hoodie.datasource.hive_sync.database=default 
--hoodie-conf hoodie.datasource.hive_sync.table=cities --hoodie-conf 
hoodie.datasource.hive_sync.username=,hoodie.datasource.hive_sync.password= 
--hoodie-conf 
hoodie.datasource.hive_sync.jdbcurl=jdbc:hive2://ip-10-0-2-11.ap-south-1.compute.internal:10000/
 --hoodie-conf 
hoodie.datasource.hive_sync.partition_extractor_class=org.apache.hudi.hive.MultiPartKeysValueExtractor
   3.
   Hive , query the data ,
   hive> select * from cities LIMIT 10;;
   OK
   20201117172523  20201117172523_1_3      4595    1000    
41d094d9-1154-4988-b197-be660eb5d8cd-0_1-23-12092_20201117172523.parquet        
2020-11-16 05:19:05.773893   4595    Akhnoor 181204  NULL    32.8455 74.4756 
true    NULL    NULL            1000
   20201117172523  20201117172523_1_6      3822    1000    
41d094d9-1154-4988-b197-be660eb5d8cd-0_1-23-12092_20201117172523.parquet        
2020-11-16 05:19:05.769144   3822    Anantnag        192101  NULL    33.7000 
75.1100 true    NULL    NULL            1000
   20201117172523  20201117172523_1_21     1964    1000    
41d094d9-1154-4988-b197-be660eb5d8cd-0_1-23-12092_20201117172523.parquet        
2020-11-16 05:19:05.757823   1964    Diver Anderbugh 193223  NULL    34.4590 
74.4486 true    NULL    NULL            1000
   20201117172523  20201117172523_1_23     1512    1000    
41d094d9-1154-4988-b197-be660eb5d8cd-0_1-23-12092_20201117172523.parquet        
2020-11-16 05:19:05.755653   1512    Jammu   180001  NULL    NULL    NULL    
true    24166   24166           1000
   20201117172523  20201117172523_40_3024  5684    1001    
f9886f5e-78cb-40b0-887a-fa59c330bfb6-0_40-23-12131_20201117172523.parquet       
2020-11-16 05:19:05.780776   5684    Pathankot       145001  NULL    32.2511 
75.6575 true    NULL    NULL            1001
   20201117172523  20201117172523_40_3025  6048    1001    
f9886f5e-78cb-40b0-887a-fa59c330bfb6-0_40-23-12131_20201117172523.parquet       
2020-11-16 05:19:05.783052   6048    Kot     175028  NULL    31.5535 77.0278 
true    NULL    NULL            1001
   20201117172523  20201117172523_40_3026  6090    1001    
f9886f5e-78cb-40b0-887a-fa59c330bfb6-0_40-23-12131_20201117172523.parquet       
2020-11-16 05:19:05.783305   6090    Shillihar       175125  NULL    31.9148 
77.2184 true    NULL    NULL            1001
   20201117172523  20201117172523_40_3027  4603    1001    
f9886f5e-78cb-40b0-887a-fa59c330bfb6-0_40-23-12131_20201117172523.parquet       
2020-11-16 05:19:05.773941   4603    Mant Khas       176215  NULL    32.2193 
76.3226 true    NULL    NULL            1001
   20201117172523  20201117172523_40_3028  6355    1001    
f9886f5e-78cb-40b0-887a-fa59c330bfb6-0_40-23-12131_20201117172523.parquet       
2020-11-16 05:19:05.785007   6355    Dharampur       173209  NULL    30.9005 
77.0150 true    NULL    NULL            1001
   20201117172523  20201117172523_40_3029  1682    1001    
f9886f5e-78cb-40b0-887a-fa59c330bfb6-0_40-23-12131_20201117172523.parquet       
2020-11-16 05:19:05.756270   1682    Bharlar 176402  NULL    32.2098 75.7375 
true    NULL    NULL            1001
   Time taken: 0.162 seconds, Fetched: 10 row(s)
   
   4.
   presto:smart_api> select * from cities limit  10;
   
   Query 20201118_045733_00001_27sh4, FAILED, 2 nodes
   http://localhost:8889/ui/query.html?20201118_045733_00001_27sh4
   Splits: 60 total, 0 done (0.00%)
   CPU Time: 0.0s total,     0 rows/s,     0B/s, 0% active
   Per Node: 0.0 parallelism,     0 rows/s,     0B/s
   Parallelism: 0.0
   Peak User Memory: 0B
   Peak Total Memory: 0B
   Peak Task Total Memory: 0B
   0:02 [0 rows, 0B] [0 rows/s, 0B/s]
   
   Query 20201118_045733_00001_27sh4 failed: org.apache.hadoop.io.LongWritable 
cannot be cast to org.apache.hadoop.hive.serde2.io.HiveDecimalWritable
   java.lang.ClassCastException: org.apache.hadoop.io.LongWritable cannot be 
cast to org.apache.hadoop.hive.serde2.io.HiveDecimalWritable
           at 
org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableHiveDecimalObjectInspector.getPrimitiveJavaObject(WritableHiveDecimalObjectInspector.java:49)
   
   
   
   **Environment Description**
   
   * Hudi version : 0.6.0
   
   * Spark version : 2.4.6
   
   * Hive version : Hive 2.3.7-amzn-1
   
   * Hadoop version : Hadoop 2.10.0-amzn-0
   
   * Storage (HDFS/S3/GCS..) : S3
   
   * Running on Docker? (yes/no) : No
   
   
   **Additional context**
   
   Add any other context about the problem here.
   
   **Stacktrace**
   
   ```Add the stacktrace of the error.```
   
   presto:smart_api> select * from cities limit  10;
   
   Query 20201118_045733_00001_27sh4, FAILED, 2 nodes
   http://localhost:8889/ui/query.html?20201118_045733_00001_27sh4
   Splits: 60 total, 0 done (0.00%)
   CPU Time: 0.0s total,     0 rows/s,     0B/s, 0% active
   Per Node: 0.0 parallelism,     0 rows/s,     0B/s
   Parallelism: 0.0
   Peak User Memory: 0B
   Peak Total Memory: 0B
   Peak Task Total Memory: 0B
   0:02 [0 rows, 0B] [0 rows/s, 0B/s]
   
   Query 20201118_045733_00001_27sh4 failed: org.apache.hadoop.io.LongWritable 
cannot be cast to org.apache.hadoop.hive.serde2.io.HiveDecimalWritable
   java.lang.ClassCastException: org.apache.hadoop.io.LongWritable cannot be 
cast to org.apache.hadoop.hive.serde2.io.HiveDecimalWritable
           at 
org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableHiveDecimalObjectInspector.getPrimitiveJavaObject(WritableHiveDecimalObjectInspector.java:49)
           at 
org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableHiveDecimalObjectInspector.getPrimitiveJavaObject(WritableHiveDecimalObjectInspector.java:26)
           at 
com.facebook.presto.hive.GenericHiveRecordCursor.parseDecimalColumn(GenericHiveRecordCursor.java:416)
           at 
com.facebook.presto.hive.GenericHiveRecordCursor.parseColumn(GenericHiveRecordCursor.java:511)
           at 
com.facebook.presto.hive.GenericHiveRecordCursor.isNull(GenericHiveRecordCursor.java:466)
           at 
com.facebook.presto.hive.HiveRecordCursor.isNull(HiveRecordCursor.java:233)
           at 
com.facebook.presto.spi.RecordPageSource.getNextPage(RecordPageSource.java:112)
           at 
com.facebook.presto.operator.TableScanOperator.getOutput(TableScanOperator.java:262)
           at 
com.facebook.presto.operator.Driver.processInternal(Driver.java:382)
           at 
com.facebook.presto.operator.Driver.lambda$processFor$8(Driver.java:284)
           at com.facebook.presto.operator.Driver.tryWithLock(Driver.java:672)
           at com.facebook.presto.operator.Driver.processFor(Driver.java:277)
           at 
com.facebook.presto.execution.SqlTaskExecution$DriverSplitRunner.processFor(SqlTaskExecution.java:1077)
           at 
com.facebook.presto.execution.executor.PrioritizedSplitRunner.process(PrioritizedSplitRunner.java:162)
           at 
com.facebook.presto.execution.executor.TaskExecutor$TaskRunner.run(TaskExecutor.java:545)
           at 
com.facebook.presto.$gen.Presto_0_238_3_amzn_0____20201117_171112_1.run(Unknown 
Source)
           at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
           at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
           at java.lang.Thread.run(Thread.java:748)
   


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to