WTa-hash commented on issue #2057:
URL: https://github.com/apache/hudi/issues/2057#issuecomment-685015564


   For the 0.6.0 issue with error: java.lang.NoSuchMethodError: 
org.apache.spark.sql.execution.datasources.PartitionedFile.(Lorg/apache/spark/sql/catalyst/InternalRow;Ljava/lang/String;JJ[Ljava/lang/String;)V
   
   ^ I can get pass this issue if I query the table using AWS Athena and remove 
Spark reads from the script.
   
   This then brings up another issue. Keep in mind, I am using a custom 
AWSDmsAvroPayload class referenced in 
https://issues.apache.org/jira/browse/HUDI-802 with Hudi 0.6.0.
   
   Using Hudi 0.6.0, I first create a Hudi table using this dataframe:
   
![df_1](https://user-images.githubusercontent.com/64644025/91885590-aac69780-ec4d-11ea-866c-7cec53efbc7a.PNG)
   
   If I create a COW table, the dataframe gets processed correctly:
   
![COW_1](https://user-images.githubusercontent.com/64644025/91885704-d5b0eb80-ec4d-11ea-8725-487db483745b.PNG)
   
   If I create a MOR table, the dataframe also gets processed correctly 
(read-optimized table):
   
![MOR_1](https://user-images.githubusercontent.com/64644025/91885750-e82b2500-ec4d-11ea-8e68-68e85ab69364.PNG)
   
   Next, I process a new dataframe with some data changes on the existing Hudi 
table:
   
![df_2](https://user-images.githubusercontent.com/64644025/91885805-fa0cc800-ec4d-11ea-811d-9dd25d56566f.PNG)
   
   The updates are processed correctly in the COW table:
   
![COW_2](https://user-images.githubusercontent.com/64644025/91885855-0c870180-ec4e-11ea-953a-485cd8b3c3a0.PNG)
   
   However, the updates did not process the deletion of ID=2 in MOR table:
   
![MOR_2](https://user-images.githubusercontent.com/64644025/91885907-2294c200-ec4e-11ea-9b0c-08f7eec80ef3.PNG)
   
   There were no errors when processing the MOR table.


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to