vaibhav-sinha commented on a change in pull request #2923:
URL: https://github.com/apache/hudi/pull/2923#discussion_r646715922



##########
File path: 
hudi-flink/src/test/java/org/apache/hudi/sink/TestWriteCopyOnWrite.java
##########
@@ -380,12 +380,12 @@ public void testUpsertWithDelete() throws Exception {
   @Test
   public void testInsertWithMiniBatches() throws Exception {
     // reset the config option
-    conf.setDouble(FlinkOptions.WRITE_BATCH_SIZE, 0.0006); // 630 bytes batch 
size
+    conf.setDouble(FlinkOptions.WRITE_BATCH_SIZE, 0.00075); // 786 bytes batch 
size

Review comment:
       I think the reason for this is as follows:
   
   For timestamp field, HoodieAvroUtils now returns an instance of 
LocalDateTime instead of Long. When Row data is being converted to 
HoodieRecordPayload, then `orderingVal` is being fetched using 
`HoodieAvroUtils.getNestedFieldVal`. Hence the structure of the payload has 
changed and is of larger size than before. 
   
   ```
   public HoodieRecordPayload<?> createPayload(GenericRecord record, boolean 
isDelete) throws Exception {
       if (shouldCombine) {
         ValidationUtils.checkState(preCombineField != null);
         Comparable<?> orderingVal = (Comparable<?>) 
HoodieAvroUtils.getNestedFieldVal(record,
             preCombineField, false);
         return (HoodieRecordPayload<?>) constructor.newInstance(
             isDelete ? null : record, orderingVal);
       } else {
         return (HoodieRecordPayload<?>) 
this.constructor.newInstance(Option.of(record));
       }
     }
   ```




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to