This is an automated email from the ASF dual-hosted git repository.

yihua pushed a commit to branch branch-0.x
in repository https://gitbox.apache.org/repos/asf/hudi.git

commit e5b28b68a65933d2fdf097de9931d1285ba76724
Author: stream2000 <[email protected]>
AuthorDate: Tue Feb 27 06:47:20 2024 +0800

    [MINOR] Fix code style for HiveAvroSerializer (#10755)
---
 .../hudi/hadoop/utils/HiveAvroSerializer.java      | 22 +++++++++++-----------
 1 file changed, 11 insertions(+), 11 deletions(-)

diff --git 
a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/utils/HiveAvroSerializer.java
 
b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/utils/HiveAvroSerializer.java
index a0d1b086e03..5f33844d60c 100644
--- 
a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/utils/HiveAvroSerializer.java
+++ 
b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/utils/HiveAvroSerializer.java
@@ -99,7 +99,7 @@ public class HiveAvroSerializer {
     List<? extends StructField> allStructFieldRefs = 
soi.getAllStructFieldRefs();
     List<Object> structFieldsDataAsList = soi.getStructFieldsDataAsList(o);
 
-    for (int i  = 0; i < size; i++) {
+    for (int i = 0; i < size; i++) {
       Schema.Field field = schema.getFields().get(i);
       if (i >= columnTypes.size()) {
         break;
@@ -134,7 +134,7 @@ public class HiveAvroSerializer {
    * Determine if an Avro schema is of type Union[T, NULL].  Avro supports 
nullable
    * types via a union of type T and null.  This is a very common use case.
    * As such, we want to silently convert it to just T and allow the value to 
be null.
-   *
+   * <p>
    * When a Hive union type is used with AVRO, the schema type becomes
    * Union[NULL, T1, T2, ...]. The NULL in the union should be silently removed
    *
@@ -266,7 +266,7 @@ public class HiveAvroSerializer {
     GenericData.Record record = new GenericData.Record(schema);
     ArrayList<TypeInfo> allStructFieldTypeInfos = 
typeInfo.getAllStructFieldTypeInfos();
 
-    for (int i  = 0; i < size; i++) {
+    for (int i = 0; i < size; i++) {
       Schema.Field field = schema.getFields().get(i);
       setUpRecordFieldFromWritable(allStructFieldTypeInfos.get(i), 
structFieldsDataAsList.get(i),
           allStructFieldRefs.get(i).getFieldObjectInspector(), record, field);
@@ -278,26 +278,26 @@ public class HiveAvroSerializer {
     switch (fieldOI.getPrimitiveCategory()) {
       case BINARY:
         if (schema.getType() == Schema.Type.BYTES) {
-          return 
AvroSerdeUtils.getBufferFromBytes((byte[])fieldOI.getPrimitiveJavaObject(structFieldData));
+          return AvroSerdeUtils.getBufferFromBytes((byte[]) 
fieldOI.getPrimitiveJavaObject(structFieldData));
         } else if (schema.getType() == Schema.Type.FIXED) {
-          GenericData.Fixed fixed = new GenericData.Fixed(schema, 
(byte[])fieldOI.getPrimitiveJavaObject(structFieldData));
+          GenericData.Fixed fixed = new GenericData.Fixed(schema, (byte[]) 
fieldOI.getPrimitiveJavaObject(structFieldData));
           return fixed;
         } else {
           throw new HoodieException("Unexpected Avro schema for Binary 
TypeInfo: " + schema.getType());
         }
       case DECIMAL:
-        HiveDecimal dec = 
(HiveDecimal)fieldOI.getPrimitiveJavaObject(structFieldData);
-        LogicalTypes.Decimal decimal = 
(LogicalTypes.Decimal)schema.getLogicalType();
+        HiveDecimal dec = (HiveDecimal) 
fieldOI.getPrimitiveJavaObject(structFieldData);
+        LogicalTypes.Decimal decimal = (LogicalTypes.Decimal) 
schema.getLogicalType();
         BigDecimal bd = new 
BigDecimal(dec.toString()).setScale(decimal.getScale());
         return HoodieAvroUtils.DECIMAL_CONVERSION.toFixed(bd, schema, decimal);
       case CHAR:
-        HiveChar ch = 
(HiveChar)fieldOI.getPrimitiveJavaObject(structFieldData);
+        HiveChar ch = (HiveChar) 
fieldOI.getPrimitiveJavaObject(structFieldData);
         return new Utf8(ch.getStrippedValue());
       case VARCHAR:
-        HiveVarchar vc = 
(HiveVarchar)fieldOI.getPrimitiveJavaObject(structFieldData);
+        HiveVarchar vc = (HiveVarchar) 
fieldOI.getPrimitiveJavaObject(structFieldData);
         return new Utf8(vc.getValue());
       case STRING:
-        String string = 
(String)fieldOI.getPrimitiveJavaObject(structFieldData);
+        String string = (String) 
fieldOI.getPrimitiveJavaObject(structFieldData);
         return new Utf8(string);
       case DATE:
         return HoodieHiveUtils.getDays(structFieldData);
@@ -364,7 +364,7 @@ public class HiveAvroSerializer {
     ObjectInspector mapValueObjectInspector = 
fieldOI.getMapValueObjectInspector();
     TypeInfo mapKeyTypeInfo = typeInfo.getMapKeyTypeInfo();
     TypeInfo mapValueTypeInfo = typeInfo.getMapValueTypeInfo();
-    Map<?,?> map = fieldOI.getMap(structFieldData);
+    Map<?, ?> map = fieldOI.getMap(structFieldData);
     Schema valueType = schema.getValueType();
 
     Map<Object, Object> deserialized = new LinkedHashMap<Object, 
Object>(fieldOI.getMapSize(structFieldData));

Reply via email to