This is an automated email from the ASF dual-hosted git repository.

danny0405 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/master by this push:
     new cf01b475b8e [MINOR] Hotfix in HdfsParquetImportProcedure for format 
parameter is not need to add (#9993)
cf01b475b8e is described below

commit cf01b475b8e938949de634be54e520bb9b6c0635
Author: xuzifu666 <[email protected]>
AuthorDate: Tue Nov 7 09:31:33 2023 +0800

    [MINOR] Hotfix in HdfsParquetImportProcedure for format parameter is not 
need to add (#9993)
---
 .../procedures/HdfsParquetImportProcedure.scala        | 18 ++++++++----------
 1 file changed, 8 insertions(+), 10 deletions(-)

diff --git 
a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HdfsParquetImportProcedure.scala
 
b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HdfsParquetImportProcedure.scala
index e050572899e..addfaca049f 100644
--- 
a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HdfsParquetImportProcedure.scala
+++ 
b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HdfsParquetImportProcedure.scala
@@ -34,11 +34,10 @@ class HdfsParquetImportProcedure extends BaseProcedure with 
ProcedureBuilder wit
     ProcedureParameter.required(4, "row_key", DataTypes.StringType),
     ProcedureParameter.required(5, "partition_key", DataTypes.StringType),
     ProcedureParameter.required(6, "schema_file_path", DataTypes.StringType),
-    ProcedureParameter.optional(7, "format", DataTypes.StringType, "parquet"),
-    ProcedureParameter.optional(8, "command", DataTypes.StringType, "insert"),
-    ProcedureParameter.optional(9, "retry", DataTypes.IntegerType, 0),
-    ProcedureParameter.optional(10, "parallelism", DataTypes.IntegerType, 
jsc.defaultParallelism),
-    ProcedureParameter.optional(11, "props_file_path", DataTypes.StringType, 
"")
+    ProcedureParameter.optional(7, "command", DataTypes.StringType, "insert"),
+    ProcedureParameter.optional(8, "retry", DataTypes.IntegerType, 0),
+    ProcedureParameter.optional(9, "parallelism", DataTypes.IntegerType, 
jsc.defaultParallelism),
+    ProcedureParameter.optional(10, "props_file_path", DataTypes.StringType, 
"")
   )
 
   private val OUTPUT_TYPE = new StructType(Array[StructField](
@@ -59,11 +58,10 @@ class HdfsParquetImportProcedure extends BaseProcedure with 
ProcedureBuilder wit
     val rowKey = getArgValueOrDefault(args, 
PARAMETERS(4)).get.asInstanceOf[String]
     val partitionKey = getArgValueOrDefault(args, 
PARAMETERS(5)).get.asInstanceOf[String]
     val schemaFilePath = getArgValueOrDefault(args, 
PARAMETERS(6)).get.asInstanceOf[String]
-    val format = getArgValueOrDefault(args, 
PARAMETERS(7)).get.asInstanceOf[String]
-    val command = getArgValueOrDefault(args, 
PARAMETERS(8)).get.asInstanceOf[String]
-    val retry = getArgValueOrDefault(args, PARAMETERS(9)).get.asInstanceOf[Int]
-    val parallelism = getArgValueOrDefault(args, 
PARAMETERS(10)).get.asInstanceOf[Int]
-    val propsFilePath = getArgValueOrDefault(args, 
PARAMETERS(11)).get.asInstanceOf[String]
+    val command = getArgValueOrDefault(args, 
PARAMETERS(7)).get.asInstanceOf[String]
+    val retry = getArgValueOrDefault(args, PARAMETERS(8)).get.asInstanceOf[Int]
+    val parallelism = getArgValueOrDefault(args, 
PARAMETERS(9)).get.asInstanceOf[Int]
+    val propsFilePath = getArgValueOrDefault(args, 
PARAMETERS(10)).get.asInstanceOf[String]
 
     val parquetImporterUtils: HDFSParquetImporterUtils = new 
HDFSParquetImporterUtils(command, srcPath, targetPath,
       tableName, tableType, rowKey, partitionKey, parallelism, schemaFilePath, 
retry, propsFilePath)

Reply via email to