This is an automated email from the ASF dual-hosted git repository.
danny0405 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/master by this push:
new cf01b475b8e [MINOR] Hotfix in HdfsParquetImportProcedure for format
parameter is not need to add (#9993)
cf01b475b8e is described below
commit cf01b475b8e938949de634be54e520bb9b6c0635
Author: xuzifu666 <[email protected]>
AuthorDate: Tue Nov 7 09:31:33 2023 +0800
[MINOR] Hotfix in HdfsParquetImportProcedure for format parameter is not
need to add (#9993)
---
.../procedures/HdfsParquetImportProcedure.scala | 18 ++++++++----------
1 file changed, 8 insertions(+), 10 deletions(-)
diff --git
a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HdfsParquetImportProcedure.scala
b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HdfsParquetImportProcedure.scala
index e050572899e..addfaca049f 100644
---
a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HdfsParquetImportProcedure.scala
+++
b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/HdfsParquetImportProcedure.scala
@@ -34,11 +34,10 @@ class HdfsParquetImportProcedure extends BaseProcedure with
ProcedureBuilder wit
ProcedureParameter.required(4, "row_key", DataTypes.StringType),
ProcedureParameter.required(5, "partition_key", DataTypes.StringType),
ProcedureParameter.required(6, "schema_file_path", DataTypes.StringType),
- ProcedureParameter.optional(7, "format", DataTypes.StringType, "parquet"),
- ProcedureParameter.optional(8, "command", DataTypes.StringType, "insert"),
- ProcedureParameter.optional(9, "retry", DataTypes.IntegerType, 0),
- ProcedureParameter.optional(10, "parallelism", DataTypes.IntegerType,
jsc.defaultParallelism),
- ProcedureParameter.optional(11, "props_file_path", DataTypes.StringType,
"")
+ ProcedureParameter.optional(7, "command", DataTypes.StringType, "insert"),
+ ProcedureParameter.optional(8, "retry", DataTypes.IntegerType, 0),
+ ProcedureParameter.optional(9, "parallelism", DataTypes.IntegerType,
jsc.defaultParallelism),
+ ProcedureParameter.optional(10, "props_file_path", DataTypes.StringType,
"")
)
private val OUTPUT_TYPE = new StructType(Array[StructField](
@@ -59,11 +58,10 @@ class HdfsParquetImportProcedure extends BaseProcedure with
ProcedureBuilder wit
val rowKey = getArgValueOrDefault(args,
PARAMETERS(4)).get.asInstanceOf[String]
val partitionKey = getArgValueOrDefault(args,
PARAMETERS(5)).get.asInstanceOf[String]
val schemaFilePath = getArgValueOrDefault(args,
PARAMETERS(6)).get.asInstanceOf[String]
- val format = getArgValueOrDefault(args,
PARAMETERS(7)).get.asInstanceOf[String]
- val command = getArgValueOrDefault(args,
PARAMETERS(8)).get.asInstanceOf[String]
- val retry = getArgValueOrDefault(args, PARAMETERS(9)).get.asInstanceOf[Int]
- val parallelism = getArgValueOrDefault(args,
PARAMETERS(10)).get.asInstanceOf[Int]
- val propsFilePath = getArgValueOrDefault(args,
PARAMETERS(11)).get.asInstanceOf[String]
+ val command = getArgValueOrDefault(args,
PARAMETERS(7)).get.asInstanceOf[String]
+ val retry = getArgValueOrDefault(args, PARAMETERS(8)).get.asInstanceOf[Int]
+ val parallelism = getArgValueOrDefault(args,
PARAMETERS(9)).get.asInstanceOf[Int]
+ val propsFilePath = getArgValueOrDefault(args,
PARAMETERS(10)).get.asInstanceOf[String]
val parquetImporterUtils: HDFSParquetImporterUtils = new
HDFSParquetImporterUtils(command, srcPath, targetPath,
tableName, tableType, rowKey, partitionKey, parallelism, schemaFilePath,
retry, propsFilePath)