This is an automated email from the ASF dual-hosted git repository. danny0405 pushed a commit to branch release-0.10.0 in repository https://gitbox.apache.org/repos/asf/hudi.git
commit 69d14353bcc20c07d53b7835940c4117cfbf5110 Author: 董可伦 <[email protected]> AuthorDate: Tue Nov 30 10:31:22 2021 +0800 [MINOR] Fix typo,rename 'getUrlEncodePartitoning' to 'getUrlEncodePartitioning' (#4130) (cherry picked from commit 3433f00cb5366fa190a7b2d5254255041a705deb) --- .../src/main/java/org/apache/hudi/common/table/HoodieTableConfig.java | 2 +- .../org/apache/spark/sql/hudi/command/DeleteHoodieTableCommand.scala | 2 +- .../apache/spark/sql/hudi/command/InsertIntoHoodieTableCommand.scala | 2 +- .../apache/spark/sql/hudi/command/MergeIntoHoodieTableCommand.scala | 2 +- .../org/apache/spark/sql/hudi/command/UpdateHoodieTableCommand.scala | 2 +- .../src/test/scala/org/apache/hudi/TestHoodieSparkSqlWriter.scala | 4 ++-- 6 files changed, 7 insertions(+), 7 deletions(-) diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableConfig.java b/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableConfig.java index 1da4130..e4b60e2 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableConfig.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableConfig.java @@ -466,7 +466,7 @@ public class HoodieTableConfig extends HoodieConfig { return getString(HIVE_STYLE_PARTITIONING_ENABLE); } - public String getUrlEncodePartitoning() { + public String getUrlEncodePartitioning() { return getString(URL_ENCODE_PARTITIONING); } diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/DeleteHoodieTableCommand.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/DeleteHoodieTableCommand.scala index 1decb5d..8475b63 100644 --- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/DeleteHoodieTableCommand.scala +++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/DeleteHoodieTableCommand.scala @@ -72,7 +72,7 @@ case class DeleteHoodieTableCommand(deleteTable: DeleteFromTable) extends Runnab "path" -> path, TBL_NAME.key -> tableConfig.getTableName, HIVE_STYLE_PARTITIONING.key -> tableConfig.getHiveStylePartitioningEnable, - URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitoning, + URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitioning, KEYGENERATOR_CLASS_NAME.key -> classOf[SqlKeyGenerator].getCanonicalName, SqlKeyGenerator.ORIGIN_KEYGEN_CLASS_NAME -> tableConfig.getKeyGeneratorClassName, OPERATION.key -> DataSourceWriteOptions.DELETE_OPERATION_OPT_VAL, diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/InsertIntoHoodieTableCommand.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/InsertIntoHoodieTableCommand.scala index 635aa64..4b6d8e0 100644 --- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/InsertIntoHoodieTableCommand.scala +++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/InsertIntoHoodieTableCommand.scala @@ -212,7 +212,7 @@ object InsertIntoHoodieTableCommand extends Logging { val partitionFields = hoodieCatalogTable.partitionFields.mkString(",") val hiveStylePartitioningEnable = Option(tableConfig.getHiveStylePartitioningEnable).getOrElse("true") - val urlEncodePartitioning = Option(tableConfig.getUrlEncodePartitoning).getOrElse("false") + val urlEncodePartitioning = Option(tableConfig.getUrlEncodePartitioning).getOrElse("false") val keyGeneratorClassName = Option(tableConfig.getKeyGeneratorClassName) .getOrElse(classOf[ComplexKeyGenerator].getCanonicalName) diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/MergeIntoHoodieTableCommand.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/MergeIntoHoodieTableCommand.scala index a746f83..76c8715 100644 --- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/MergeIntoHoodieTableCommand.scala +++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/MergeIntoHoodieTableCommand.scala @@ -449,7 +449,7 @@ case class MergeIntoHoodieTableCommand(mergeInto: MergeIntoTable) extends Runnab PARTITIONPATH_FIELD.key -> tableConfig.getPartitionFieldProp, PAYLOAD_CLASS_NAME.key -> classOf[ExpressionPayload].getCanonicalName, HIVE_STYLE_PARTITIONING.key -> tableConfig.getHiveStylePartitioningEnable, - URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitoning, + URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitioning, KEYGENERATOR_CLASS_NAME.key -> classOf[SqlKeyGenerator].getCanonicalName, SqlKeyGenerator.ORIGIN_KEYGEN_CLASS_NAME -> tableConfig.getKeyGeneratorClassName, META_SYNC_ENABLED.key -> enableHive.toString, diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/UpdateHoodieTableCommand.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/UpdateHoodieTableCommand.scala index a6a0cb1..60122ce 100644 --- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/UpdateHoodieTableCommand.scala +++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/UpdateHoodieTableCommand.scala @@ -98,7 +98,7 @@ case class UpdateHoodieTableCommand(updateTable: UpdateTable) extends RunnableCo PRECOMBINE_FIELD.key -> preCombineColumn, TBL_NAME.key -> hoodieCatalogTable.tableName, HIVE_STYLE_PARTITIONING.key -> tableConfig.getHiveStylePartitioningEnable, - URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitoning, + URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitioning, KEYGENERATOR_CLASS_NAME.key -> classOf[SqlKeyGenerator].getCanonicalName, SqlKeyGenerator.ORIGIN_KEYGEN_CLASS_NAME -> tableConfig.getKeyGeneratorClassName, OPERATION.key -> UPSERT_OPERATION_OPT_VAL, diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/TestHoodieSparkSqlWriter.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/TestHoodieSparkSqlWriter.scala index fbdfb69..48f8408 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/TestHoodieSparkSqlWriter.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/TestHoodieSparkSqlWriter.scala @@ -853,7 +853,7 @@ class TestHoodieSparkSqlWriter { .setConf(spark.sparkContext.hadoopConfiguration) .setBasePath(tablePath1).build().getTableConfig assert(tableConfig1.getHiveStylePartitioningEnable == "true") - assert(tableConfig1.getUrlEncodePartitoning == "false") + assert(tableConfig1.getUrlEncodePartitioning == "false") assert(tableConfig1.getKeyGeneratorClassName == classOf[ComplexKeyGenerator].getName) df.write.format("hudi") .options(options) @@ -874,7 +874,7 @@ class TestHoodieSparkSqlWriter { .setConf(spark.sparkContext.hadoopConfiguration) .setBasePath(tablePath2).build().getTableConfig assert(tableConfig2.getHiveStylePartitioningEnable == "false") - assert(tableConfig2.getUrlEncodePartitoning == "true") + assert(tableConfig2.getUrlEncodePartitioning == "true") assert(tableConfig2.getKeyGeneratorClassName == classOf[SimpleKeyGenerator].getName) val df2 = Seq((2, "a2", 20, 1000, "2021-10-16")).toDF("id", "name", "value", "ts", "dt")
