This is an automated email from the ASF dual-hosted git repository.

yihua pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new 11f7a2eed6e [DOCS] Fix dataframe write option in schema_evolution docs 
(#7481)
11f7a2eed6e is described below

commit 11f7a2eed6e92e0c727315d9385f545c99c35250
Author: Zouxxyy <[email protected]>
AuthorDate: Mon Mar 4 07:43:43 2024 +0800

    [DOCS] Fix dataframe write option in schema_evolution docs (#7481)
    
    Co-authored-by: Y Ethan Guo <[email protected]>
---
 website/docs/schema_evolution.md                         | 16 ++++++++--------
 .../versioned_docs/version-0.10.0/schema_evolution.md    | 16 ++++++++--------
 .../versioned_docs/version-0.10.1/schema_evolution.md    | 16 ++++++++--------
 .../versioned_docs/version-0.11.0/schema_evolution.md    | 16 ++++++++--------
 .../versioned_docs/version-0.11.1/schema_evolution.md    | 16 ++++++++--------
 .../versioned_docs/version-0.12.0/schema_evolution.md    | 16 ++++++++--------
 .../versioned_docs/version-0.12.1/schema_evolution.md    | 16 ++++++++--------
 .../versioned_docs/version-0.12.2/schema_evolution.md    | 16 ++++++++--------
 .../versioned_docs/version-0.13.0/schema_evolution.md    | 16 ++++++++--------
 .../versioned_docs/version-0.13.1/schema_evolution.md    | 16 ++++++++--------
 .../versioned_docs/version-0.14.0/schema_evolution.md    | 16 ++++++++--------
 .../versioned_docs/version-0.14.1/schema_evolution.md    | 16 ++++++++--------
 12 files changed, 96 insertions(+), 96 deletions(-)

diff --git a/website/docs/schema_evolution.md b/website/docs/schema_evolution.md
index ca086b6dfe0..e8cfc53bff2 100755
--- a/website/docs/schema_evolution.md
+++ b/website/docs/schema_evolution.md
@@ -207,11 +207,11 @@ val data1 = Seq(Row("row_1", "part_0", 0L, "bob", "v_0", 
0),
 var dfFromData1 = spark.createDataFrame(data1, schema)
 dfFromData1.write.format("hudi").
    options(getQuickstartWriteConfigs).
-   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+   option(PRECOMBINE_FIELD.key, "preComb").
+   option(RECORDKEY_FIELD.key, "rowId").
+   option(PARTITIONPATH_FIELD.key, "partitionId").
    option("hoodie.index.type","SIMPLE").
-   option(TABLE_NAME.key, tableName).
+   option(TBL_NAME.key, tableName).
    mode(Overwrite).
    save(basePath)
 
@@ -266,11 +266,11 @@ val data2 = Seq(Row("row_2", "part_0", 5L, "john", "v_3", 
3L, "newField_1"),
 var dfFromData2 = spark.createDataFrame(data2, newSchema)
 dfFromData2.write.format("hudi").
     options(getQuickstartWriteConfigs).
-    option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    option(PRECOMBINE_FIELD.key, "preComb").
+    option(RECORDKEY_FIELD.key, "rowId").
+    option(PARTITIONPATH_FIELD.key, "partitionId").
     option("hoodie.index.type","SIMPLE").
-    option(TABLE_NAME.key, tableName).
+    option(TBL_NAME.key, tableName).
     mode(Append).
     save(basePath)
 
diff --git a/website/versioned_docs/version-0.10.0/schema_evolution.md 
b/website/versioned_docs/version-0.10.0/schema_evolution.md
index bd80f0d290c..481ed236f52 100644
--- a/website/versioned_docs/version-0.10.0/schema_evolution.md
+++ b/website/versioned_docs/version-0.10.0/schema_evolution.md
@@ -87,11 +87,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob", 
"v_0", 0),
 scala> var dfFromData1 = spark.createDataFrame(data1, schema)
 scala> dfFromData1.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Overwrite).
     |   save(basePath)
 
@@ -147,11 +147,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john", 
"v_3", 3L, "newField_1
 scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
 scala> dfFromData2.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Append).
     |   save(basePath)
 
diff --git a/website/versioned_docs/version-0.10.1/schema_evolution.md 
b/website/versioned_docs/version-0.10.1/schema_evolution.md
index bd80f0d290c..481ed236f52 100644
--- a/website/versioned_docs/version-0.10.1/schema_evolution.md
+++ b/website/versioned_docs/version-0.10.1/schema_evolution.md
@@ -87,11 +87,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob", 
"v_0", 0),
 scala> var dfFromData1 = spark.createDataFrame(data1, schema)
 scala> dfFromData1.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Overwrite).
     |   save(basePath)
 
@@ -147,11 +147,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john", 
"v_3", 3L, "newField_1
 scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
 scala> dfFromData2.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Append).
     |   save(basePath)
 
diff --git a/website/versioned_docs/version-0.11.0/schema_evolution.md 
b/website/versioned_docs/version-0.11.0/schema_evolution.md
index 06b755d71e4..fd4521e5f0d 100755
--- a/website/versioned_docs/version-0.11.0/schema_evolution.md
+++ b/website/versioned_docs/version-0.11.0/schema_evolution.md
@@ -265,11 +265,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob", 
"v_0", 0),
 scala> var dfFromData1 = spark.createDataFrame(data1, schema)
 scala> dfFromData1.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Overwrite).
     |   save(basePath)
 
@@ -325,11 +325,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john", 
"v_3", 3L, "newField_1
 scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
 scala> dfFromData2.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Append).
     |   save(basePath)
 
diff --git a/website/versioned_docs/version-0.11.1/schema_evolution.md 
b/website/versioned_docs/version-0.11.1/schema_evolution.md
index 06b755d71e4..fd4521e5f0d 100755
--- a/website/versioned_docs/version-0.11.1/schema_evolution.md
+++ b/website/versioned_docs/version-0.11.1/schema_evolution.md
@@ -265,11 +265,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob", 
"v_0", 0),
 scala> var dfFromData1 = spark.createDataFrame(data1, schema)
 scala> dfFromData1.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Overwrite).
     |   save(basePath)
 
@@ -325,11 +325,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john", 
"v_3", 3L, "newField_1
 scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
 scala> dfFromData2.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Append).
     |   save(basePath)
 
diff --git a/website/versioned_docs/version-0.12.0/schema_evolution.md 
b/website/versioned_docs/version-0.12.0/schema_evolution.md
index ed4a1f7d539..d1cac665a59 100755
--- a/website/versioned_docs/version-0.12.0/schema_evolution.md
+++ b/website/versioned_docs/version-0.12.0/schema_evolution.md
@@ -266,11 +266,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob", 
"v_0", 0),
 scala> var dfFromData1 = spark.createDataFrame(data1, schema)
 scala> dfFromData1.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Overwrite).
     |   save(basePath)
 
@@ -326,11 +326,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john", 
"v_3", 3L, "newField_1
 scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
 scala> dfFromData2.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Append).
     |   save(basePath)
 
diff --git a/website/versioned_docs/version-0.12.1/schema_evolution.md 
b/website/versioned_docs/version-0.12.1/schema_evolution.md
index ed4a1f7d539..d1cac665a59 100755
--- a/website/versioned_docs/version-0.12.1/schema_evolution.md
+++ b/website/versioned_docs/version-0.12.1/schema_evolution.md
@@ -266,11 +266,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob", 
"v_0", 0),
 scala> var dfFromData1 = spark.createDataFrame(data1, schema)
 scala> dfFromData1.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Overwrite).
     |   save(basePath)
 
@@ -326,11 +326,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john", 
"v_3", 3L, "newField_1
 scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
 scala> dfFromData2.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Append).
     |   save(basePath)
 
diff --git a/website/versioned_docs/version-0.12.2/schema_evolution.md 
b/website/versioned_docs/version-0.12.2/schema_evolution.md
index ed4a1f7d539..d1cac665a59 100755
--- a/website/versioned_docs/version-0.12.2/schema_evolution.md
+++ b/website/versioned_docs/version-0.12.2/schema_evolution.md
@@ -266,11 +266,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob", 
"v_0", 0),
 scala> var dfFromData1 = spark.createDataFrame(data1, schema)
 scala> dfFromData1.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Overwrite).
     |   save(basePath)
 
@@ -326,11 +326,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john", 
"v_3", 3L, "newField_1
 scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
 scala> dfFromData2.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Append).
     |   save(basePath)
 
diff --git a/website/versioned_docs/version-0.13.0/schema_evolution.md 
b/website/versioned_docs/version-0.13.0/schema_evolution.md
index ed4a1f7d539..d1cac665a59 100755
--- a/website/versioned_docs/version-0.13.0/schema_evolution.md
+++ b/website/versioned_docs/version-0.13.0/schema_evolution.md
@@ -266,11 +266,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob", 
"v_0", 0),
 scala> var dfFromData1 = spark.createDataFrame(data1, schema)
 scala> dfFromData1.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Overwrite).
     |   save(basePath)
 
@@ -326,11 +326,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john", 
"v_3", 3L, "newField_1
 scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
 scala> dfFromData2.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Append).
     |   save(basePath)
 
diff --git a/website/versioned_docs/version-0.13.1/schema_evolution.md 
b/website/versioned_docs/version-0.13.1/schema_evolution.md
index ed4a1f7d539..d1cac665a59 100755
--- a/website/versioned_docs/version-0.13.1/schema_evolution.md
+++ b/website/versioned_docs/version-0.13.1/schema_evolution.md
@@ -266,11 +266,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob", 
"v_0", 0),
 scala> var dfFromData1 = spark.createDataFrame(data1, schema)
 scala> dfFromData1.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Overwrite).
     |   save(basePath)
 
@@ -326,11 +326,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john", 
"v_3", 3L, "newField_1
 scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
 scala> dfFromData2.write.format("hudi").
     |   options(getQuickstartWriteConfigs).
-    |   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    |   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    |   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    |   option(PRECOMBINE_FIELD.key, "preComb").
+    |   option(RECORDKEY_FIELD.key, "rowId").
+    |   option(PARTITIONPATH_FIELD.key, "partitionId").
     |   option("hoodie.index.type","SIMPLE").
-    |   option(TABLE_NAME.key, tableName).
+    |   option(TBL_NAME.key, tableName).
     |   mode(Append).
     |   save(basePath)
 
diff --git a/website/versioned_docs/version-0.14.0/schema_evolution.md 
b/website/versioned_docs/version-0.14.0/schema_evolution.md
index 8fe04d65238..d01723f496d 100755
--- a/website/versioned_docs/version-0.14.0/schema_evolution.md
+++ b/website/versioned_docs/version-0.14.0/schema_evolution.md
@@ -190,11 +190,11 @@ val data1 = Seq(Row("row_1", "part_0", 0L, "bob", "v_0", 
0),
 var dfFromData1 = spark.createDataFrame(data1, schema)
 dfFromData1.write.format("hudi").
    options(getQuickstartWriteConfigs).
-   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+   option(PRECOMBINE_FIELD.key, "preComb").
+   option(RECORDKEY_FIELD.key, "rowId").
+   option(PARTITIONPATH_FIELD.key, "partitionId").
    option("hoodie.index.type","SIMPLE").
-   option(TABLE_NAME.key, tableName).
+   option(TBL_NAME.key, tableName).
    mode(Overwrite).
    save(basePath)
 
@@ -249,11 +249,11 @@ val data2 = Seq(Row("row_2", "part_0", 5L, "john", "v_3", 
3L, "newField_1"),
 var dfFromData2 = spark.createDataFrame(data2, newSchema)
 dfFromData2.write.format("hudi").
     options(getQuickstartWriteConfigs).
-    option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    option(PRECOMBINE_FIELD.key, "preComb").
+    option(RECORDKEY_FIELD.key, "rowId").
+    option(PARTITIONPATH_FIELD.key, "partitionId").
     option("hoodie.index.type","SIMPLE").
-    option(TABLE_NAME.key, tableName).
+    option(TBL_NAME.key, tableName).
     mode(Append).
     save(basePath)
 
diff --git a/website/versioned_docs/version-0.14.1/schema_evolution.md 
b/website/versioned_docs/version-0.14.1/schema_evolution.md
index ca086b6dfe0..e8cfc53bff2 100755
--- a/website/versioned_docs/version-0.14.1/schema_evolution.md
+++ b/website/versioned_docs/version-0.14.1/schema_evolution.md
@@ -207,11 +207,11 @@ val data1 = Seq(Row("row_1", "part_0", 0L, "bob", "v_0", 
0),
 var dfFromData1 = spark.createDataFrame(data1, schema)
 dfFromData1.write.format("hudi").
    options(getQuickstartWriteConfigs).
-   option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-   option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-   option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+   option(PRECOMBINE_FIELD.key, "preComb").
+   option(RECORDKEY_FIELD.key, "rowId").
+   option(PARTITIONPATH_FIELD.key, "partitionId").
    option("hoodie.index.type","SIMPLE").
-   option(TABLE_NAME.key, tableName).
+   option(TBL_NAME.key, tableName).
    mode(Overwrite).
    save(basePath)
 
@@ -266,11 +266,11 @@ val data2 = Seq(Row("row_2", "part_0", 5L, "john", "v_3", 
3L, "newField_1"),
 var dfFromData2 = spark.createDataFrame(data2, newSchema)
 dfFromData2.write.format("hudi").
     options(getQuickstartWriteConfigs).
-    option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
-    option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
-    option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+    option(PRECOMBINE_FIELD.key, "preComb").
+    option(RECORDKEY_FIELD.key, "rowId").
+    option(PARTITIONPATH_FIELD.key, "partitionId").
     option("hoodie.index.type","SIMPLE").
-    option(TABLE_NAME.key, tableName).
+    option(TBL_NAME.key, tableName).
     mode(Append).
     save(basePath)
 

Reply via email to