This is an automated email from the ASF dual-hosted git repository.
yihua pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/asf-site by this push:
new 11f7a2eed6e [DOCS] Fix dataframe write option in schema_evolution docs
(#7481)
11f7a2eed6e is described below
commit 11f7a2eed6e92e0c727315d9385f545c99c35250
Author: Zouxxyy <[email protected]>
AuthorDate: Mon Mar 4 07:43:43 2024 +0800
[DOCS] Fix dataframe write option in schema_evolution docs (#7481)
Co-authored-by: Y Ethan Guo <[email protected]>
---
website/docs/schema_evolution.md | 16 ++++++++--------
.../versioned_docs/version-0.10.0/schema_evolution.md | 16 ++++++++--------
.../versioned_docs/version-0.10.1/schema_evolution.md | 16 ++++++++--------
.../versioned_docs/version-0.11.0/schema_evolution.md | 16 ++++++++--------
.../versioned_docs/version-0.11.1/schema_evolution.md | 16 ++++++++--------
.../versioned_docs/version-0.12.0/schema_evolution.md | 16 ++++++++--------
.../versioned_docs/version-0.12.1/schema_evolution.md | 16 ++++++++--------
.../versioned_docs/version-0.12.2/schema_evolution.md | 16 ++++++++--------
.../versioned_docs/version-0.13.0/schema_evolution.md | 16 ++++++++--------
.../versioned_docs/version-0.13.1/schema_evolution.md | 16 ++++++++--------
.../versioned_docs/version-0.14.0/schema_evolution.md | 16 ++++++++--------
.../versioned_docs/version-0.14.1/schema_evolution.md | 16 ++++++++--------
12 files changed, 96 insertions(+), 96 deletions(-)
diff --git a/website/docs/schema_evolution.md b/website/docs/schema_evolution.md
index ca086b6dfe0..e8cfc53bff2 100755
--- a/website/docs/schema_evolution.md
+++ b/website/docs/schema_evolution.md
@@ -207,11 +207,11 @@ val data1 = Seq(Row("row_1", "part_0", 0L, "bob", "v_0",
0),
var dfFromData1 = spark.createDataFrame(data1, schema)
dfFromData1.write.format("hudi").
options(getQuickstartWriteConfigs).
- option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ option(PRECOMBINE_FIELD.key, "preComb").
+ option(RECORDKEY_FIELD.key, "rowId").
+ option(PARTITIONPATH_FIELD.key, "partitionId").
option("hoodie.index.type","SIMPLE").
- option(TABLE_NAME.key, tableName).
+ option(TBL_NAME.key, tableName).
mode(Overwrite).
save(basePath)
@@ -266,11 +266,11 @@ val data2 = Seq(Row("row_2", "part_0", 5L, "john", "v_3",
3L, "newField_1"),
var dfFromData2 = spark.createDataFrame(data2, newSchema)
dfFromData2.write.format("hudi").
options(getQuickstartWriteConfigs).
- option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ option(PRECOMBINE_FIELD.key, "preComb").
+ option(RECORDKEY_FIELD.key, "rowId").
+ option(PARTITIONPATH_FIELD.key, "partitionId").
option("hoodie.index.type","SIMPLE").
- option(TABLE_NAME.key, tableName).
+ option(TBL_NAME.key, tableName).
mode(Append).
save(basePath)
diff --git a/website/versioned_docs/version-0.10.0/schema_evolution.md
b/website/versioned_docs/version-0.10.0/schema_evolution.md
index bd80f0d290c..481ed236f52 100644
--- a/website/versioned_docs/version-0.10.0/schema_evolution.md
+++ b/website/versioned_docs/version-0.10.0/schema_evolution.md
@@ -87,11 +87,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob",
"v_0", 0),
scala> var dfFromData1 = spark.createDataFrame(data1, schema)
scala> dfFromData1.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Overwrite).
| save(basePath)
@@ -147,11 +147,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john",
"v_3", 3L, "newField_1
scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
scala> dfFromData2.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Append).
| save(basePath)
diff --git a/website/versioned_docs/version-0.10.1/schema_evolution.md
b/website/versioned_docs/version-0.10.1/schema_evolution.md
index bd80f0d290c..481ed236f52 100644
--- a/website/versioned_docs/version-0.10.1/schema_evolution.md
+++ b/website/versioned_docs/version-0.10.1/schema_evolution.md
@@ -87,11 +87,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob",
"v_0", 0),
scala> var dfFromData1 = spark.createDataFrame(data1, schema)
scala> dfFromData1.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Overwrite).
| save(basePath)
@@ -147,11 +147,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john",
"v_3", 3L, "newField_1
scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
scala> dfFromData2.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Append).
| save(basePath)
diff --git a/website/versioned_docs/version-0.11.0/schema_evolution.md
b/website/versioned_docs/version-0.11.0/schema_evolution.md
index 06b755d71e4..fd4521e5f0d 100755
--- a/website/versioned_docs/version-0.11.0/schema_evolution.md
+++ b/website/versioned_docs/version-0.11.0/schema_evolution.md
@@ -265,11 +265,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob",
"v_0", 0),
scala> var dfFromData1 = spark.createDataFrame(data1, schema)
scala> dfFromData1.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Overwrite).
| save(basePath)
@@ -325,11 +325,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john",
"v_3", 3L, "newField_1
scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
scala> dfFromData2.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Append).
| save(basePath)
diff --git a/website/versioned_docs/version-0.11.1/schema_evolution.md
b/website/versioned_docs/version-0.11.1/schema_evolution.md
index 06b755d71e4..fd4521e5f0d 100755
--- a/website/versioned_docs/version-0.11.1/schema_evolution.md
+++ b/website/versioned_docs/version-0.11.1/schema_evolution.md
@@ -265,11 +265,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob",
"v_0", 0),
scala> var dfFromData1 = spark.createDataFrame(data1, schema)
scala> dfFromData1.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Overwrite).
| save(basePath)
@@ -325,11 +325,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john",
"v_3", 3L, "newField_1
scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
scala> dfFromData2.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Append).
| save(basePath)
diff --git a/website/versioned_docs/version-0.12.0/schema_evolution.md
b/website/versioned_docs/version-0.12.0/schema_evolution.md
index ed4a1f7d539..d1cac665a59 100755
--- a/website/versioned_docs/version-0.12.0/schema_evolution.md
+++ b/website/versioned_docs/version-0.12.0/schema_evolution.md
@@ -266,11 +266,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob",
"v_0", 0),
scala> var dfFromData1 = spark.createDataFrame(data1, schema)
scala> dfFromData1.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Overwrite).
| save(basePath)
@@ -326,11 +326,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john",
"v_3", 3L, "newField_1
scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
scala> dfFromData2.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Append).
| save(basePath)
diff --git a/website/versioned_docs/version-0.12.1/schema_evolution.md
b/website/versioned_docs/version-0.12.1/schema_evolution.md
index ed4a1f7d539..d1cac665a59 100755
--- a/website/versioned_docs/version-0.12.1/schema_evolution.md
+++ b/website/versioned_docs/version-0.12.1/schema_evolution.md
@@ -266,11 +266,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob",
"v_0", 0),
scala> var dfFromData1 = spark.createDataFrame(data1, schema)
scala> dfFromData1.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Overwrite).
| save(basePath)
@@ -326,11 +326,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john",
"v_3", 3L, "newField_1
scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
scala> dfFromData2.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Append).
| save(basePath)
diff --git a/website/versioned_docs/version-0.12.2/schema_evolution.md
b/website/versioned_docs/version-0.12.2/schema_evolution.md
index ed4a1f7d539..d1cac665a59 100755
--- a/website/versioned_docs/version-0.12.2/schema_evolution.md
+++ b/website/versioned_docs/version-0.12.2/schema_evolution.md
@@ -266,11 +266,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob",
"v_0", 0),
scala> var dfFromData1 = spark.createDataFrame(data1, schema)
scala> dfFromData1.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Overwrite).
| save(basePath)
@@ -326,11 +326,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john",
"v_3", 3L, "newField_1
scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
scala> dfFromData2.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Append).
| save(basePath)
diff --git a/website/versioned_docs/version-0.13.0/schema_evolution.md
b/website/versioned_docs/version-0.13.0/schema_evolution.md
index ed4a1f7d539..d1cac665a59 100755
--- a/website/versioned_docs/version-0.13.0/schema_evolution.md
+++ b/website/versioned_docs/version-0.13.0/schema_evolution.md
@@ -266,11 +266,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob",
"v_0", 0),
scala> var dfFromData1 = spark.createDataFrame(data1, schema)
scala> dfFromData1.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Overwrite).
| save(basePath)
@@ -326,11 +326,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john",
"v_3", 3L, "newField_1
scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
scala> dfFromData2.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Append).
| save(basePath)
diff --git a/website/versioned_docs/version-0.13.1/schema_evolution.md
b/website/versioned_docs/version-0.13.1/schema_evolution.md
index ed4a1f7d539..d1cac665a59 100755
--- a/website/versioned_docs/version-0.13.1/schema_evolution.md
+++ b/website/versioned_docs/version-0.13.1/schema_evolution.md
@@ -266,11 +266,11 @@ scala> val data1 = Seq(Row("row_1", "part_0", 0L, "bob",
"v_0", 0),
scala> var dfFromData1 = spark.createDataFrame(data1, schema)
scala> dfFromData1.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Overwrite).
| save(basePath)
@@ -326,11 +326,11 @@ scala> val data2 = Seq(Row("row_2", "part_0", 5L, "john",
"v_3", 3L, "newField_1
scala> var dfFromData2 = spark.createDataFrame(data2, newSchema)
scala> dfFromData2.write.format("hudi").
| options(getQuickstartWriteConfigs).
- | option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- | option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- | option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ | option(PRECOMBINE_FIELD.key, "preComb").
+ | option(RECORDKEY_FIELD.key, "rowId").
+ | option(PARTITIONPATH_FIELD.key, "partitionId").
| option("hoodie.index.type","SIMPLE").
- | option(TABLE_NAME.key, tableName).
+ | option(TBL_NAME.key, tableName).
| mode(Append).
| save(basePath)
diff --git a/website/versioned_docs/version-0.14.0/schema_evolution.md
b/website/versioned_docs/version-0.14.0/schema_evolution.md
index 8fe04d65238..d01723f496d 100755
--- a/website/versioned_docs/version-0.14.0/schema_evolution.md
+++ b/website/versioned_docs/version-0.14.0/schema_evolution.md
@@ -190,11 +190,11 @@ val data1 = Seq(Row("row_1", "part_0", 0L, "bob", "v_0",
0),
var dfFromData1 = spark.createDataFrame(data1, schema)
dfFromData1.write.format("hudi").
options(getQuickstartWriteConfigs).
- option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ option(PRECOMBINE_FIELD.key, "preComb").
+ option(RECORDKEY_FIELD.key, "rowId").
+ option(PARTITIONPATH_FIELD.key, "partitionId").
option("hoodie.index.type","SIMPLE").
- option(TABLE_NAME.key, tableName).
+ option(TBL_NAME.key, tableName).
mode(Overwrite).
save(basePath)
@@ -249,11 +249,11 @@ val data2 = Seq(Row("row_2", "part_0", 5L, "john", "v_3",
3L, "newField_1"),
var dfFromData2 = spark.createDataFrame(data2, newSchema)
dfFromData2.write.format("hudi").
options(getQuickstartWriteConfigs).
- option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ option(PRECOMBINE_FIELD.key, "preComb").
+ option(RECORDKEY_FIELD.key, "rowId").
+ option(PARTITIONPATH_FIELD.key, "partitionId").
option("hoodie.index.type","SIMPLE").
- option(TABLE_NAME.key, tableName).
+ option(TBL_NAME.key, tableName).
mode(Append).
save(basePath)
diff --git a/website/versioned_docs/version-0.14.1/schema_evolution.md
b/website/versioned_docs/version-0.14.1/schema_evolution.md
index ca086b6dfe0..e8cfc53bff2 100755
--- a/website/versioned_docs/version-0.14.1/schema_evolution.md
+++ b/website/versioned_docs/version-0.14.1/schema_evolution.md
@@ -207,11 +207,11 @@ val data1 = Seq(Row("row_1", "part_0", 0L, "bob", "v_0",
0),
var dfFromData1 = spark.createDataFrame(data1, schema)
dfFromData1.write.format("hudi").
options(getQuickstartWriteConfigs).
- option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ option(PRECOMBINE_FIELD.key, "preComb").
+ option(RECORDKEY_FIELD.key, "rowId").
+ option(PARTITIONPATH_FIELD.key, "partitionId").
option("hoodie.index.type","SIMPLE").
- option(TABLE_NAME.key, tableName).
+ option(TBL_NAME.key, tableName).
mode(Overwrite).
save(basePath)
@@ -266,11 +266,11 @@ val data2 = Seq(Row("row_2", "part_0", 5L, "john", "v_3",
3L, "newField_1"),
var dfFromData2 = spark.createDataFrame(data2, newSchema)
dfFromData2.write.format("hudi").
options(getQuickstartWriteConfigs).
- option(PRECOMBINE_FIELD_OPT_KEY.key, "preComb").
- option(RECORDKEY_FIELD_OPT_KEY.key, "rowId").
- option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionId").
+ option(PRECOMBINE_FIELD.key, "preComb").
+ option(RECORDKEY_FIELD.key, "rowId").
+ option(PARTITIONPATH_FIELD.key, "partitionId").
option("hoodie.index.type","SIMPLE").
- option(TABLE_NAME.key, tableName).
+ option(TBL_NAME.key, tableName).
mode(Append).
save(basePath)