This is an automated email from the ASF dual-hosted git repository.
danny0405 pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/asf-site by this push:
new 4960e61f6ca Revert [HUDI-6438] Config parameter
'MAKE_NEW_COLUMNS_NULLABLE' (#10783)
4960e61f6ca is described below
commit 4960e61f6ca25a6f636b9256542bb3594583eb65
Author: Geser Dugarov, PhD <[email protected]>
AuthorDate: Fri Mar 1 07:52:46 2024 +0700
Revert [HUDI-6438] Config parameter 'MAKE_NEW_COLUMNS_NULLABLE' (#10783)
---
website/docs/configurations.md | 2 --
1 file changed, 2 deletions(-)
diff --git a/website/docs/configurations.md b/website/docs/configurations.md
index e52f0a52a75..0f0ba7df7e7 100644
--- a/website/docs/configurations.md
+++ b/website/docs/configurations.md
@@ -161,7 +161,6 @@ Options useful for writing tables via
`write.format.option(...)`
|
[hoodie.datasource.write.insert.drop.duplicates](#hoodiedatasourcewriteinsertdropduplicates)
| false
| If set to true, records from the incoming
dataframe will not overwrite existing records with the same key during the
write operation. <br /> **Note** Just for Insert operation in Spark SQL writing
since 0.14.0, users can switch to the config
`hoodie.datasource.insert.dup.policy` [...]
|
[hoodie.datasource.write.keygenerator.class](#hoodiedatasourcewritekeygeneratorclass)
|
org.apache.hudi.keygen.SimpleKeyGenerator | Key generator
class, that implements `org.apache.hudi.keygen.KeyGenerator`<br />`Config
Param: KEYGENERATOR_CLASS_NAME`
[...]
|
[hoodie.datasource.write.keygenerator.consistent.logical.timestamp.enabled](#hoodiedatasourcewritekeygeneratorconsistentlogicaltimestampenabled)
| false | When set to
true, consistent value will be generated for a logical timestamp type column,
like timestamp-millis and timestamp-micros, irrespective of whether row-writer
is enabled. Disabled by default so as not to break the pipeline that deploy
either fully row-writer path or non [...]
-|
[hoodie.datasource.write.new.columns.nullable](#hoodiedatasourcewritenewcolumnsnullable)
| false
| When a non-nullable column is added
to datasource during a write operation, the write operation will fail schema
compatibility check. Set this option to true will make the newly added column
nullable to successfully complete the write operation.<br />`Config Param:
MAKE_NEW_COL [...]
|
[hoodie.datasource.write.partitionpath.urlencode](#hoodiedatasourcewritepartitionpathurlencode)
| false
| Should we url encode the partition path
value, before creating the folder structure.<br />`Config Param:
URL_ENCODE_PARTITIONING`
[...]
| [hoodie.datasource.write.payload.class](#hoodiedatasourcewritepayloadclass)
|
org.apache.hudi.common.model.OverwriteWithLatestAvroPayload | Payload class
used. Override this, if you like to roll your own merge logic, when
upserting/inserting. This will render any value set for
PRECOMBINE_FIELD_OPT_VAL in-effective<br />`Config Param: PAYLOAD_CLASS_NAME`
[...]
|
[hoodie.datasource.write.reconcile.schema](#hoodiedatasourcewritereconcileschema)
| false
| This config controls how
writer's schema will be selected based on the incoming batch's schema as well
as existing table's one. When schema reconciliation is DISABLED, incoming
batch's schema will be picked as a writer-schema (therefore updating table's
schema). When schema recon [...]
@@ -359,7 +358,6 @@ The following set of configurations are common across Hudi.
| [hoodie.memory.compaction.max.size](#hoodiememorycompactionmaxsize)
| (N/A) | Maximum amount of memory used in bytes for
compaction operations in bytes , before spilling to local storage.<br />`Config
Param: MAX_MEMORY_FOR_COMPACTION`
[...]
|
[hoodie.common.diskmap.compression.enabled](#hoodiecommondiskmapcompressionenabled)
| true | Turn on compression for BITCASK disk map used by the
External Spillable Map<br />`Config Param:
DISK_MAP_BITCASK_COMPRESSION_ENABLED`
[...]
| [hoodie.common.spillable.diskmap.type](#hoodiecommonspillablediskmaptype)
| BITCASK | When handling input data that cannot be held in
memory, to merge with a file on storage, a spillable diskmap is employed. By
default, we use a persistent hashmap based loosely on bitcask, that offers O(1)
inserts, lookups. Change this to `ROCKS_DB` to prefer using rocksDB, for
handling the spill.<br />`Config Param: SPILLABLE_DISK_MAP_TYPE`
[...]
-|
[hoodie.datasource.write.new.columns.nullable](#hoodiedatasourcewritenewcolumnsnullable)
| false | When a non-nullable column is added to datasource during a write
operation, the write operation will fail schema compatibility check. Set this
option to true will make the newly added column nullable to successfully
complete the write operation.<br />`Config Param: MAKE_NEW_COLUMNS_NULLABLE`<br
/>`Since Version: 0.14.0`
[...]
|
[hoodie.datasource.write.reconcile.schema](#hoodiedatasourcewritereconcileschema)
| false | This config controls how writer's schema will be selected
based on the incoming batch's schema as well as existing table's one. When
schema reconciliation is DISABLED, incoming batch's schema will be picked as a
writer-schema (therefore updating table's schema). When schema reconciliation
is ENABLED, writer-schema will be picked such that table's schema (after txn)
is either kept the [...]
| [hoodie.fs.atomic_creation.support](#hoodiefsatomic_creationsupport)
| | This config is used to specify the file system which
supports atomic file creation . atomic means that an operation either succeeds
and has an effect or has fails and has no effect; now this feature is used by
FileSystemLockProvider to guaranteeing that only one writer can create the lock
file at a time. since some FS does not support atomic file creation (eg: S3),
we decide the FileS [...]
| [hoodie.memory.dfs.buffer.max.size](#hoodiememorydfsbuffermaxsize)
| 16777216 | Property to control the max memory in bytes for dfs
input stream buffer size<br />`Config Param: MAX_DFS_STREAM_BUFFER_SIZE`
[...]