This is an automated email from the ASF dual-hosted git repository.

vinoth pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new 1541e99  [HUDI-3076] - Docs for config file details (#4423)
1541e99 is described below

commit 1541e99fa1535a88d8f540174c8f2b21101ececc
Author: Kyle Weller <[email protected]>
AuthorDate: Thu Jan 6 14:42:04 2022 -0800

    [HUDI-3076] - Docs for config file details (#4423)
    
    * added config file details to docs
    
    * Minor cosmetic tweaks
    
    * Minor cosmetic tweaks
    
    Co-authored-by: vinoth chandar <[email protected]>
---
 website/docs/configurations.md                          | 6 ++++++
 website/versioned_docs/version-0.10.0/configurations.md | 6 ++++++
 2 files changed, 12 insertions(+)

diff --git a/website/docs/configurations.md b/website/docs/configurations.md
index 02363e7..e2cb35b 100644
--- a/website/docs/configurations.md
+++ b/website/docs/configurations.md
@@ -17,6 +17,12 @@ This page covers the different ways of configuring your job 
to write/read Hudi t
 - [**Kafka Connect Configs**](#KAFKA_CONNECT): These set of configs are used 
for Kafka Connect Sink Connector for writing Hudi Tables
 - [**Amazon Web Services Configs**](#AWS): Please fill in the description for 
Config Group Name: Amazon Web Services Configs
 
+## Externalized Config File
+Instead of directly passing configuration settings to every Hudi job, you can 
also centrally set them in a configuration 
+file `hudi-default.conf`. By default, Hudi would load the configuration file 
under `/etc/hudi/conf` directory. You can 
+specify a different configuration directory location by setting the 
`HUDI_CONF_DIR` environment variable. This can be 
+useful for uniformly enforcing repeated configs (like Hive sync or write/index 
tuning), across your entire data lake. 
+
 ## Spark Datasource Configs {#SPARK_DATASOURCE}
 These configs control the Hudi Spark Datasource, providing ability to define 
keys/partitioning, pick out the write operation, specify how to merge records 
or choosing query type to read.
 
diff --git a/website/versioned_docs/version-0.10.0/configurations.md 
b/website/versioned_docs/version-0.10.0/configurations.md
index 02363e7..4521ce7 100644
--- a/website/versioned_docs/version-0.10.0/configurations.md
+++ b/website/versioned_docs/version-0.10.0/configurations.md
@@ -17,6 +17,12 @@ This page covers the different ways of configuring your job 
to write/read Hudi t
 - [**Kafka Connect Configs**](#KAFKA_CONNECT): These set of configs are used 
for Kafka Connect Sink Connector for writing Hudi Tables
 - [**Amazon Web Services Configs**](#AWS): Please fill in the description for 
Config Group Name: Amazon Web Services Configs
 
+## Externalized Config File
+Instead of directly passing configuration settings to every Hudi job, you can 
also centrally set them in a configuration
+file `hudi-default.conf`. By default, Hudi would load the configuration file 
under `/etc/hudi/conf` directory. You can
+specify a different configuration directory location by setting the 
`HUDI_CONF_DIR` environment variable. This can be
+useful for uniformly enforcing repeated configs (like Hive sync or write/index 
tuning), across your entire data lake.
+
 ## Spark Datasource Configs {#SPARK_DATASOURCE}
 These configs control the Hudi Spark Datasource, providing ability to define 
keys/partitioning, pick out the write operation, specify how to merge records 
or choosing query type to read.
 

Reply via email to