This is an automated email from the ASF dual-hosted git repository. vinoth pushed a commit to branch asf-site in repository https://gitbox.apache.org/repos/asf/incubator-hudi.git
The following commit(s) were added to refs/heads/asf-site by this push: new edc13c1 [HUDI-262] Update website to reflect name change in InputFormat class name edc13c1 is described below commit edc13c125f2bebb147fd21eb192b8a50ad568ec8 Author: Bhavani Sudha Saktheeswaran <bhasu...@uber.com> AuthorDate: Wed Sep 18 12:32:40 2019 -0700 [HUDI-262] Update website to reflect name change in InputFormat class name --- docs/README.md | 2 +- docs/querying_data.cn.md | 6 +++--- docs/querying_data.md | 6 +++--- 3 files changed, 7 insertions(+), 7 deletions(-) diff --git a/docs/README.md b/docs/README.md index 4307a6a..74c78e1 100644 --- a/docs/README.md +++ b/docs/README.md @@ -5,7 +5,7 @@ This folder contains resources that build the [Apache Hudi website](https://hudi ### Building docs -The site is based on a [Jekyll](https://jekyllrb.com/) theme hosted [here](idratherbewriting.com/documentation-theme-jekyll/) with detailed instructions. +The site is based on a [Jekyll](https://jekyllrb.com/) theme hosted [here](https://idratherbewriting.com/documentation-theme-jekyll/) with detailed instructions. #### Docker diff --git a/docs/querying_data.cn.md b/docs/querying_data.cn.md index 3a6fd0f..1653b08 100644 --- a/docs/querying_data.cn.md +++ b/docs/querying_data.cn.md @@ -14,8 +14,8 @@ bundle has been provided, the dataset can be queried by popular query engines li Specifically, there are two Hive tables named off [table name](configurations.html#TABLE_NAME_OPT_KEY) passed during write. For e.g, if `table name = hudi_tbl`, then we get - - `hudi_tbl` realizes the read optimized view of the dataset backed by `HoodieInputFormat`, exposing purely columnar data. - - `hudi_tbl_rt` realizes the real time view of the dataset backed by `HoodieRealtimeInputFormat`, exposing merged view of base and log data. + - `hudi_tbl` realizes the read optimized view of the dataset backed by `HoodieParquetInputFormat`, exposing purely columnar data. + - `hudi_tbl_rt` realizes the real time view of the dataset backed by `HoodieParquetRealtimeInputFormat`, exposing merged view of base and log data. As discussed in the concepts section, the one key primitive needed for [incrementally processing](https://www.oreilly.com/ideas/ubers-case-for-incremental-processing-on-hadoop), is `incremental pulls` (to obtain a change stream/log from a dataset). Hudi datasets can be pulled incrementally, which means you can get ALL and ONLY the updated & new rows @@ -33,7 +33,7 @@ classes with its dependencies are available for query planning & execution. ### Read Optimized table {#hive-ro-view} In addition to setup above, for beeline cli access, the `hive.input.format` variable needs to be set to the fully qualified path name of the -inputformat `org.apache.hudi.hadoop.HoodieInputFormat`. For Tez, additionally the `hive.tez.input.format` needs to be set +inputformat `org.apache.hudi.hadoop.HoodieParquetInputFormat`. For Tez, additionally the `hive.tez.input.format` needs to be set to `org.apache.hadoop.hive.ql.io.HiveInputFormat` ### Real time table {#hive-rt-view} diff --git a/docs/querying_data.md b/docs/querying_data.md index 3a6fd0f..1653b08 100644 --- a/docs/querying_data.md +++ b/docs/querying_data.md @@ -14,8 +14,8 @@ bundle has been provided, the dataset can be queried by popular query engines li Specifically, there are two Hive tables named off [table name](configurations.html#TABLE_NAME_OPT_KEY) passed during write. For e.g, if `table name = hudi_tbl`, then we get - - `hudi_tbl` realizes the read optimized view of the dataset backed by `HoodieInputFormat`, exposing purely columnar data. - - `hudi_tbl_rt` realizes the real time view of the dataset backed by `HoodieRealtimeInputFormat`, exposing merged view of base and log data. + - `hudi_tbl` realizes the read optimized view of the dataset backed by `HoodieParquetInputFormat`, exposing purely columnar data. + - `hudi_tbl_rt` realizes the real time view of the dataset backed by `HoodieParquetRealtimeInputFormat`, exposing merged view of base and log data. As discussed in the concepts section, the one key primitive needed for [incrementally processing](https://www.oreilly.com/ideas/ubers-case-for-incremental-processing-on-hadoop), is `incremental pulls` (to obtain a change stream/log from a dataset). Hudi datasets can be pulled incrementally, which means you can get ALL and ONLY the updated & new rows @@ -33,7 +33,7 @@ classes with its dependencies are available for query planning & execution. ### Read Optimized table {#hive-ro-view} In addition to setup above, for beeline cli access, the `hive.input.format` variable needs to be set to the fully qualified path name of the -inputformat `org.apache.hudi.hadoop.HoodieInputFormat`. For Tez, additionally the `hive.tez.input.format` needs to be set +inputformat `org.apache.hudi.hadoop.HoodieParquetInputFormat`. For Tez, additionally the `hive.tez.input.format` needs to be set to `org.apache.hadoop.hive.ql.io.HiveInputFormat` ### Real time table {#hive-rt-view}