This is an automated email from the ASF dual-hosted git repository. chenliang613 pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/carbondata.git
commit 42d59be0be3eb58b26cb50ed16b4f2b2dd43cfb1 Author: bieremayi <[email protected]> AuthorDate: Tue Dec 7 17:06:29 2021 +0800 Revert "remove useless numerical value , revert some typo issues" This reverts commit 7af81adbf6800655902754c9b1dac5c1fb263f9d. --- docs/addsegment-guide.md | 24 ++++++++++++------------ docs/faq.md | 6 +++--- 2 files changed, 15 insertions(+), 15 deletions(-) diff --git a/docs/addsegment-guide.md b/docs/addsegment-guide.md index 619293f..3985beb 100644 --- a/docs/addsegment-guide.md +++ b/docs/addsegment-guide.md @@ -40,7 +40,7 @@ Users can add the existing data as a segment to the carbon table provided the sc | ------------------------------------------------------------ | ------------------------------------------------------------ | | [PATH](#path) | User external old table path | | [FORMAT](#format) | User external old table file format | -| [PARTITION](#partition) | Partition info for partition table , should be form of "a:int, b:string" | +| [PARTITION](#partition) | Extract partition info for partition table , should be form of "a:int, b:string" | - @@ -60,7 +60,7 @@ Users can add the existing data as a segment to the carbon table provided the sc OPTIONS('FORMAT'='parquet') ``` - ##### PARTITION: - Partition info for partition table , should be form of "a:int, b:string" + Extract partition info for partition table , should be form of "a:int, b:string" ``` OPTIONS('PARTITION'='a:int, b:string') @@ -95,8 +95,8 @@ LOCATION 'hdfs://bieremayi/user/hive/warehouse/log_parquet_par'; Parquet File Location : ``` -/user/hive/warehouse/log_parquet_par/day=20211123/hour=12/type=0 -/user/hive/warehouse/log_parquet_par/day=20211123/hour=12/type=1 +25.1 K 75.2 K /user/hive/warehouse/log_parquet_par/day=20211123/hour=12/type=0 +8.7 K 26.2 K /user/hive/warehouse/log_parquet_par/day=20211123/hour=12/type=1 ``` @@ -117,8 +117,8 @@ LOCATION 'hdfs://bieremayi/user/hive/warehouse/log_orc_par'; Orc File Location : ``` -/user/hive/warehouse/log_orc_par/day=20211123/hour=12/type=0 -/user/hive/warehouse/log_orc_par/day=20211123/hour=12/type=1 +19.4 K 58.2 K /user/hive/warehouse/log_orc_par/day=20211123/hour=12/type=0 +7.1 K 21.4 K /user/hive/warehouse/log_orc_par/day=20211123/hour=12/type=1 ``` **Steps:** @@ -156,13 +156,13 @@ step3: Result check. SHOW SEGMENTS FOR TABLE default.log_carbon_par; ``` -| ID | Status | Load Start Time | Load Time Taken | Partition | Data Size | Index Size | File Format | +|SegmentSequenceId|Status|Load Time|Elapsed Time |Partition|Data Size|Index Size|File Format| |:---|:---|:---|:---|:---|:---|:---|:---| -|4 |Success |2021-11-29 17:59:40.819 |7.026S |{day=20211123,hour=12,type=1}, ... |xxx| xxx |columnar_v3| -|3 |Success |2021-11-29 16:34:28.106 |0.418S |{day=20211123,hour=12,type=0} |xxx |NA | orc| -|2 |Success |2021-11-29 16:34:27.733 |0.222S |{day=20211123,hour=12,type=1} |xxx |NA | orc| -|1 |Success |2021-11-29 16:30:17.207 |0.275S |{day=20211123,hour=12,type=0} |xxx |NA | parquet| -|0 |Success |2021-11-29 16:30:16.48 |0.452S |{day=20211123,hour=12,type=1} |xxx |NA | parquet| +|4 |Success |2021-11-29 17:59:40.819 |7.026S |{day=20211123,hour=12,type=1}, ... |29.86KB| 4.80KB |columnar_v3| +|3 |Success |2021-11-29 16:34:28.106 |0.418S |{day=20211123,hour=12,type=0} |19.40KB |NA | orc| +|2 |Success |2021-11-29 16:34:27.733 |0.222S |{day=20211123,hour=12,type=1} |7.12KB |NA | orc| +|1 |Success |2021-11-29 16:30:17.207 |0.275S |{day=20211123,hour=12,type=0} |25.08KB |NA | parquet| +|0 |Success |2021-11-29 16:30:16.48 |0.452S |{day=20211123,hour=12,type=1} |8.72KB |NA | parquet| ### Changes to tablestatus file diff --git a/docs/faq.md b/docs/faq.md index 304e826..4daf2f5 100644 --- a/docs/faq.md +++ b/docs/faq.md @@ -29,7 +29,7 @@ * [Why different time zone result for select query output when query SDK writer output?](#why-different-time-zone-result-for-select-query-output-when-query-sdk-writer-output) * [How to check LRU cache memory footprint?](#how-to-check-lru-cache-memory-footprint) * [How to deal with the trailing task in query?](#How-to-deal-with-the-trailing-task-in-query) -* [How to manage hybrid file format in carbondata table?](#How-to-manage-hybrid-file-format-in-carbondata-table) +* [How to manage mix file format in carbondata table?](#How-to-manage-mix-file-format-in-carbondata-table) # TroubleShooting @@ -214,9 +214,9 @@ spark.speculation is a group of configuration, that can monitor trailing tasks a spark.blacklist.enabled, avoid reduction of available executors due to blacklist mechanism. -## How to manage hybrid file format in carbondata table +## How to manage mix file format in carbondata table -Refer [Heterogeneous format segments in carbondata](./addsegment-guide.md) +[Heterogeneous format segments in carbondata](./addsegment-guide.md) ## Getting tablestatus.lock issues When loading data
