This is an automated email from the ASF dual-hosted git repository. chenliang613 pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/carbondata.git
commit 7af81adbf6800655902754c9b1dac5c1fb263f9d Author: bieremayi <[email protected]> AuthorDate: Sat Dec 4 16:42:07 2021 +0800 remove useless numerical value , revert some typo issues --- docs/addsegment-guide.md | 24 ++++++++++++------------ docs/faq.md | 6 +++--- 2 files changed, 15 insertions(+), 15 deletions(-) diff --git a/docs/addsegment-guide.md b/docs/addsegment-guide.md index 3985beb..619293f 100644 --- a/docs/addsegment-guide.md +++ b/docs/addsegment-guide.md @@ -40,7 +40,7 @@ Users can add the existing data as a segment to the carbon table provided the sc | ------------------------------------------------------------ | ------------------------------------------------------------ | | [PATH](#path) | User external old table path | | [FORMAT](#format) | User external old table file format | -| [PARTITION](#partition) | Extract partition info for partition table , should be form of "a:int, b:string" | +| [PARTITION](#partition) | Partition info for partition table , should be form of "a:int, b:string" | - @@ -60,7 +60,7 @@ Users can add the existing data as a segment to the carbon table provided the sc OPTIONS('FORMAT'='parquet') ``` - ##### PARTITION: - Extract partition info for partition table , should be form of "a:int, b:string" + Partition info for partition table , should be form of "a:int, b:string" ``` OPTIONS('PARTITION'='a:int, b:string') @@ -95,8 +95,8 @@ LOCATION 'hdfs://bieremayi/user/hive/warehouse/log_parquet_par'; Parquet File Location : ``` -25.1 K 75.2 K /user/hive/warehouse/log_parquet_par/day=20211123/hour=12/type=0 -8.7 K 26.2 K /user/hive/warehouse/log_parquet_par/day=20211123/hour=12/type=1 +/user/hive/warehouse/log_parquet_par/day=20211123/hour=12/type=0 +/user/hive/warehouse/log_parquet_par/day=20211123/hour=12/type=1 ``` @@ -117,8 +117,8 @@ LOCATION 'hdfs://bieremayi/user/hive/warehouse/log_orc_par'; Orc File Location : ``` -19.4 K 58.2 K /user/hive/warehouse/log_orc_par/day=20211123/hour=12/type=0 -7.1 K 21.4 K /user/hive/warehouse/log_orc_par/day=20211123/hour=12/type=1 +/user/hive/warehouse/log_orc_par/day=20211123/hour=12/type=0 +/user/hive/warehouse/log_orc_par/day=20211123/hour=12/type=1 ``` **Steps:** @@ -156,13 +156,13 @@ step3: Result check. SHOW SEGMENTS FOR TABLE default.log_carbon_par; ``` -|SegmentSequenceId|Status|Load Time|Elapsed Time |Partition|Data Size|Index Size|File Format| +| ID | Status | Load Start Time | Load Time Taken | Partition | Data Size | Index Size | File Format | |:---|:---|:---|:---|:---|:---|:---|:---| -|4 |Success |2021-11-29 17:59:40.819 |7.026S |{day=20211123,hour=12,type=1}, ... |29.86KB| 4.80KB |columnar_v3| -|3 |Success |2021-11-29 16:34:28.106 |0.418S |{day=20211123,hour=12,type=0} |19.40KB |NA | orc| -|2 |Success |2021-11-29 16:34:27.733 |0.222S |{day=20211123,hour=12,type=1} |7.12KB |NA | orc| -|1 |Success |2021-11-29 16:30:17.207 |0.275S |{day=20211123,hour=12,type=0} |25.08KB |NA | parquet| -|0 |Success |2021-11-29 16:30:16.48 |0.452S |{day=20211123,hour=12,type=1} |8.72KB |NA | parquet| +|4 |Success |2021-11-29 17:59:40.819 |7.026S |{day=20211123,hour=12,type=1}, ... |xxx| xxx |columnar_v3| +|3 |Success |2021-11-29 16:34:28.106 |0.418S |{day=20211123,hour=12,type=0} |xxx |NA | orc| +|2 |Success |2021-11-29 16:34:27.733 |0.222S |{day=20211123,hour=12,type=1} |xxx |NA | orc| +|1 |Success |2021-11-29 16:30:17.207 |0.275S |{day=20211123,hour=12,type=0} |xxx |NA | parquet| +|0 |Success |2021-11-29 16:30:16.48 |0.452S |{day=20211123,hour=12,type=1} |xxx |NA | parquet| ### Changes to tablestatus file diff --git a/docs/faq.md b/docs/faq.md index 4daf2f5..304e826 100644 --- a/docs/faq.md +++ b/docs/faq.md @@ -29,7 +29,7 @@ * [Why different time zone result for select query output when query SDK writer output?](#why-different-time-zone-result-for-select-query-output-when-query-sdk-writer-output) * [How to check LRU cache memory footprint?](#how-to-check-lru-cache-memory-footprint) * [How to deal with the trailing task in query?](#How-to-deal-with-the-trailing-task-in-query) -* [How to manage mix file format in carbondata table?](#How-to-manage-mix-file-format-in-carbondata-table) +* [How to manage hybrid file format in carbondata table?](#How-to-manage-hybrid-file-format-in-carbondata-table) # TroubleShooting @@ -214,9 +214,9 @@ spark.speculation is a group of configuration, that can monitor trailing tasks a spark.blacklist.enabled, avoid reduction of available executors due to blacklist mechanism. -## How to manage mix file format in carbondata table +## How to manage hybrid file format in carbondata table -[Heterogeneous format segments in carbondata](./addsegment-guide.md) +Refer [Heterogeneous format segments in carbondata](./addsegment-guide.md) ## Getting tablestatus.lock issues When loading data
