This is an automated email from the ASF dual-hosted git repository.
zykkk pushed a commit to branch branch-2.1
in repository https://gitbox.apache.org/repos/asf/doris.git
The following commit(s) were added to refs/heads/branch-2.1 by this push:
new a9f9113c48f [branch-2.1][test](external)move hive cases from p2 to p0
(#37149)
a9f9113c48f is described below
commit a9f9113c48fc76bb0b3db91de1068ee5b3383dc0
Author: zy-kkk <[email protected]>
AuthorDate: Wed Jul 3 19:44:52 2024 +0800
[branch-2.1][test](external)move hive cases from p2 to p0 (#37149)
pk (#36855)
test_hive_same_db_table_name
test_hive_special_char_partition
test_complex_types
test_wide_table
---
.../test_complex_types/create_table.hql | 138 +++++
.../data/multi_catalog/test_complex_types/run.sh | 22 +
.../test_hive_same_db_table_name/create_table.hql | 22 +
.../test_hive_same_db_table_name/data.tar.gz | Bin 0 -> 183 bytes
.../test_hive_same_db_table_name/run.sh | 12 +
.../create_table.hql | 22 +
.../test_hive_special_char_partition/data.tar.gz | Bin 0 -> 894 bytes
.../test_hive_special_char_partition/run.sh | 12 +
.../multi_catalog/test_wide_table/create_table.hql | 654 +++++++++++++++++++++
.../data/multi_catalog/test_wide_table/run.sh | 22 +
.../hive/test_complex_types.out | 54 +-
.../hive/test_hive_same_db_table_name.out | 15 +
.../hive/test_hive_special_char_partition.out | 50 ++
.../external_table_p0/hive/test_wide_table.out | 45 ++
.../external_table_p2/hive/test_wide_table.out | 23 -
.../hive/test_complex_types.groovy | 39 +-
.../hive/test_hive_same_db_table_name.groovy | 31 +-
.../hive/test_hive_special_char_partition.groovy | 31 +-
.../hive/test_wide_table.groovy | 33 +-
19 files changed, 1132 insertions(+), 93 deletions(-)
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/create_table.hql
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/create_table.hql
new file mode 100644
index 00000000000..f22d7ff246b
--- /dev/null
+++
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/create_table.hql
@@ -0,0 +1,138 @@
+create database if not exists multi_catalog;
+
+use multi_catalog;
+
+CREATE TABLE `byd`(
+ `id` int,
+ `capacity` array<double>,
+ `singles` map<string,double>,
+ `favor` struct<name:string,age:int,tip:timestamp>)
+ROW FORMAT SERDE
+ 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
+STORED AS INPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
+OUTPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
+LOCATION
+ '/user/doris/suites/multi_catalog/byd'
+TBLPROPERTIES (
+ 'transient_lastDdlTime'='1690356922');
+
+msck repair table byd;
+
+CREATE TABLE `complex_offsets_check`(
+ `id` int,
+ `array1` array<int>,
+ `array2` array<array<int>>,
+ `map1` map<string,int>,
+ `struct1` struct<s1:string,s2:int>)
+ROW FORMAT SERDE
+ 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
+STORED AS INPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
+OUTPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
+LOCATION
+ '/user/doris/suites/multi_catalog/complex_offsets_check'
+TBLPROPERTIES (
+ 'transient_lastDdlTime'='1690974653');
+
+msck repair table complex_offsets_check;
+
+CREATE TABLE `parquet_all_types`(
+ `t_null_string` string,
+ `t_null_varchar` varchar(65535),
+ `t_null_char` char(10),
+ `t_null_decimal_precision_2` decimal(2,1),
+ `t_null_decimal_precision_4` decimal(4,2),
+ `t_null_decimal_precision_8` decimal(8,4),
+ `t_null_decimal_precision_17` decimal(17,8),
+ `t_null_decimal_precision_18` decimal(18,8),
+ `t_null_decimal_precision_38` decimal(38,16),
+ `t_empty_string` string,
+ `t_string` string,
+ `t_empty_varchar` varchar(65535),
+ `t_varchar` varchar(65535),
+ `t_varchar_max_length` varchar(65535),
+ `t_char` char(10),
+ `t_int` int,
+ `t_bigint` bigint,
+ `t_float` float,
+ `t_double` double,
+ `t_boolean_true` boolean,
+ `t_boolean_false` boolean,
+ `t_decimal_precision_2` decimal(2,1),
+ `t_decimal_precision_4` decimal(4,2),
+ `t_decimal_precision_8` decimal(8,4),
+ `t_decimal_precision_17` decimal(17,8),
+ `t_decimal_precision_18` decimal(18,8),
+ `t_decimal_precision_38` decimal(38,16),
+ `t_binary` binary,
+ `t_map_string` map<string,string>,
+ `t_map_varchar` map<varchar(65535),varchar(65535)>,
+ `t_map_char` map<char(10),char(10)>,
+ `t_map_int` map<int,int>,
+ `t_map_bigint` map<bigint,bigint>,
+ `t_map_float` map<float,float>,
+ `t_map_double` map<double,double>,
+ `t_map_boolean` map<boolean,boolean>,
+ `t_map_decimal_precision_2` map<decimal(2,1),decimal(2,1)>,
+ `t_map_decimal_precision_4` map<decimal(4,2),decimal(4,2)>,
+ `t_map_decimal_precision_8` map<decimal(8,4),decimal(8,4)>,
+ `t_map_decimal_precision_17` map<decimal(17,8),decimal(17,8)>,
+ `t_map_decimal_precision_18` map<decimal(18,8),decimal(18,8)>,
+ `t_map_decimal_precision_38` map<decimal(38,16),decimal(38,16)>,
+ `t_array_string` array<string>,
+ `t_array_int` array<int>,
+ `t_array_bigint` array<bigint>,
+ `t_array_float` array<float>,
+ `t_array_double` array<double>,
+ `t_array_boolean` array<boolean>,
+ `t_array_varchar` array<varchar(65535)>,
+ `t_array_char` array<char(10)>,
+ `t_array_decimal_precision_2` array<decimal(2,1)>,
+ `t_array_decimal_precision_4` array<decimal(4,2)>,
+ `t_array_decimal_precision_8` array<decimal(8,4)>,
+ `t_array_decimal_precision_17` array<decimal(17,8)>,
+ `t_array_decimal_precision_18` array<decimal(18,8)>,
+ `t_array_decimal_precision_38` array<decimal(38,16)>,
+ `t_struct_bigint` struct<s_bigint:bigint>,
+ `t_complex` map<string,array<struct<s_int:int>>>,
+ `t_struct_nested` struct<struct_field:array<string>>,
+ `t_struct_null` struct<struct_field_null:string,struct_field_null2:string>,
+ `t_struct_non_nulls_after_nulls`
struct<struct_non_nulls_after_nulls1:int,struct_non_nulls_after_nulls2:string>,
+ `t_nested_struct_non_nulls_after_nulls`
struct<struct_field1:int,struct_field2:string,strict_field3:struct<nested_struct_field1:int,nested_struct_field2:string>>,
+ `t_map_null_value` map<string,string>,
+ `t_array_string_starting_with_nulls` array<string>,
+ `t_array_string_with_nulls_in_between` array<string>,
+ `t_array_string_ending_with_nulls` array<string>,
+ `t_array_string_all_nulls` array<string>)
+ROW FORMAT SERDE
+ 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
+STORED AS INPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
+OUTPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
+LOCATION
+ '/user/doris/suites/multi_catalog/parquet_all_types'
+TBLPROPERTIES (
+ 'transient_lastDdlTime'='1692347490');
+
+msck repair table parquet_all_types;
+
+CREATE TABLE `date_dict`(
+ `date1` date,
+ `date2` date,
+ `date3` date)
+ROW FORMAT SERDE
+ 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
+STORED AS INPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
+OUTPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
+LOCATION
+ '/user/doris/suites/multi_catalog/date_dict'
+TBLPROPERTIES (
+ 'transient_lastDdlTime'='1693396885');
+
+msck repair table date_dict;
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/run.sh
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/run.sh
new file mode 100755
index 00000000000..947fb796eeb
--- /dev/null
+++
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_complex_types/run.sh
@@ -0,0 +1,22 @@
+#!/bin/bash
+set -x
+
+CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)"
+
+if [[ ! -d "${CUR_DIR}/data" ]]; then
+ echo "${CUR_DIR}/data does not exist"
+ cd "${CUR_DIR}" && rm -f data.tar.gz \
+ && curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_complex_types/data.tar.gz
\
+ && tar xzf data.tar.gz
+ cd -
+else
+ echo "${CUR_DIR}/data exist, continue !"
+fi
+
+## mkdir and put data to hdfs
+hadoop fs -mkdir -p /user/doris/suites/multi_catalog/
+hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/
+
+# create table
+hive -f "${CUR_DIR}"/create_table.hql
+
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/create_table.hql
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/create_table.hql
new file mode 100644
index 00000000000..3d672596230
--- /dev/null
+++
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/create_table.hql
@@ -0,0 +1,22 @@
+create database if not exists multi_catalog;
+
+use multi_catalog;
+
+CREATE TABLE `region`(
+ `r_regionkey` int,
+ `r_name` char(25))
+ROW FORMAT SERDE
+ 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
+WITH SERDEPROPERTIES (
+ 'field.delim'='|',
+ 'serialization.format'='|')
+STORED AS INPUTFORMAT
+ 'org.apache.hadoop.mapred.TextInputFormat'
+OUTPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
+LOCATION
+ '/user/doris/suites/multi_catalog/region'
+TBLPROPERTIES (
+ 'transient_lastDdlTime'='1670483235');
+
+msck repair table region;
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/data.tar.gz
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/data.tar.gz
new file mode 100644
index 00000000000..c5367912ddf
Binary files /dev/null and
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/data.tar.gz
differ
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/run.sh
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/run.sh
new file mode 100644
index 00000000000..a52ee0579a2
--- /dev/null
+++
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_same_db_table_name/run.sh
@@ -0,0 +1,12 @@
+#!/bin/bash
+set -x
+
+CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)"
+
+## mkdir and put data to hdfs
+cd "${CUR_DIR}" && rm -rf data/ && tar xzf data.tar.gz
+hadoop fs -mkdir -p /user/doris/suites/multi_catalog/
+hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/
+
+# create table
+hive -f "${CUR_DIR}"/create_table.hql
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/create_table.hql
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/create_table.hql
new file mode 100644
index 00000000000..2631d1360c9
--- /dev/null
+++
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/create_table.hql
@@ -0,0 +1,22 @@
+create database if not exists multi_catalog;
+
+use multi_catalog;
+
+CREATE TABLE `special_character_1_partition`(
+ `name` string)
+PARTITIONED BY (
+ `part` string)
+ROW FORMAT SERDE
+ 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
+STORED AS INPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
+OUTPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
+LOCATION
+ '/user/doris/suites/multi_catalog/special_character_1_partition'
+TBLPROPERTIES (
+ 'transient_lastDdlTime'='1689575322');
+
+set hive.msck.path.validation=ignore;
+
+msck repair table special_character_1_partition;
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/data.tar.gz
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/data.tar.gz
new file mode 100644
index 00000000000..bce0219357e
Binary files /dev/null and
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/data.tar.gz
differ
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/run.sh
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/run.sh
new file mode 100644
index 00000000000..a52ee0579a2
--- /dev/null
+++
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_hive_special_char_partition/run.sh
@@ -0,0 +1,12 @@
+#!/bin/bash
+set -x
+
+CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)"
+
+## mkdir and put data to hdfs
+cd "${CUR_DIR}" && rm -rf data/ && tar xzf data.tar.gz
+hadoop fs -mkdir -p /user/doris/suites/multi_catalog/
+hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/
+
+# create table
+hive -f "${CUR_DIR}"/create_table.hql
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/create_table.hql
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/create_table.hql
new file mode 100644
index 00000000000..e57a9148383
--- /dev/null
+++
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/create_table.hql
@@ -0,0 +1,654 @@
+create database if not exists multi_catalog;
+
+use multi_catalog;
+
+CREATE TABLE `wide_table1_orc`(
+ `col1` decimal(16,0),
+ `col2` string,
+ `col3` string,
+ `col4` string,
+ `col5` string,
+ `col6` string,
+ `col7` string,
+ `col8` string,
+ `col9` string,
+ `col10` string,
+ `col11` string,
+ `col12` string,
+ `col13` string,
+ `col14` string,
+ `col15` decimal(2,0),
+ `col16` decimal(2,0),
+ `col17` decimal(2,0),
+ `col18` decimal(2,0),
+ `col19` decimal(2,0),
+ `col20` decimal(10,0),
+ `col21` string,
+ `col22` decimal(10,0),
+ `col23` string,
+ `col24` decimal(12,0),
+ `col25` string,
+ `col26` decimal(8,0),
+ `col27` decimal(8,0),
+ `col28` decimal(8,0),
+ `col29` decimal(8,0),
+ `col30` decimal(2,0),
+ `col31` string,
+ `col32` decimal(12,0),
+ `col33` string,
+ `col34` decimal(8,0),
+ `col35` decimal(12,0),
+ `col36` string,
+ `col37` decimal(12,0),
+ `col38` string,
+ `col39` decimal(12,0),
+ `col40` string,
+ `col41` int,
+ `col42` int,
+ `col43` string,
+ `col44` decimal(12,0),
+ `col45` string,
+ `col46` int,
+ `col47` decimal(12,0),
+ `col48` string,
+ `col49` decimal(8,0),
+ `col50` string,
+ `col51` string,
+ `col52` string,
+ `col53` string,
+ `col54` string,
+ `col55` string,
+ `col56` string,
+ `col57` string,
+ `col58` string,
+ `col59` string,
+ `col60` string,
+ `col61` string,
+ `col62` string,
+ `col63` string,
+ `col64` decimal(12,0),
+ `col65` string,
+ `col66` string,
+ `col67` string,
+ `col68` decimal(12,0),
+ `col69` string,
+ `col70` decimal(16,2),
+ `col71` decimal(8,0),
+ `col72` string,
+ `col73` decimal(16,2),
+ `col74` decimal(16,2),
+ `col75` decimal(16,2),
+ `col76` decimal(16,2),
+ `col77` string,
+ `col78` string,
+ `col79` string,
+ `col80` decimal(16,2),
+ `col81` decimal(16,2),
+ `col82` decimal(16,2),
+ `col83` decimal(16,2),
+ `col84` decimal(16,2),
+ `col85` decimal(16,2),
+ `col86` decimal(16,2),
+ `col87` decimal(16,2),
+ `col88` decimal(16,2),
+ `col89` decimal(16,2),
+ `col90` decimal(16,2),
+ `col91` decimal(16,2),
+ `col92` decimal(16,2),
+ `col93` decimal(16,2),
+ `col94` decimal(16,2),
+ `col95` decimal(16,2),
+ `col96` decimal(16,2),
+ `col97` decimal(19,6),
+ `col98` decimal(12,5),
+ `col99` decimal(12,5),
+ `col100` decimal(12,5),
+ `col101` decimal(16,2),
+ `col102` decimal(16,2),
+ `col103` decimal(16,2),
+ `col104` decimal(16,2),
+ `col105` decimal(16,2),
+ `col106` decimal(16,2),
+ `col107` decimal(16,2),
+ `col108` decimal(16,2),
+ `col109` decimal(16,2),
+ `col110` decimal(16,2),
+ `col111` decimal(16,2),
+ `col112` decimal(16,2),
+ `col113` decimal(16,2),
+ `col114` decimal(16,2),
+ `col115` decimal(16,2),
+ `col116` decimal(16,2),
+ `col117` decimal(8,0),
+ `col118` decimal(8,0),
+ `col119` decimal(8,0),
+ `col120` decimal(8,0),
+ `col121` decimal(8,0),
+ `col122` decimal(8,0),
+ `col123` decimal(8,0),
+ `col124` decimal(8,0),
+ `col125` decimal(8,0),
+ `col126` decimal(8,0),
+ `col127` decimal(8,0),
+ `col128` decimal(8,0),
+ `col129` decimal(8,0),
+ `col130` decimal(8,0),
+ `col131` decimal(8,0),
+ `col132` decimal(16,2),
+ `col133` decimal(16,2),
+ `col134` decimal(16,2),
+ `col135` decimal(16,2),
+ `col136` decimal(16,2),
+ `col137` decimal(16,2),
+ `col138` decimal(16,2),
+ `col139` decimal(16,2),
+ `col140` decimal(16,2),
+ `col141` decimal(16,2),
+ `col142` decimal(16,2),
+ `col143` decimal(16,2),
+ `col144` decimal(16,2),
+ `col145` decimal(16,2),
+ `col146` decimal(16,2),
+ `col147` decimal(16,2),
+ `col148` decimal(16,2),
+ `col149` decimal(16,2),
+ `col150` decimal(16,2),
+ `col151` decimal(16,2),
+ `col152` decimal(16,2),
+ `col153` decimal(16,2),
+ `col154` decimal(16,2),
+ `col155` decimal(16,2),
+ `col156` decimal(16,2),
+ `col157` decimal(16,2),
+ `col158` decimal(16,2),
+ `col159` decimal(16,2),
+ `col160` decimal(16,2),
+ `col161` decimal(16,2),
+ `col162` decimal(16,2),
+ `col163` decimal(16,2),
+ `col164` decimal(16,2),
+ `col165` decimal(16,2),
+ `col166` decimal(16,2),
+ `col167` decimal(16,2),
+ `col168` decimal(16,2),
+ `col169` decimal(16,2),
+ `col170` decimal(16,2),
+ `col171` decimal(16,2),
+ `col172` decimal(16,2),
+ `col173` decimal(16,2),
+ `col174` decimal(16,2),
+ `col175` decimal(16,2),
+ `col176` decimal(16,2),
+ `col177` decimal(16,2),
+ `col178` decimal(16,2),
+ `col179` decimal(16,2),
+ `col180` decimal(16,2),
+ `col181` decimal(16,2),
+ `col182` decimal(16,2),
+ `col183` decimal(16,2),
+ `col184` decimal(16,2),
+ `col185` decimal(16,2),
+ `col186` decimal(16,2),
+ `col187` decimal(16,2),
+ `col188` decimal(16,2),
+ `col189` decimal(16,2),
+ `col190` decimal(16,2),
+ `col191` decimal(16,2),
+ `col192` decimal(16,2),
+ `col193` decimal(16,2),
+ `col194` decimal(16,2),
+ `col195` decimal(16,2),
+ `col196` decimal(16,2),
+ `col197` decimal(16,2),
+ `col198` decimal(16,2),
+ `col199` decimal(16,2),
+ `col200` decimal(16,2),
+ `col201` decimal(16,2),
+ `col202` decimal(16,2),
+ `col203` decimal(16,2),
+ `col204` decimal(16,2),
+ `col205` decimal(16,2),
+ `col206` decimal(16,2),
+ `col207` decimal(16,2),
+ `col208` decimal(16,2),
+ `col209` decimal(16,2),
+ `col210` decimal(16,2),
+ `col211` decimal(16,2),
+ `col212` decimal(16,2),
+ `col213` decimal(16,2),
+ `col214` decimal(16,2),
+ `col215` decimal(16,2),
+ `col216` decimal(16,2),
+ `col217` decimal(16,2),
+ `col218` decimal(16,2),
+ `col219` decimal(16,2),
+ `col220` decimal(16,2),
+ `col221` decimal(16,2),
+ `col222` decimal(16,2),
+ `col223` decimal(16,2),
+ `col224` decimal(16,2),
+ `col225` decimal(16,2),
+ `col226` decimal(16,2),
+ `col227` decimal(16,2),
+ `col228` decimal(16,2),
+ `col229` decimal(16,2),
+ `col230` decimal(16,2),
+ `col231` decimal(22,6),
+ `col232` decimal(22,6),
+ `col233` decimal(22,6),
+ `col234` decimal(22,6),
+ `col235` decimal(22,6),
+ `col236` decimal(22,6),
+ `col237` decimal(22,6),
+ `col238` decimal(22,6),
+ `col239` decimal(22,6),
+ `col240` decimal(22,6),
+ `col241` decimal(22,6),
+ `col242` decimal(22,6),
+ `col243` decimal(22,6),
+ `col244` decimal(22,6),
+ `col245` decimal(22,6),
+ `col246` decimal(22,6),
+ `col247` decimal(22,6),
+ `col248` decimal(22,6),
+ `col249` decimal(22,6),
+ `col250` decimal(22,6),
+ `col251` decimal(22,6),
+ `col252` decimal(16,2),
+ `col253` decimal(16,2),
+ `col254` decimal(16,2),
+ `col255` decimal(16,2),
+ `col256` decimal(16,2),
+ `col257` decimal(16,2),
+ `col258` decimal(16,2),
+ `col259` decimal(16,2),
+ `col260` decimal(16,2),
+ `col261` decimal(16,2),
+ `col262` decimal(16,2),
+ `col263` decimal(16,2),
+ `col264` decimal(16,2),
+ `col265` decimal(16,2),
+ `col266` decimal(16,2),
+ `col267` decimal(16,2),
+ `col268` decimal(16,2),
+ `col269` decimal(16,2),
+ `col270` decimal(16,2),
+ `col271` decimal(16,2),
+ `col272` decimal(16,2),
+ `col273` decimal(16,2),
+ `col274` decimal(16,2),
+ `col275` decimal(16,2),
+ `col276` decimal(16,2),
+ `col277` decimal(16,2),
+ `col278` decimal(16,2),
+ `col279` decimal(16,2),
+ `col280` decimal(16,2),
+ `col281` decimal(16,2),
+ `col282` decimal(16,2),
+ `col283` decimal(16,2),
+ `col284` decimal(16,2),
+ `col285` decimal(16,2),
+ `col286` decimal(16,2),
+ `col287` decimal(16,2),
+ `col288` decimal(16,2),
+ `col289` decimal(16,2),
+ `col290` decimal(16,2),
+ `col291` decimal(16,2),
+ `col292` decimal(16,2),
+ `col293` decimal(16,2),
+ `col294` decimal(16,2),
+ `col295` decimal(16,2),
+ `col296` decimal(16,2),
+ `col297` decimal(16,2),
+ `col298` decimal(16,2),
+ `col299` decimal(16,2),
+ `col300` decimal(16,2),
+ `col301` decimal(16,2),
+ `col302` decimal(16,2),
+ `col303` decimal(16,2),
+ `col304` decimal(16,2),
+ `col305` decimal(16,2),
+ `col306` decimal(16,2),
+ `col307` decimal(16,2),
+ `col308` decimal(16,2),
+ `col309` decimal(16,2),
+ `col310` decimal(16,2),
+ `col311` decimal(16,2),
+ `col312` decimal(16,2),
+ `col313` decimal(16,2),
+ `col314` decimal(16,2),
+ `col315` decimal(16,2),
+ `col316` decimal(16,2),
+ `col317` decimal(16,2),
+ `col318` decimal(16,2),
+ `col319` decimal(16,2),
+ `col320` decimal(16,2),
+ `col321` decimal(16,2),
+ `col322` decimal(16,2),
+ `col323` decimal(16,2),
+ `col324` decimal(16,2),
+ `col325` decimal(16,2),
+ `col326` decimal(16,2),
+ `col327` decimal(16,2),
+ `col328` decimal(16,2),
+ `col329` decimal(16,2),
+ `col330` decimal(16,2),
+ `col331` decimal(16,2),
+ `col332` decimal(16,2),
+ `col333` decimal(16,2),
+ `col334` decimal(16,2),
+ `col335` decimal(16,2),
+ `col336` decimal(16,2),
+ `col337` decimal(16,2),
+ `col338` decimal(16,2),
+ `col339` decimal(22,6),
+ `col340` decimal(22,6),
+ `col341` decimal(22,6),
+ `col342` decimal(22,6),
+ `col343` decimal(22,6),
+ `col344` decimal(22,6),
+ `col345` decimal(22,6),
+ `col346` decimal(22,6),
+ `col347` decimal(22,6),
+ `col348` decimal(22,6),
+ `col349` decimal(22,6),
+ `col350` decimal(22,6),
+ `col351` decimal(22,6),
+ `col352` decimal(22,6),
+ `col353` decimal(16,2),
+ `col354` decimal(16,2),
+ `col355` decimal(16,2),
+ `col356` decimal(16,2),
+ `col357` decimal(16,2),
+ `col358` decimal(16,2),
+ `col359` decimal(16,2),
+ `col360` decimal(16,2),
+ `col361` decimal(16,2),
+ `col362` decimal(16,2),
+ `col363` decimal(16,2),
+ `col364` decimal(16,2),
+ `col365` decimal(16,2),
+ `col366` decimal(16,2),
+ `col367` decimal(16,2),
+ `col368` decimal(16,2),
+ `col369` decimal(16,2),
+ `col370` decimal(16,2),
+ `col371` decimal(16,2),
+ `col372` decimal(16,2),
+ `col373` decimal(16,2),
+ `col374` decimal(16,2),
+ `col375` decimal(16,2),
+ `col376` decimal(16,2),
+ `col377` decimal(16,2),
+ `col378` decimal(16,2),
+ `col379` decimal(16,2),
+ `col380` decimal(16,2),
+ `col381` decimal(16,2),
+ `col382` decimal(16,2),
+ `col383` decimal(16,2),
+ `col384` decimal(16,2),
+ `col385` decimal(16,2),
+ `col386` decimal(16,2),
+ `col387` decimal(16,2),
+ `col388` decimal(16,2),
+ `col389` decimal(16,2),
+ `col390` decimal(16,2),
+ `col391` decimal(16,2),
+ `col392` decimal(16,2),
+ `col393` decimal(16,2),
+ `col394` decimal(16,2),
+ `col395` decimal(8,0),
+ `col396` decimal(8,0),
+ `col397` string,
+ `col398` string,
+ `col399` decimal(16,2),
+ `col400` decimal(16,2),
+ `col401` decimal(16,2),
+ `col402` decimal(16,2),
+ `col403` decimal(16,2),
+ `col404` decimal(16,2),
+ `col405` decimal(16,2),
+ `col406` decimal(16,2),
+ `col407` decimal(16,2),
+ `col408` decimal(16,2),
+ `col409` decimal(16,2),
+ `col410` decimal(16,2),
+ `col411` decimal(16,2),
+ `col412` decimal(16,2),
+ `col413` decimal(16,2),
+ `col414` decimal(16,2),
+ `col415` decimal(16,2),
+ `col416` decimal(16,2),
+ `col417` decimal(16,2),
+ `col418` decimal(16,2),
+ `col419` decimal(16,2),
+ `col420` int,
+ `col421` int,
+ `col422` int,
+ `col423` decimal(16,2),
+ `col424` decimal(16,2),
+ `col425` decimal(16,2),
+ `col426` decimal(16,2),
+ `col427` decimal(16,2),
+ `col428` decimal(16,2),
+ `col429` string,
+ `col430` string,
+ `col431` string,
+ `col432` decimal(16,2),
+ `col433` decimal(16,2),
+ `col434` decimal(16,2),
+ `col435` decimal(16,2),
+ `col436` decimal(16,2),
+ `col437` decimal(16,2),
+ `col438` decimal(16,2),
+ `col439` decimal(16,2),
+ `col440` decimal(22,6),
+ `col441` decimal(22,6),
+ `col442` decimal(16,3),
+ `col443` decimal(16,3),
+ `col444` decimal(16,3),
+ `col445` decimal(16,2),
+ `col446` decimal(16,2),
+ `col447` decimal(9,6),
+ `col448` decimal(9,6),
+ `col449` decimal(16,2),
+ `col450` decimal(16,2),
+ `col451` decimal(16,2),
+ `col452` decimal(16,2),
+ `col453` decimal(16,2),
+ `col454` int,
+ `col455` decimal(16,2),
+ `col456` decimal(16,2),
+ `col457` decimal(16,2),
+ `col458` decimal(16,2),
+ `col459` decimal(16,2),
+ `col460` decimal(16,2),
+ `col461` decimal(16,2),
+ `col462` decimal(16,2),
+ `col463` decimal(22,6),
+ `col464` decimal(22,6),
+ `col465` decimal(22,6),
+ `col466` decimal(22,6),
+ `col467` decimal(22,6),
+ `col468` decimal(22,6),
+ `col469` decimal(22,6),
+ `col470` decimal(22,6),
+ `col471` decimal(16,2),
+ `col472` decimal(16,2),
+ `col473` decimal(16,2),
+ `col474` decimal(16,2),
+ `col475` decimal(16,2),
+ `col476` decimal(16,2),
+ `col477` decimal(16,2),
+ `col478` decimal(16,2),
+ `col479` decimal(16,2),
+ `col480` decimal(16,2),
+ `col481` decimal(16,2),
+ `col482` decimal(16,2),
+ `col483` decimal(16,2),
+ `col484` decimal(16,2),
+ `col485` decimal(16,2),
+ `col486` decimal(16,2),
+ `col487` decimal(16,2),
+ `col488` decimal(16,2),
+ `col489` decimal(16,2),
+ `col490` decimal(16,2),
+ `col491` decimal(16,2),
+ `col492` decimal(16,2),
+ `col493` decimal(16,2),
+ `col494` decimal(16,2),
+ `col495` decimal(16,2),
+ `col496` decimal(16,2),
+ `col497` decimal(16,2),
+ `col498` decimal(16,2),
+ `col499` decimal(16,2),
+ `col500` decimal(16,2),
+ `col501` decimal(16,2),
+ `col502` decimal(16,2),
+ `col503` decimal(16,2),
+ `col504` decimal(16,2),
+ `col505` decimal(16,2),
+ `col506` decimal(16,2),
+ `col507` decimal(16,2),
+ `col508` decimal(8,0),
+ `col509` decimal(8,0),
+ `col510` decimal(8,0),
+ `col511` decimal(8,0),
+ `col512` decimal(8,0),
+ `col513` decimal(8,0),
+ `col514` decimal(8,0),
+ `col515` decimal(8,0),
+ `col516` decimal(8,0),
+ `col517` decimal(8,0),
+ `col518` decimal(8,0),
+ `col519` decimal(8,0),
+ `col520` decimal(8,0),
+ `col521` decimal(8,0),
+ `col522` decimal(8,0),
+ `col523` decimal(8,0),
+ `col524` decimal(8,0),
+ `col525` decimal(8,0),
+ `col526` decimal(8,0),
+ `col527` decimal(8,0),
+ `col528` decimal(8,0),
+ `col529` decimal(8,0),
+ `col530` decimal(16,2),
+ `col531` decimal(16,2),
+ `col532` decimal(16,2),
+ `col533` decimal(16,2),
+ `col534` decimal(22,6),
+ `col535` decimal(16,2),
+ `col536` decimal(16,2),
+ `col537` decimal(16,2),
+ `col538` decimal(16,2),
+ `col539` decimal(22,6),
+ `col540` decimal(22,6),
+ `col541` decimal(22,6),
+ `col542` string,
+ `col543` decimal(16,2),
+ `col544` decimal(16,2),
+ `col545` decimal(16,2),
+ `col546` decimal(16,2),
+ `col547` decimal(16,2),
+ `col548` decimal(16,2),
+ `col549` decimal(16,2),
+ `col550` decimal(16,2),
+ `col551` decimal(16,2),
+ `col552` decimal(16,2),
+ `col553` decimal(16,2),
+ `col554` decimal(16,2),
+ `col555` decimal(16,2),
+ `col556` decimal(16,2),
+ `col557` string,
+ `col558` string,
+ `col559` string,
+ `col560` string,
+ `col561` string,
+ `col562` string,
+ `col563` string,
+ `col564` string,
+ `col565` decimal(16,2),
+ `col566` decimal(22,6),
+ `col567` decimal(22,6),
+ `col568` decimal(22,6),
+ `col569` string,
+ `col570` string,
+ `col571` decimal(16,2),
+ `col572` decimal(12,0),
+ `col573` string,
+ `col574` string,
+ `col575` decimal(22,6),
+ `col576` decimal(22,6),
+ `col577` decimal(22,6),
+ `col578` decimal(16,2),
+ `col579` decimal(22,2),
+ `col580` decimal(22,2),
+ `col581` decimal(22,2),
+ `col582` decimal(16,2),
+ `col583` decimal(16,2),
+ `col584` decimal(16,2),
+ `col585` decimal(16,2),
+ `col586` decimal(16,2),
+ `col587` decimal(16,2),
+ `col588` decimal(16,2),
+ `col589` decimal(16,2),
+ `col590` decimal(16,2),
+ `col591` decimal(16,2),
+ `col592` decimal(16,2),
+ `col593` decimal(16,2),
+ `col594` decimal(16,2),
+ `col595` decimal(16,2),
+ `col596` decimal(16,2),
+ `col597` string,
+ `col598` string,
+ `col599` decimal(16,2),
+ `col600` decimal(16,2),
+ `col601` decimal(16,2),
+ `col602` decimal(16,2),
+ `col603` decimal(16,2),
+ `col604` decimal(16,2),
+ `col605` decimal(16,2),
+ `col606` decimal(16,2),
+ `col607` int,
+ `col608` string,
+ `col609` string,
+ `col610` string,
+ `col611` decimal(22,2),
+ `col612` string,
+ `col613` string,
+ `col614` string,
+ `col615` string,
+ `col616` string,
+ `col617` string,
+ `col618` string,
+ `col619` string,
+ `col620` string,
+ `col621` string,
+ `col622` string,
+ `col623` string,
+ `col624` string,
+ `col625` string,
+ `col626` string,
+ `col627` string,
+ `col628` string,
+ `col629` decimal(16,2),
+ `col630` decimal(16,2),
+ `col631` string,
+ `col632` string,
+ `col633` string,
+ `col634` string,
+ `col635` string,
+ `col636` string,
+ `col637` decimal(16,2))
+ROW FORMAT SERDE
+ 'org.apache.hadoop.hive.ql.io.orc.OrcSerde'
+STORED AS INPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat'
+OUTPUTFORMAT
+ 'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat'
+LOCATION
+ '/user/doris/suites/multi_catalog/wide_table1_orc'
+TBLPROPERTIES (
+ 'transient_lastDdlTime'='1680503244');
+
+msck repair table wide_table1_orc;
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/run.sh
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/run.sh
new file mode 100755
index 00000000000..bfaeadfad06
--- /dev/null
+++
b/docker/thirdparties/docker-compose/hive/scripts/data/multi_catalog/test_wide_table/run.sh
@@ -0,0 +1,22 @@
+#!/bin/bash
+set -x
+
+CUR_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" &>/dev/null && pwd)"
+
+if [[ ! -d "${CUR_DIR}/data" ]]; then
+ echo "${CUR_DIR}/data does not exist"
+ cd "${CUR_DIR}" && rm -f data.tar.gz \
+ && curl -O
https://s3BucketName.s3Endpoint/regression/datalake/pipeline_data/multi_catalog/test_wide_table/data.tar.gz
\
+ && tar xzf data.tar.gz
+ cd -
+else
+ echo "${CUR_DIR}/data exist, continue !"
+fi
+
+## mkdir and put data to hdfs
+hadoop fs -mkdir -p /user/doris/suites/multi_catalog/
+hadoop fs -put "${CUR_DIR}"/data/* /user/doris/suites/multi_catalog/
+
+# create table
+hive -f "${CUR_DIR}"/create_table.hql
+
diff --git a/regression-test/data/external_table_p2/hive/test_complex_types.out
b/regression-test/data/external_table_p0/hive/test_complex_types.out
similarity index 50%
rename from regression-test/data/external_table_p2/hive/test_complex_types.out
rename to regression-test/data/external_table_p0/hive/test_complex_types.out
index 31cff0e45b2..4b94a4f1511 100644
--- a/regression-test/data/external_table_p2/hive/test_complex_types.out
+++ b/regression-test/data/external_table_p0/hive/test_complex_types.out
@@ -17,12 +17,6 @@
-- !array_max --
11028
--- !array_filter --
-11028
-
--- !array_last --
-0.9899828598260161
-
-- !null_struct_element_orc --
0
@@ -41,11 +35,53 @@
-- !array_max_orc --
11028
--- !array_filter_orc --
+-- !offsets_check --
+0 [1, 2] [[], [3], null] {"a":1, "b":2} {"s1": "e", "s2": null}
+1 [] [] {} \N
+2 \N \N \N {"s1": "h", "s2": 10}
+3 [5, null] [[6, 7], [8, null], null] {"f":1, "g":null}
{"s1": null, "s2": 9}
+
+-- !map_with_nullable_key --
+\N \N \N \N \N \N \N \N \N
test test
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
[...]
+
+-- !date_dict --
+2036-12-28 1898-12-28 2539-12-28
+
+-- !null_struct_element --
+0
+
+-- !map_key_select --
+38111 0.770169659057425
+
+-- !map_keys --
+["9wXr9n-TBm9Wyt-r8H-SkAq", "CPDH4G-ZXGPkku-3wY-ktaQ",
"RvNlMt-HHjHN5M-VjP-xHAI", "qKIhKy-Ws344os-haX-2pmT",
"DOJJ5l-UEkwVMs-x9F-HifD", "m871g8-1eFi7jt-oBq-S0yc",
"wXugVP-v2fc6IF-DeU-On3T", "B0mXFX-QvgUgo7-Dih-6rDu",
"E9zv3F-xMqSbMa-il4-FuDg", "msuFIN-ZkKO8TY-tu4-veH0",
"0rSUyl-Un07aIW-KAx-WHnX", "XvbmO8-WA6oAqc-ihc-s8IL",
"G6B6RD-AicAlZb-16u-Pn1I", "coDK0Q-tMg1294-JMQ-ZWQu",
"4c0aWh-yhL6BOX-rRu-1n0r", "G4iUcG-ZhWw62v-VLt-n6lH",
"IIB7qD-WQistwT-Vux-0c9B", "7cTyuR-5ssXm2S-sJR-JTIZ", "3KPh [...]
+
+-- !map_values --
+[0.9805502029231666, 0.5330291595754054, 0.30024744873379805,
0.48563601750302665, 0.7687106425158624, 0.6993506644925102,
0.2849354808825807, 0.3473417455186141, 0.13500129443045072,
0.9708132103700939, 0.18583042639943448, 0.4886337264552073,
0.36354741695157655, 0.5640845268971175, 0.1374134087807577,
0.7766547647451623, 0.5835323296668318, 0.3654459547110349, 0.5479776709993764,
0.8379932542117192, 0.15665046278350814, 0.03371222042250388,
0.1699781825927229, 0.3579630495075078, 0.02 [...]
+
+-- !map_contains_key --
+1077 [0.7805560995873845, 0.9303489002269559, 0.2529522997521877,
0.662270811026298, 0.664725297532439, 0.1019441091764477, 0.9614059300688174,
0.5278126009983843, 0.5287505841216708, 0.426116738236779, 0.42300502393871175,
0.5327026330053651, 0.6025481777942603, 0.2710733647257627, 0.613792118138183,
0.0021003027835629906, 0.3200675048728582, 0.5485611014660204,
0.5121510581313707, 0.5145136652805358]
{"9wXr9n-TBm9Wyt-r8H-SkAq":0.9338329010480995,
"CPDH4G-ZXGPkku-3wY-ktaQ":0.43552569633 [...]
+
+-- !array_max --
11028
--- !array_last_orc --
-0.9899828598260161
+-- !null_struct_element_orc --
+0
+
+-- !map_key_select_orc --
+38111 0.770169659057425
+
+-- !map_keys_orc --
+["9wXr9n-TBm9Wyt-r8H-SkAq", "CPDH4G-ZXGPkku-3wY-ktaQ",
"RvNlMt-HHjHN5M-VjP-xHAI", "qKIhKy-Ws344os-haX-2pmT",
"DOJJ5l-UEkwVMs-x9F-HifD", "m871g8-1eFi7jt-oBq-S0yc",
"wXugVP-v2fc6IF-DeU-On3T", "B0mXFX-QvgUgo7-Dih-6rDu",
"E9zv3F-xMqSbMa-il4-FuDg", "msuFIN-ZkKO8TY-tu4-veH0",
"0rSUyl-Un07aIW-KAx-WHnX", "XvbmO8-WA6oAqc-ihc-s8IL",
"G6B6RD-AicAlZb-16u-Pn1I", "coDK0Q-tMg1294-JMQ-ZWQu",
"4c0aWh-yhL6BOX-rRu-1n0r", "G4iUcG-ZhWw62v-VLt-n6lH",
"IIB7qD-WQistwT-Vux-0c9B", "7cTyuR-5ssXm2S-sJR-JTIZ", "3KPh [...]
+
+-- !map_values_orc --
+[0.9805502029231666, 0.5330291595754054, 0.30024744873379805,
0.48563601750302665, 0.7687106425158624, 0.6993506644925102,
0.2849354808825807, 0.3473417455186141, 0.13500129443045072,
0.9708132103700939, 0.18583042639943448, 0.4886337264552073,
0.36354741695157655, 0.5640845268971175, 0.1374134087807577,
0.7766547647451623, 0.5835323296668318, 0.3654459547110349, 0.5479776709993764,
0.8379932542117192, 0.15665046278350814, 0.03371222042250388,
0.1699781825927229, 0.3579630495075078, 0.02 [...]
+
+-- !map_contains_key_orc --
+1077 [0.7805560995873845, 0.9303489002269559, 0.2529522997521877,
0.662270811026298, 0.664725297532439, 0.1019441091764477, 0.9614059300688174,
0.5278126009983843, 0.5287505841216708, 0.426116738236779, 0.42300502393871175,
0.5327026330053651, 0.6025481777942603, 0.2710733647257627, 0.613792118138183,
0.0021003027835629906, 0.3200675048728582, 0.5485611014660204,
0.5121510581313707, 0.5145136652805358]
{"9wXr9n-TBm9Wyt-r8H-SkAq":0.9338329010480995,
"CPDH4G-ZXGPkku-3wY-ktaQ":0.43552569633 [...]
+
+-- !array_max_orc --
+11028
-- !offsets_check --
0 [1, 2] [[], [3], null] {"a":1, "b":2} {"s1": "e", "s2": null}
diff --git
a/regression-test/data/external_table_p2/hive/test_hive_same_db_table_name.out
b/regression-test/data/external_table_p0/hive/test_hive_same_db_table_name.out
similarity index 62%
rename from
regression-test/data/external_table_p2/hive/test_hive_same_db_table_name.out
rename to
regression-test/data/external_table_p0/hive/test_hive_same_db_table_name.out
index 03e8be6475d..b3ff544066f 100644
---
a/regression-test/data/external_table_p2/hive/test_hive_same_db_table_name.out
+++
b/regression-test/data/external_table_p0/hive/test_hive_same_db_table_name.out
@@ -14,3 +14,18 @@
-- !6 --
1 name1
+-- !1 --
+
+-- !2 --
+
+-- !3 --
+
+-- !4 --
+1 name1
+
+-- !5 --
+1 name1
+
+-- !6 --
+1 name1
+
diff --git
a/regression-test/data/external_table_p2/hive/test_hive_special_char_partition.out
b/regression-test/data/external_table_p0/hive/test_hive_special_char_partition.out
similarity index 54%
rename from
regression-test/data/external_table_p2/hive/test_hive_special_char_partition.out
rename to
regression-test/data/external_table_p0/hive/test_hive_special_char_partition.out
index 0bd26b12760..f81719d2d0e 100644
---
a/regression-test/data/external_table_p2/hive/test_hive_special_char_partition.out
+++
b/regression-test/data/external_table_p0/hive/test_hive_special_char_partition.out
@@ -49,3 +49,53 @@ name6 2023%01%01
-- !13 --
name# 2023#01#01
+-- !1 --
+name# 2023#01#01
+name1 2023/01/01
+name10 2023<01><01>
+name11 2023\\01\\01
+name12 2023.01.01
+name2 2023 01 01
+name3 2023:01:01
+name4 2023?01?01
+name5 2023=01=01
+name6 2023%01%01
+name8 2023"01"01
+name9 2023'01'01
+
+-- !2 --
+name2
+
+-- !3 --
+name1
+
+-- !4 --
+name4 2023?01?01
+
+-- !5 --
+name12 2023.01.01
+
+-- !6 --
+name10 2023<01><01>
+
+-- !7 --
+name3 2023:01:01
+
+-- !8 --
+name5 2023=01=01
+
+-- !9 --
+name8 2023"01"01
+
+-- !10 --
+name9 2023'01'01
+
+-- !11 --
+name11 2023\\01\\01
+
+-- !12 --
+name6 2023%01%01
+
+-- !13 --
+name# 2023#01#01
+
diff --git a/regression-test/data/external_table_p0/hive/test_wide_table.out
b/regression-test/data/external_table_p0/hive/test_wide_table.out
new file mode 100644
index 00000000000..9b5b21a8d98
--- /dev/null
+++ b/regression-test/data/external_table_p0/hive/test_wide_table.out
@@ -0,0 +1,45 @@
+-- This file is automatically generated. You should know what you did if you
want to edit this
+-- !01 --
+6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
+
+-- !02 --
+6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
+
+-- !03 --
+
+-- !04 --
+
+-- !05 --
+6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
+
+-- !06 --
+6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
+
+-- !07 --
+6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
+
+-- !08 --
+9999999541515682 99999218685068.86 99999869 221095586.11
27.54254 61077635638.763621
+
+-- !01 --
+6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
+
+-- !02 --
+6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
+
+-- !03 --
+
+-- !04 --
+
+-- !05 --
+6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
+
+-- !06 --
+6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
+
+-- !07 --
+6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
+
+-- !08 --
+9999999541515682 99999218685068.86 99999869 221095586.11
27.54254 61077635638.763621
+
diff --git a/regression-test/data/external_table_p2/hive/test_wide_table.out
b/regression-test/data/external_table_p2/hive/test_wide_table.out
deleted file mode 100644
index 143aeb9bf50..00000000000
--- a/regression-test/data/external_table_p2/hive/test_wide_table.out
+++ /dev/null
@@ -1,23 +0,0 @@
--- This file is automatically generated. You should know what you did if you
want to edit this
--- !01 --
-6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
-
--- !02 --
-6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
-
--- !03 --
-
--- !04 --
-
--- !05 --
-6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
-
--- !06 --
-6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
-
--- !07 --
-6117920261 28156890937818.64 11058113 84788841307158.93
9988065.83660 8116313253956313.527443
-
--- !08 --
-9999999541515682.000000000 99999218685068.860000000
99999869.000000000 221095586.110000000 27.542540000
61077635638.763621000
-
diff --git
a/regression-test/suites/external_table_p2/hive/test_complex_types.groovy
b/regression-test/suites/external_table_p0/hive/test_complex_types.groovy
similarity index 64%
rename from
regression-test/suites/external_table_p2/hive/test_complex_types.groovy
rename to
regression-test/suites/external_table_p0/hive/test_complex_types.groovy
index 1a90570ff42..e8cba5df36d 100644
--- a/regression-test/suites/external_table_p2/hive/test_complex_types.groovy
+++ b/regression-test/suites/external_table_p0/hive/test_complex_types.groovy
@@ -15,20 +15,23 @@
// specific language governing permissions and limitations
// under the License.
-suite("test_complex_types",
"p2,external,hive,external_remote,external_remote_hive") {
- String enabled = context.config.otherConfigs.get("enableExternalHiveTest")
- if (enabled != null && enabled.equalsIgnoreCase("true")) {
- String extHiveHmsHost =
context.config.otherConfigs.get("extHiveHmsHost")
- String extHiveHmsPort =
context.config.otherConfigs.get("extHiveHmsPort")
- String catalog_name = "test_complex_types"
- sql """drop catalog if exists ${catalog_name};"""
- sql """
- create catalog if not exists ${catalog_name} properties (
- 'type'='hms',
- 'hadoop.username' = 'hadoop',
- 'hive.metastore.uris' =
'thrift://${extHiveHmsHost}:${extHiveHmsPort}'
- );
- """
+suite("test_complex_types",
"p0,external,hive,external_docker,external_docker_hive") {
+ String enabled = context.config.otherConfigs.get("enableHiveTest")
+ if (enabled == null || !enabled.equalsIgnoreCase("true")) {
+ logger.info("diable Hive test.")
+ return;
+ }
+
+ for (String hivePrefix : ["hive2", "hive3"]) {
+ String hms_port = context.config.otherConfigs.get(hivePrefix +
"HmsPort")
+ String catalog_name = "${hivePrefix}_test_complex_types"
+ String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
+
+ sql """drop catalog if exists ${catalog_name}"""
+ sql """create catalog if not exists ${catalog_name} properties (
+ "type"="hms",
+ 'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}'
+ );"""
logger.info("catalog " + catalog_name + " created")
sql """switch ${catalog_name};"""
logger.info("switched to catalog " + catalog_name)
@@ -47,9 +50,9 @@ suite("test_complex_types",
"p2,external,hive,external_remote,external_remote_hi
qt_array_max """select count(array_max(capacity)) from byd where
array_max(capacity) > 0.99"""
- qt_array_filter """select count(array_size(array_filter(i -> (i >
0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99),
capacity))"""
+ // qt_array_filter """select count(array_size(array_filter(i -> (i >
0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99),
capacity))"""
- qt_array_last """select max(array_last(i -> i > 0, capacity)) from byd
where array_last(i -> i > 0, capacity) < 0.99"""
+ // qt_array_last """select max(array_last(i -> i > 0, capacity)) from
byd where array_last(i -> i > 0, capacity) < 0.99"""
qt_null_struct_element_orc """select count(struct_element(favor,
'tip')) from byd where id % 13 = 0"""
@@ -63,9 +66,9 @@ suite("test_complex_types",
"p2,external,hive,external_remote,external_remote_hi
qt_array_max_orc """select count(array_max(capacity)) from byd where
array_max(capacity) > 0.99"""
- qt_array_filter_orc """select count(array_size(array_filter(i -> (i >
0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99),
capacity))"""
+ // qt_array_filter_orc """select count(array_size(array_filter(i -> (i
> 0.99), capacity))) from byd where array_size(array_filter(i -> (i > 0.99),
capacity))"""
- qt_array_last_orc """select max(array_last(i -> i > 0, capacity)) from
byd where array_last(i -> i > 0, capacity) < 0.99"""
+ // qt_array_last_orc """select max(array_last(i -> i > 0, capacity))
from byd where array_last(i -> i > 0, capacity) < 0.99"""
qt_offsets_check """select * from complex_offsets_check order by id"""
diff --git
a/regression-test/suites/external_table_p2/hive/test_hive_same_db_table_name.groovy
b/regression-test/suites/external_table_p0/hive/test_hive_same_db_table_name.groovy
similarity index 66%
rename from
regression-test/suites/external_table_p2/hive/test_hive_same_db_table_name.groovy
rename to
regression-test/suites/external_table_p0/hive/test_hive_same_db_table_name.groovy
index 1ae209f53f1..d881facbba2 100644
---
a/regression-test/suites/external_table_p2/hive/test_hive_same_db_table_name.groovy
+++
b/regression-test/suites/external_table_p0/hive/test_hive_same_db_table_name.groovy
@@ -15,20 +15,23 @@
// specific language governing permissions and limitations
// under the License.
-suite("test_hive_same_db_table_name",
"p2,external,hive,external_remote,external_remote_hive") {
- String enabled = context.config.otherConfigs.get("enableExternalHiveTest")
- if (enabled != null && enabled.equalsIgnoreCase("true")) {
- String extHiveHmsHost =
context.config.otherConfigs.get("extHiveHmsHost")
- String extHiveHmsPort =
context.config.otherConfigs.get("extHiveHmsPort")
- String catalog_name = "test_hive_same_db_table_name"
- sql """drop catalog if exists ${catalog_name};"""
- sql """
- create catalog if not exists ${catalog_name} properties (
- 'type'='hms',
- 'hadoop.username' = 'hadoop',
- 'hive.metastore.uris' =
'thrift://${extHiveHmsHost}:${extHiveHmsPort}'
- );
- """
+suite("test_hive_same_db_table_name",
"p0,external,hive,external_docker,external_docker_hive") {
+ String enabled = context.config.otherConfigs.get("enableHiveTest")
+ if (enabled == null || !enabled.equalsIgnoreCase("true")) {
+ logger.info("diable Hive test.")
+ return;
+ }
+
+ for (String hivePrefix : ["hive2", "hive3"]) {
+ String hms_port = context.config.otherConfigs.get(hivePrefix +
"HmsPort")
+ String catalog_name = "${hivePrefix}_test_hive_same_db_table_name"
+ String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
+
+ sql """drop catalog if exists ${catalog_name}"""
+ sql """create catalog if not exists ${catalog_name} properties (
+ "type"="hms",
+ 'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}'
+ );"""
sql """switch internal;"""
sql """create database if not exists multi_catalog;"""
sql """use multi_catalog;"""
diff --git
a/regression-test/suites/external_table_p2/hive/test_hive_special_char_partition.groovy
b/regression-test/suites/external_table_p0/hive/test_hive_special_char_partition.groovy
similarity index 71%
rename from
regression-test/suites/external_table_p2/hive/test_hive_special_char_partition.groovy
rename to
regression-test/suites/external_table_p0/hive/test_hive_special_char_partition.groovy
index 563a3a68c94..8b78ab2e3ce 100644
---
a/regression-test/suites/external_table_p2/hive/test_hive_special_char_partition.groovy
+++
b/regression-test/suites/external_table_p0/hive/test_hive_special_char_partition.groovy
@@ -15,20 +15,23 @@
// specific language governing permissions and limitations
// under the License.
-suite("test_hive_special_char_partition",
"p2,external,hive,external_remote,external_remote_hive") {
- String enabled = context.config.otherConfigs.get("enableExternalHiveTest")
- if (enabled != null && enabled.equalsIgnoreCase("true")) {
- String extHiveHmsHost =
context.config.otherConfigs.get("extHiveHmsHost")
- String extHiveHmsPort =
context.config.otherConfigs.get("extHiveHmsPort")
- String catalog_name = "test_hive_special_char_partition"
- sql """drop catalog if exists ${catalog_name};"""
- sql """
- create catalog if not exists ${catalog_name} properties (
- 'type'='hms',
- 'hadoop.username' = 'hadoop',
- 'hive.metastore.uris' =
'thrift://${extHiveHmsHost}:${extHiveHmsPort}'
- );
- """
+suite("test_hive_special_char_partition",
"p0,external,hive,external_docker,external_docker_hive") {
+ String enabled = context.config.otherConfigs.get("enableHiveTest")
+ if (enabled == null || !enabled.equalsIgnoreCase("true")) {
+ logger.info("diable Hive test.")
+ return;
+ }
+
+ for (String hivePrefix : ["hive2", "hive3"]) {
+ String hms_port = context.config.otherConfigs.get(hivePrefix +
"HmsPort")
+ String catalog_name = "${hivePrefix}_test_hive_special_char_partition"
+ String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
+
+ sql """drop catalog if exists ${catalog_name}"""
+ sql """create catalog if not exists ${catalog_name} properties (
+ "type"="hms",
+ 'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}'
+ );"""
logger.info("catalog " + catalog_name + " created")
sql """switch ${catalog_name};"""
logger.info("switched to catalog " + catalog_name)
diff --git
a/regression-test/suites/external_table_p2/hive/test_wide_table.groovy
b/regression-test/suites/external_table_p0/hive/test_wide_table.groovy
similarity index 77%
rename from regression-test/suites/external_table_p2/hive/test_wide_table.groovy
rename to regression-test/suites/external_table_p0/hive/test_wide_table.groovy
index e8de2380eb5..1c266d22c67 100644
--- a/regression-test/suites/external_table_p2/hive/test_wide_table.groovy
+++ b/regression-test/suites/external_table_p0/hive/test_wide_table.groovy
@@ -15,7 +15,7 @@
// specific language governing permissions and limitations
// under the License.
-suite("test_wide_table",
"p2,external,hive,external_remote,external_remote_hive") {
+suite("test_wide_table",
"p0,external,hive,external_docker,external_docker_hive") {
def formats = ["_orc"]
def decimal_test1 = """select col1, col70, col71, col81, col100, col534
from wide_table1SUFFIX where col1 is not null order by col1 limit 1;"""
@@ -39,24 +39,27 @@ suite("test_wide_table",
"p2,external,hive,external_remote,external_remote_hive"
"""
def decimal_test8 = """select max(col1), max(col70), max(col71),
min(col81), min(col100), min(col534) from wide_table1SUFFIX;"""
- String enabled = context.config.otherConfigs.get("enableExternalHiveTest")
- if (enabled != null && enabled.equalsIgnoreCase("true")) {
- String extHiveHmsHost =
context.config.otherConfigs.get("extHiveHmsHost")
- String extHiveHmsPort =
context.config.otherConfigs.get("extHiveHmsPort")
- String catalog_name = "external_wide_table"
+ String enabled = context.config.otherConfigs.get("enableHiveTest")
+ if (enabled == null || !enabled.equalsIgnoreCase("true")) {
+ logger.info("diable Hive test.")
+ return;
+ }
+
+ for (String hivePrefix : ["hive2", "hive3"]) {
+ String hms_port = context.config.otherConfigs.get(hivePrefix +
"HmsPort")
+ String catalog_name = "${hivePrefix}_test_complex_types"
+ String externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
- sql """drop catalog if exists ${catalog_name};"""
- sql """
- create catalog if not exists ${catalog_name} properties (
- 'type'='hms',
- 'hive.metastore.uris' =
'thrift://${extHiveHmsHost}:${extHiveHmsPort}'
- );
- """
+ sql """drop catalog if exists ${catalog_name}"""
+ sql """create catalog if not exists ${catalog_name} properties (
+ "type"="hms",
+ 'hive.metastore.uris' = 'thrift://${externalEnvIp}:${hms_port}'
+ );"""
logger.info("catalog " + catalog_name + " created")
sql """switch ${catalog_name};"""
logger.info("switched to catalog " + catalog_name)
- sql """use wide_tables;"""
- logger.info("use wide_tables")
+ sql """use multi_catalog;"""
+ logger.info("use multi_catalog")
for (String format in formats) {
logger.info("Process format " + format)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]