This is an automated email from the ASF dual-hosted git repository.
kxiao pushed a commit to branch branch-2.0
in repository https://gitbox.apache.org/repos/asf/doris.git
The following commit(s) were added to refs/heads/branch-2.0 by this push:
new 7cc8ab26c69 Revert "[refactor](Nereids) refactor infer predicate rule
to avoid lost cast (#25637)"
7cc8ab26c69 is described below
commit 7cc8ab26c69a39180bf2e6934e54a14171cbfe97
Author: Kang <[email protected]>
AuthorDate: Thu Oct 26 10:37:50 2023 +0800
Revert "[refactor](Nereids) refactor infer predicate rule to avoid lost
cast (#25637)"
This reverts commit 5ca2fc6b1ede1141b557a17a3db166e5b93edd3a, since it's
mixed by mistake
---
be/src/olap/rowset/segment_v2/segment_iterator.cpp | 40 +++++++--------------
be/src/olap/rowset/segment_v2/segment_iterator.h | 1 -
be/src/vec/columns/column.h | 2 --
be/src/vec/columns/column_map.cpp | 21 -----------
be/src/vec/columns/column_map.h | 4 +--
be/src/vec/columns/column_nullable.h | 2 --
be/src/vec/columns/column_struct.cpp | 15 --------
be/src/vec/columns/column_struct.h | 4 ---
.../test_nestedtypes_csv_insert_into_with_s3.out | 2 +-
.../test_nestedtypes_json_insert_into_with_s3.out | 6 ++--
regression-test/data/export/test_struct_export.out | Bin 442 -> 469 bytes
.../data/insert_p0/test_struct_insert.out | Bin 417 -> 444 bytes
.../data/load_p0/stream_load/map_char_test.csv | 4 ---
.../data/load_p0/stream_load/test_stream_load.out | 6 ----
.../aggregate/aggregate_group_by_metric_type.out | Bin 154 -> 170 bytes
.../load_p0/stream_load/test_stream_load.groovy | 40 +--------------------
16 files changed, 19 insertions(+), 128 deletions(-)
diff --git a/be/src/olap/rowset/segment_v2/segment_iterator.cpp
b/be/src/olap/rowset/segment_v2/segment_iterator.cpp
index ce04b05d4b4..bd2b9df914c 100644
--- a/be/src/olap/rowset/segment_v2/segment_iterator.cpp
+++ b/be/src/olap/rowset/segment_v2/segment_iterator.cpp
@@ -1498,38 +1498,24 @@ bool
SegmentIterator::_can_evaluated_by_vectorized(ColumnPredicate* predicate) {
}
}
-bool SegmentIterator::_has_char_type(const Field& column_desc) {
- switch (column_desc.type()) {
- case FieldType::OLAP_FIELD_TYPE_CHAR:
- return true;
- case FieldType::OLAP_FIELD_TYPE_ARRAY:
- return _has_char_type(*column_desc.get_sub_field(0));
- case FieldType::OLAP_FIELD_TYPE_MAP:
- return _has_char_type(*column_desc.get_sub_field(0)) ||
- _has_char_type(*column_desc.get_sub_field(1));
- case FieldType::OLAP_FIELD_TYPE_STRUCT:
- for (int idx = 0; idx < column_desc.get_sub_field_count(); ++idx) {
- if (_has_char_type(*column_desc.get_sub_field(idx))) {
- return true;
- }
- }
- return false;
- default:
- return false;
- }
-};
-
void SegmentIterator::_vec_init_char_column_id() {
for (size_t i = 0; i < _schema->num_column_ids(); i++) {
auto cid = _schema->column_id(i);
- const Field* column_desc = _schema->column(cid);
+ auto column_desc = _schema->column(cid);
- if (_has_char_type(*column_desc)) {
- _char_type_idx.emplace_back(i);
- if (i != 0) {
- _char_type_idx_no_0.emplace_back(i);
+ do {
+ if (column_desc->type() == FieldType::OLAP_FIELD_TYPE_CHAR) {
+ _char_type_idx.emplace_back(i);
+ if (i != 0) {
+ _char_type_idx_no_0.emplace_back(i);
+ }
+ break;
+ } else if (column_desc->type() !=
FieldType::OLAP_FIELD_TYPE_ARRAY) {
+ break;
}
- }
+ // for Array<Char> or Array<Array<Char>>
+ column_desc = column_desc->get_sub_field(0);
+ } while (column_desc != nullptr);
}
}
diff --git a/be/src/olap/rowset/segment_v2/segment_iterator.h
b/be/src/olap/rowset/segment_v2/segment_iterator.h
index 33d3a3f5f9c..d53ad9d62b0 100644
--- a/be/src/olap/rowset/segment_v2/segment_iterator.h
+++ b/be/src/olap/rowset/segment_v2/segment_iterator.h
@@ -201,7 +201,6 @@ private:
// CHAR type in storage layer padding the 0 in length. But query engine
need ignore the padding 0.
// so segment iterator need to shrink char column before output it. only
use in vec query engine.
void _vec_init_char_column_id();
- bool _has_char_type(const Field& column_desc);
uint32_t segment_id() const { return _segment->id(); }
uint32_t num_rows() const { return _segment->num_rows(); }
diff --git a/be/src/vec/columns/column.h b/be/src/vec/columns/column.h
index 139a23f7935..df4c66fa14c 100644
--- a/be/src/vec/columns/column.h
+++ b/be/src/vec/columns/column.h
@@ -670,8 +670,6 @@ public:
virtual bool is_column_map() const { return false; }
- virtual bool is_column_struct() const { return false; }
-
/// If the only value column can contain is NULL.
virtual bool only_null() const { return false; }
diff --git a/be/src/vec/columns/column_map.cpp
b/be/src/vec/columns/column_map.cpp
index 9a550097af9..f10890b0093 100644
--- a/be/src/vec/columns/column_map.cpp
+++ b/be/src/vec/columns/column_map.cpp
@@ -442,27 +442,6 @@ void ColumnMap::replicate(const uint32_t* indexs, size_t
target_size, IColumn& c
->replicate(indexs, target_size,
res.values_column->assume_mutable_ref());
}
-MutableColumnPtr ColumnMap::get_shrinked_column() {
- MutableColumns new_columns(2);
-
- if (keys_column->is_column_string() || keys_column->is_column_array() ||
- keys_column->is_column_map() || keys_column->is_column_struct()) {
- new_columns[0] = keys_column->get_shrinked_column();
- } else {
- new_columns[0] = keys_column->get_ptr();
- }
-
- if (values_column->is_column_string() || values_column->is_column_array()
||
- values_column->is_column_map() || values_column->is_column_struct()) {
- new_columns[1] = values_column->get_shrinked_column();
- } else {
- new_columns[1] = values_column->get_ptr();
- }
-
- return ColumnMap::create(new_columns[0]->assume_mutable(),
new_columns[1]->assume_mutable(),
- offsets_column->assume_mutable());
-}
-
void ColumnMap::reserve(size_t n) {
get_offsets().reserve(n);
keys_column->reserve(n);
diff --git a/be/src/vec/columns/column_map.h b/be/src/vec/columns/column_map.h
index f8bc8a4812d..0c09f0a9dae 100644
--- a/be/src/vec/columns/column_map.h
+++ b/be/src/vec/columns/column_map.h
@@ -36,8 +36,6 @@
#include "vec/columns/column.h"
#include "vec/columns/column_array.h"
#include "vec/columns/column_impl.h"
-#include "vec/columns/column_nullable.h"
-#include "vec/columns/column_struct.h"
#include "vec/columns/column_vector.h"
#include "vec/common/assert_cast.h"
#include "vec/common/cow.h"
@@ -112,7 +110,7 @@ public:
const char* deserialize_and_insert_from_arena(const char* pos) override;
void update_hash_with_value(size_t n, SipHash& hash) const override;
- MutableColumnPtr get_shrinked_column() override;
+
ColumnPtr filter(const Filter& filt, ssize_t result_size_hint) const
override;
size_t filter(const Filter& filter) override;
ColumnPtr permute(const Permutation& perm, size_t limit) const override;
diff --git a/be/src/vec/columns/column_nullable.h
b/be/src/vec/columns/column_nullable.h
index 4350bead62d..465e7dc2d96 100644
--- a/be/src/vec/columns/column_nullable.h
+++ b/be/src/vec/columns/column_nullable.h
@@ -262,8 +262,6 @@ public:
bool is_column_decimal() const override { return
get_nested_column().is_column_decimal(); }
bool is_column_string() const override { return
get_nested_column().is_column_string(); }
bool is_column_array() const override { return
get_nested_column().is_column_array(); }
- bool is_column_map() const override { return
get_nested_column().is_column_map(); }
- bool is_column_struct() const override { return
get_nested_column().is_column_struct(); }
bool is_fixed_and_contiguous() const override { return false; }
bool values_have_fixed_size() const override { return
nested_column->values_have_fixed_size(); }
diff --git a/be/src/vec/columns/column_struct.cpp
b/be/src/vec/columns/column_struct.cpp
index a2d1eae5709..88391370398 100644
--- a/be/src/vec/columns/column_struct.cpp
+++ b/be/src/vec/columns/column_struct.cpp
@@ -293,21 +293,6 @@ void ColumnStruct::replicate(const uint32_t* indexs,
size_t target_size, IColumn
}
}
-MutableColumnPtr ColumnStruct::get_shrinked_column() {
- const size_t tuple_size = columns.size();
- MutableColumns new_columns(tuple_size);
-
- for (size_t i = 0; i < tuple_size; ++i) {
- if (columns[i]->is_column_string() || columns[i]->is_column_array() ||
- columns[i]->is_column_map() || columns[i]->is_column_struct()) {
- new_columns[i] = columns[i]->get_shrinked_column();
- } else {
- new_columns[i] = columns[i]->get_ptr();
- }
- }
- return ColumnStruct::create(std::move(new_columns));
-}
-
MutableColumns ColumnStruct::scatter(ColumnIndex num_columns, const Selector&
selector) const {
const size_t tuple_size = columns.size();
std::vector<MutableColumns> scattered_tuple_elements(tuple_size);
diff --git a/be/src/vec/columns/column_struct.h
b/be/src/vec/columns/column_struct.h
index db9633f6032..9d219c3c989 100644
--- a/be/src/vec/columns/column_struct.h
+++ b/be/src/vec/columns/column_struct.h
@@ -82,7 +82,6 @@ public:
}
std::string get_name() const override;
- bool is_column_struct() const override { return true; }
const char* get_family_name() const override { return "Struct"; }
bool can_be_inside_nullable() const override { return true; }
MutableColumnPtr clone_empty() const override;
@@ -161,9 +160,6 @@ public:
LOG(FATAL) << "compare_at not implemented";
}
void get_extremes(Field& min, Field& max) const override;
-
- MutableColumnPtr get_shrinked_column() override;
-
void reserve(size_t n) override;
void resize(size_t n) override;
size_t byte_size() const override;
diff --git
a/regression-test/data/datatype_p0/nested_types/query/test_nestedtypes_csv_insert_into_with_s3.out
b/regression-test/data/datatype_p0/nested_types/query/test_nestedtypes_csv_insert_into_with_s3.out
index 0774e566854..3aeb3276518 100644
---
a/regression-test/data/datatype_p0/nested_types/query/test_nestedtypes_csv_insert_into_with_s3.out
+++
b/regression-test/data/datatype_p0/nested_types/query/test_nestedtypes_csv_insert_into_with_s3.out
@@ -15,7 +15,7 @@
0 {"true":true,"false":false}
{"69":-94,"101":-115,"-64":66,"-71":101,"0":38,"-49":-24,"88":45,"-98":20,"5":88,"120":25,"-59":-66,"-91":92,"-113":61,"-39":109,"-22":-103,"37":-56,"25":32,"-50":-52,"-8":60,"-76":-5,"93":48,"57":-119,"-127":34,"106":125,"-12":15,"84":-16,"61":68,"-72":33,"-44":0,"60":-107,"117":12,"102":-11,"-11":-12,"28":60,"-26":93,"21":122,"92":-52,"65":105,"-119":-1,"-4":10,"-23":93,"-55":-67,"-19":5,"41":46,"2":-3,"-83":-74,"-94":-54,"-47":109,"-42":53,"44":19,"-5":56
[...]
-- !sql_arr_csv_as_json_doris --
-0 {1:1, 0:0} {69:-94, 101:-115, -64:66, -71:101, 0:38, -49:-24,
88:45, -98:20, 5:88, 120:25, -59:-66, -91:92, -113:61, -39:109, -22:-103,
37:-56, 25:32, -50:-52, -8:60, -76:-5, 93:48, 57:-119, -127:34, 106:125,
-12:15, 84:-16, 61:68, -72:33, -44:0, 60:-107, 117:12, 102:-11, -11:-12, 28:60,
-26:93, 21:122, 92:-52, 65:105, -119:-1, -4:10, -23:93, -55:-67, -19:5, 41:46,
2:-3, -83:-74, -94:-54, -47:109, -42:53, 44:19, -5:56, 27:14, 54:124, -84:-69,
86:-124, -20:9, 76:-39, 39:25, 98:7, -52:-1 [...]
+0 {1:1, 0:0} {69:-94, 101:-115, -64:66, -71:101, 0:38, -49:-24,
88:45, -98:20, 5:88, 120:25, -59:-66, -91:92, -113:61, -39:109, -22:-103,
37:-56, 25:32, -50:-52, -8:60, -76:-5, 93:48, 57:-119, -127:34, 106:125,
-12:15, 84:-16, 61:68, -72:33, -44:0, 60:-107, 117:12, 102:-11, -11:-12, 28:60,
-26:93, 21:122, 92:-52, 65:105, -119:-1, -4:10, -23:93, -55:-67, -19:5, 41:46,
2:-3, -83:-74, -94:-54, -47:109, -42:53, 44:19, -5:56, 27:14, 54:124, -84:-69,
86:-124, -20:9, 76:-39, 39:25, 98:7, -52:-1 [...]
-- !sql_arr_csv_as_json_s3 --
[{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"true":true,"false":true},{"tru
[...]
diff --git
a/regression-test/data/datatype_p0/nested_types/query/test_nestedtypes_json_insert_into_with_s3.out
b/regression-test/data/datatype_p0/nested_types/query/test_nestedtypes_json_insert_into_with_s3.out
index 7595d7693c7..21fb51ac176 100644
---
a/regression-test/data/datatype_p0/nested_types/query/test_nestedtypes_json_insert_into_with_s3.out
+++
b/regression-test/data/datatype_p0/nested_types/query/test_nestedtypes_json_insert_into_with_s3.out
@@ -15,19 +15,19 @@
0 {"false":true,"true":true}
{"-64":-114,"5":87,"-128":-34,"-59":100,"-54":-124,"-91":2,"-7":14,"-96":-88,"25":6,"-108":91,"46":89,"93":6,"-3":-21,"78":118,"29":103,"-127":5,"106":75,"121":-63,"84":-107,"-104":-53,"61":-14,"-44":98,"6":-50,"60":-120,"117":86,"38":50,"70":45,"33":39,"65":-43,"-119":-115,"-4":-68,"-36":80,"9":36,"53":95,"-55":-70,"77":115,"-19":-63,"13":60,"-51":23,"41":109,"73":-82,"32":-47,"34":122,"-42":-16,"44":17,"118":-103,"-101":14,"39":31,"-52":-98,"-106":-85,"48":
[...]
-- !sql_arr_json_without_quote_doris --
-0 {0:1, 1:1} {-64:-114, 5:87, -128:-34, -59:100, -54:-124, -91:2,
-7:14, -96:-88, 25:6, -108:91, 46:89, 93:6, -3:-21, 78:118, 29:103, -127:5,
106:75, 121:-63, 84:-107, -104:-53, 61:-14, -44:98, 6:-50, 60:-120, 117:86,
38:50, 70:45, 33:39, 65:-43, -119:-115, -4:-68, -36:80, 9:36, 53:95, -55:-70,
77:115, -19:-63, 13:60, -51:23, 41:109, 73:-82, 32:-47, 34:122, -42:-16, 44:17,
118:-103, -101:14, 39:31, -52:-98, -106:-85, 48:126, 63:92, -46:1, -1:-89,
16:-38, -14:97, -21:-69, 11:-6, 72:57, -3 [...]
+0 {0:1, 1:1} {-64:-114, 5:87, -128:-34, -59:100, -54:-124, -91:2,
-7:14, -96:-88, 25:6, -108:91, 46:89, 93:6, -3:-21, 78:118, 29:103, -127:5,
106:75, 121:-63, 84:-107, -104:-53, 61:-14, -44:98, 6:-50, 60:-120, 117:86,
38:50, 70:45, 33:39, 65:-43, -119:-115, -4:-68, -36:80, 9:36, 53:95, -55:-70,
77:115, -19:-63, 13:60, -51:23, 41:109, 73:-82, 32:-47, 34:122, -42:-16, 44:17,
118:-103, -101:14, 39:31, -52:-98, -106:-85, 48:126, 63:92, -46:1, -1:-89,
16:-38, -14:97, -21:-69, 11:-6, 72:57, -3 [...]
-- !sql_arr_json_without_quote_s3 --
0
[{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true,"true":false},{"false":true
[...]
-- !sql_arr_json_without_quote_doris --
-0 [{0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1,
1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0},
{0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1,
1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0},
{0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1,
1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0},
{0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0 [...]
+0 [{0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1,
1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0},
{0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1,
1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0},
{0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1,
1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0},
{0:1, 1:0}, {0:1, 1:0}, {0:1, 1:0 [...]
-- !sql_arr_json_without_quote_s3 --
0 {"true":[true,false,false],"false":[false,true,false]}
{"101":[-38,-115,35],"-59":[-113,-128,110],"10":[-78,-127,16],"56":[108,-40,124],"-22":[-63,-88,-24],"37":[-7,81,-20],"14":[-114,95,-3],"110":[71,-72,-84],"125":[1,110,56],"-40":[-13,19,80],"20":[-41,-115,107],"46":[82,-104,-30],"57":[116,68,88],"-104":[-64,119,31],"-72":[-65,-44,27],"1":[-127,-73,81],"74":[81,-82,-32],"-99":[-70,37,55],"60":[89,112,124],"-58":[-69,50,-56],"102":[27,-17,-9],"-68":[-112,20,91],"-43":[75,21,76],"33":
[...]
-- !sql_arr_json_without_quote_doris --
-0 {1:[1, 0, 0], 0:[0, 1, 0]} {101:[-38, -115, 35], -59:[-113, -128,
110], 10:[-78, -127, 16], 56:[108, -40, 124], -22:[-63, -88, -24], 37:[-7, 81,
-20], 14:[-114, 95, -3], 110:[71, -72, -84], 125:[1, 110, 56], -40:[-13, 19,
80], 20:[-41, -115, 107], 46:[82, -104, -30], 57:[116, 68, 88], -104:[-64, 119,
31], -72:[-65, -44, 27], 1:[-127, -73, 81], 74:[81, -82, -32], -99:[-70, 37,
55], 60:[89, 112, 124], -58:[-69, 50, -56], 102:[27, -17, -9], -68:[-112, 20,
91], -43:[75, 21, 76], 33:[103, 6 [...]
+0 {1:[1, 0, 0], 0:[0, 1, 0]} {101:[-38, -115, 35], -59:[-113, -128,
110], 10:[-78, -127, 16], 56:[108, -40, 124], -22:[-63, -88, -24], 37:[-7, 81,
-20], 14:[-114, 95, -3], 110:[71, -72, -84], 125:[1, 110, 56], -40:[-13, 19,
80], 20:[-41, -115, 107], 46:[82, -104, -30], 57:[116, 68, 88], -104:[-64, 119,
31], -72:[-65, -44, 27], 1:[-127, -73, 81], 74:[81, -82, -32], -99:[-70, 37,
55], 60:[89, 112, 124], -58:[-69, 50, -56], 102:[27, -17, -9], -68:[-112, 20,
91], -43:[75, 21, 76], 33:[103, 6 [...]
-- !sql_arr --
100
diff --git a/regression-test/data/export/test_struct_export.out
b/regression-test/data/export/test_struct_export.out
index 21141b417d3..0dc07fa7cb6 100644
Binary files a/regression-test/data/export/test_struct_export.out and
b/regression-test/data/export/test_struct_export.out differ
diff --git a/regression-test/data/insert_p0/test_struct_insert.out
b/regression-test/data/insert_p0/test_struct_insert.out
index 7bd960b0e3d..2c202a651e7 100644
Binary files a/regression-test/data/insert_p0/test_struct_insert.out and
b/regression-test/data/insert_p0/test_struct_insert.out differ
diff --git a/regression-test/data/load_p0/stream_load/map_char_test.csv
b/regression-test/data/load_p0/stream_load/map_char_test.csv
deleted file mode 100644
index 4b545d1ea36..00000000000
--- a/regression-test/data/load_p0/stream_load/map_char_test.csv
+++ /dev/null
@@ -1,4 +0,0 @@
-1 {1:"1", 2:"22", 3:"333", 4:"4444", 5:"55555", 6:"666666", 7:"7777777"}
-3 {1:"11", 2:"22", 3:"33", 4:"44", 5:"55", 6:"66", 7:"77"}
-2 {1:"1", 2:"2", 3:"3", 4:"4", 5:"5", 6:"6", 7:"7"}
-4 {1:"111", 2:"22", 3:"333", 4:"444", 5:"55", 6:"66", 7:"777"}
diff --git a/regression-test/data/load_p0/stream_load/test_stream_load.out
b/regression-test/data/load_p0/stream_load/test_stream_load.out
index 505b3110cf2..8816fce94b8 100644
--- a/regression-test/data/load_p0/stream_load/test_stream_load.out
+++ b/regression-test/data/load_p0/stream_load/test_stream_load.out
@@ -6,12 +6,6 @@
-- !sql1 --
2019 9 9 9 7.700 a 2019-09-09
1970-01-01T08:33:39 k7 9.0 9.0
--- !map11 --
-1 {1:"1", 2:"22", 3:"333", 4:"4444", 5:"55555", 6:"666666", 7:"7777777"}
-2 {1:"1", 2:"2", 3:"3", 4:"4", 5:"5", 6:"6", 7:"7"}
-3 {1:"11", 2:"22", 3:"33", 4:"44", 5:"55", 6:"66", 7:"77"}
-4 {1:"111", 2:"22", 3:"333", 4:"444", 5:"55", 6:"66", 7:"777"}
-
-- !all11 --
2500
diff --git
a/regression-test/data/query_p0/aggregate/aggregate_group_by_metric_type.out
b/regression-test/data/query_p0/aggregate/aggregate_group_by_metric_type.out
index d37e5e62bb0..32cdb5a79e9 100644
Binary files
a/regression-test/data/query_p0/aggregate/aggregate_group_by_metric_type.out
and
b/regression-test/data/query_p0/aggregate/aggregate_group_by_metric_type.out
differ
diff --git a/regression-test/suites/load_p0/stream_load/test_stream_load.groovy
b/regression-test/suites/load_p0/stream_load/test_stream_load.groovy
index 0884330f2d3..d9cf9c72938 100644
--- a/regression-test/suites/load_p0/stream_load/test_stream_load.groovy
+++ b/regression-test/suites/load_p0/stream_load/test_stream_load.groovy
@@ -191,8 +191,6 @@ suite("test_stream_load", "p0") {
def tableName7 = "test_unique_key_with_delete"
def tableName8 = "test_array"
def tableName10 = "test_struct"
- def tableName11 = "test_map"
-
sql """ DROP TABLE IF EXISTS ${tableName3} """
sql """ DROP TABLE IF EXISTS ${tableName4} """
sql """ DROP TABLE IF EXISTS ${tableName5} """
@@ -200,7 +198,6 @@ suite("test_stream_load", "p0") {
sql """ DROP TABLE IF EXISTS ${tableName7} """
sql """ DROP TABLE IF EXISTS ${tableName8} """
sql """ DROP TABLE IF EXISTS ${tableName10} """
- sql """ DROP TABLE IF EXISTS ${tableName11} """
sql """
CREATE TABLE IF NOT EXISTS ${tableName3} (
`k1` int(11) NULL,
@@ -284,7 +281,7 @@ suite("test_stream_load", "p0") {
`k4` ARRAY<BIGINT> NULL COMMENT "",
`k5` ARRAY<CHAR> NULL COMMENT "",
`k6` ARRAY<VARCHAR(20)> NULL COMMENT "",
- `k7` ARRAY<DATE> NULL COMMENT "",
+ `k7` ARRAY<DATE> NULL COMMENT "",
`k8` ARRAY<DATETIME> NULL COMMENT "",
`k9` ARRAY<FLOAT> NULL COMMENT "",
`k10` ARRAY<DOUBLE> NULL COMMENT "",
@@ -319,41 +316,6 @@ suite("test_stream_load", "p0") {
);
"""
- sql """
- CREATE TABLE IF NOT EXISTS ${tableName11} (
- `k1` int(11) NULL,
- `k2` map<int, char(7)> NULL
- ) ENGINE=OLAP
- DUPLICATE KEY(`k1`)
- DISTRIBUTED BY HASH(`k1`) BUCKETS 3
- PROPERTIES (
- "replication_allocation" = "tag.location.default: 1"
- );
- """
-
- // load map with specific-length char with non-specific-length data
- streamLoad {
- table "${tableName11}"
-
- set 'column_separator', '\t'
-
- file 'map_char_test.csv'
- time 10000 // limit inflight 10s
-
- check { result, exception, startTime, endTime ->
- if (exception != null) {
- throw exception
- }
- log.info("Stream load result: ${result}".toString())
- def json = parseJson(result)
- assertEquals("success", json.Status.toLowerCase())
- assertEquals(4, json.NumberTotalRows)
- assertEquals(0, json.NumberFilteredRows)
- }
- }
- sql "sync"
- order_qt_map11 "SELECT * FROM ${tableName11} order by k1"
-
// load all columns
streamLoad {
table "${tableName3}"
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]