This is an automated email from the ASF dual-hosted git repository.
yiguolei pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/doris.git
The following commit(s) were added to refs/heads/master by this push:
new 01c94a554d9 [fix](autoinc) Fix broker load when target table has
autoinc column (#28402)
01c94a554d9 is described below
commit 01c94a554d9c5739b21b2b565ef2ef0f2f11a04e
Author: bobhan1 <[email protected]>
AuthorDate: Thu Dec 14 18:02:54 2023 +0800
[fix](autoinc) Fix broker load when target table has autoinc column (#28402)
---
.../data_case/autoinc/auto_inc_basic.csv | 9 ++
.../data_case/autoinc/auto_inc_with_null.csv | 9 ++
.../doris/load/loadv2/LoadingTaskPlanner.java | 1 +
.../hive/test_autoinc_broker_load.out | 45 ++++++++
.../hive/test_autoinc_broker_load.groovy | 113 +++++++++++++++++++++
5 files changed, 177 insertions(+)
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/preinstalled_data/data_case/autoinc/auto_inc_basic.csv
b/docker/thirdparties/docker-compose/hive/scripts/preinstalled_data/data_case/autoinc/auto_inc_basic.csv
new file mode 100644
index 00000000000..d4820564703
--- /dev/null
+++
b/docker/thirdparties/docker-compose/hive/scripts/preinstalled_data/data_case/autoinc/auto_inc_basic.csv
@@ -0,0 +1,9 @@
+Bob, 100
+Alice, 200
+Tom, 300
+Test, 400
+Carter, 500
+Smith, 600
+Beata, 700
+Doris, 800
+Nereids, 900
\ No newline at end of file
diff --git
a/docker/thirdparties/docker-compose/hive/scripts/preinstalled_data/data_case/autoinc/auto_inc_with_null.csv
b/docker/thirdparties/docker-compose/hive/scripts/preinstalled_data/data_case/autoinc/auto_inc_with_null.csv
new file mode 100644
index 00000000000..14390ca3cfd
--- /dev/null
+++
b/docker/thirdparties/docker-compose/hive/scripts/preinstalled_data/data_case/autoinc/auto_inc_with_null.csv
@@ -0,0 +1,9 @@
+null, Bob, 100
+null, Alice, 200
+null, Tom, 300
+null, Test, 400
+4, Carter, 500
+5, Smith, 600
+6, Beata, 700
+7, Doris, 800
+8, Nereids, 900
\ No newline at end of file
diff --git
a/fe/fe-core/src/main/java/org/apache/doris/load/loadv2/LoadingTaskPlanner.java
b/fe/fe-core/src/main/java/org/apache/doris/load/loadv2/LoadingTaskPlanner.java
index e2f3bdcbe9e..7a7687329b2 100644
---
a/fe/fe-core/src/main/java/org/apache/doris/load/loadv2/LoadingTaskPlanner.java
+++
b/fe/fe-core/src/main/java/org/apache/doris/load/loadv2/LoadingTaskPlanner.java
@@ -154,6 +154,7 @@ public class LoadingTaskPlanner {
slotDesc.setIsMaterialized(true);
slotDesc.setColumn(col);
slotDesc.setIsNullable(col.isAllowNull());
+ slotDesc.setAutoInc(col.isAutoInc());
SlotDescriptor scanSlotDesc =
descTable.addSlotDescriptor(scanTupleDesc);
scanSlotDesc.setIsMaterialized(true);
scanSlotDesc.setColumn(col);
diff --git
a/regression-test/data/external_table_p0/hive/test_autoinc_broker_load.out
b/regression-test/data/external_table_p0/hive/test_autoinc_broker_load.out
new file mode 100644
index 00000000000..cf14e8ea6a8
--- /dev/null
+++ b/regression-test/data/external_table_p0/hive/test_autoinc_broker_load.out
@@ -0,0 +1,45 @@
+-- This file is automatically generated. You should know what you did if you
want to edit this
+-- !sql --
+0 Bob 100
+1 Alice 200
+2 Tom 300
+3 Test 400
+4 Carter 500
+5 Smith 600
+6 Beata 700
+7 Doris 800
+8 Nereids 900
+
+-- !sql --
+0 Bob 123
+1 Alice 200
+2 Tom 323
+3 Test 400
+4 Carter 523
+5 Smith 600
+6 Beata 700
+7 Doris 800
+8 Nereids 900
+
+-- !sql --
+0 Bob 100
+1 Alice 200
+2 Tom 300
+3 Test 400
+4 Carter 500
+5 Smith 600
+6 Beata 700
+7 Doris 800
+8 Nereids 900
+
+-- !sql --
+0 Bob 123
+1 Alice 200
+2 Tom 323
+3 Test 400
+4 Carter 523
+5 Smith 600
+6 Beata 700
+7 Doris 800
+8 Nereids 900
+
diff --git
a/regression-test/suites/external_table_p0/hive/test_autoinc_broker_load.groovy
b/regression-test/suites/external_table_p0/hive/test_autoinc_broker_load.groovy
new file mode 100644
index 00000000000..540714b3071
--- /dev/null
+++
b/regression-test/suites/external_table_p0/hive/test_autoinc_broker_load.groovy
@@ -0,0 +1,113 @@
+
+// Licensed to the Apache Software Foundation (ASF) under one
+// or more contributor license agreements. See the NOTICE file
+// distributed with this work for additional information
+// regarding copyright ownership. The ASF licenses this file
+// to you under the Apache License, Version 2.0 (the
+// "License"); you may not use this file except in compliance
+// with the License. You may obtain a copy of the License at
+//
+// http://www.apache.org/licenses/LICENSE-2.0
+//
+// Unless required by applicable law or agreed to in writing,
+// software distributed under the License is distributed on an
+// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+// KIND, either express or implied. See the License for the
+// specific language governing permissions and limitations
+// under the License.
+
+suite("test_autoinc_broker_load",
"p0,external,hive,external_docker,external_docker_hive") {
+
+ String enabled = context.config.otherConfigs.get("enableHiveTest")
+ if (enabled != null && enabled.equalsIgnoreCase("true")) {
+ brokerName = getBrokerName()
+ hdfsUser = getHdfsUser()
+ hdfsPasswd = getHdfsPasswd()
+ hdfs_port = context.config.otherConfigs.get("hdfs_port")
+ externalEnvIp = context.config.otherConfigs.get("externalEnvIp")
+
+ def test_dir = "user/doris/preinstalled_data/data_case/autoinc"
+
+ def load_from_hdfs = {columns, testTable, label, testFile, format,
brokerName, hdfsUser, hdfsPasswd ->
+ def result1= sql """ LOAD LABEL ${label} (
+ DATA
INFILE("hdfs://${externalEnvIp}:${hdfs_port}/${test_dir}/${testFile}")
+ INTO TABLE ${testTable}
+ COLUMNS TERMINATED BY ","
+ (${columns})
+ ) with HDFS (
+
"fs.defaultFS"="hdfs://${externalEnvIp}:${hdfs_port}",
+ "hadoop.username"="${hdfsUser}")
+ PROPERTIES (
+ "timeout"="1200",
+ "max_filter_ratio"="0");"""
+ assertTrue(result1.size() == 1)
+ assertTrue(result1[0].size() == 1)
+ assertTrue(result1[0][0] == 0, "Query OK, 0 rows affected")
+ }
+
+ def wait_for_load_result = {checklabel, testTable ->
+ max_try_milli_secs = 10000
+ while(max_try_milli_secs) {
+ result = sql "show load where label = '${checklabel}'"
+ if(result[0][2] == "FINISHED") {
+ break
+ } else {
+ sleep(1000) // wait 1 second every time
+ max_try_milli_secs -= 1000
+ if(max_try_milli_secs <= 0) {
+ assertEquals(1, 2)
+ }
+ }
+ }
+ }
+
+ table = "test_autoinc_broker_load"
+ sql "drop table if exists ${table}"
+ sql """ CREATE TABLE IF NOT EXISTS `${table}` (
+ `id` BIGINT NOT NULL AUTO_INCREMENT COMMENT "用户 ID",
+ `name` varchar(65533) NOT NULL COMMENT "用户姓名",
+ `value` int(11) NOT NULL COMMENT "用户得分"
+ ) ENGINE=OLAP
+ UNIQUE KEY(`id`)
+ COMMENT "OLAP"
+ DISTRIBUTED BY HASH(`id`) BUCKETS 1
+ PROPERTIES (
+ "replication_allocation" = "tag.location.default: 1",
+ "in_memory" = "false",
+ "storage_format" = "V2",
+ "enable_unique_key_merge_on_write" = "true") """
+
+ def test_load_label = UUID.randomUUID().toString().replaceAll("-", "")
+ load_from_hdfs("name, value", table, test_load_label,
"auto_inc_basic.csv", "csv", brokerName, hdfsUser, hdfsPasswd)
+ wait_for_load_result(test_load_label, table)
+ qt_sql "select * from ${table};"
+ sql """ insert into ${table} values(0, "Bob", 123), (2, "Tom", 323),
(4, "Carter", 523);"""
+ qt_sql "select * from ${table} order by id"
+ sql "drop table if exists ${table};"
+
+
+ table = "test_autoinc_broker_load2"
+ sql "drop table if exists ${table}"
+ sql """ CREATE TABLE IF NOT EXISTS `${table}` (
+ `id` BIGINT NOT NULL AUTO_INCREMENT COMMENT "用户 ID",
+ `name` varchar(65533) NOT NULL COMMENT "用户姓名",
+ `value` int(11) NOT NULL COMMENT "用户得分"
+ ) ENGINE=OLAP
+ UNIQUE KEY(`id`)
+ COMMENT "OLAP"
+ DISTRIBUTED BY HASH(`id`) BUCKETS 1
+ PROPERTIES (
+ "replication_allocation" = "tag.location.default: 1",
+ "in_memory" = "false",
+ "storage_format" = "V2",
+ "enable_unique_key_merge_on_write" = "true");"""
+ test_load_label = UUID.randomUUID().toString().replaceAll("-", "")
+ load_from_hdfs("id, name, value", table, test_load_label,
"auto_inc_with_null.csv", "csv", brokerName, hdfsUser, hdfsPasswd)
+ wait_for_load_result(test_load_label, table)
+ sql "sync"
+ qt_sql "select * from ${table};"
+ sql """ insert into ${table} values(0, "Bob", 123), (2, "Tom", 323),
(4, "Carter", 523);"""
+ qt_sql "select * from ${table} order by id"
+ sql "drop table if exists ${table};"
+ }
+}
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]