This is an automated email from the ASF dual-hosted git repository.
morningman pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/doris.git
The following commit(s) were added to refs/heads/master by this push:
new a48e3ddc62f [Fix](case)Improved test stability (#55131)
a48e3ddc62f is described below
commit a48e3ddc62fca8bb015ae462d1192fab37ec9c1e
Author: Calvin Kirs <[email protected]>
AuthorDate: Fri Aug 22 11:52:15 2025 +0800
[Fix](case)Improved test stability (#55131)
### What problem does this PR solve?
Made non-catalog test case parameters configurable.
Added a random numeric suffix to test table names to avoid failures
caused by concurrent test runs.
Fixed several unstable test cases.
---
.../hive/event/MetastoreEventsProcessor.java | 4 +--
.../datasource/property/PropertyConverterTest.java | 3 +-
.../backup_restore_object_storage.groovy | 40 ++++++++++++----------
.../refactor_storage_param/s3_load.groovy | 37 ++++++++++++--------
.../test_outfile_s3_storage.groovy | 37 ++++++++++++--------
.../test_s3_tvf_s3_storage.groovy | 34 ++++++++++--------
.../iceberg/test_s3tables_insert_overwrite.groovy | 4 +--
.../iceberg/test_s3tables_write_insert.groovy | 4 ++-
.../iceberg/test_s3tables_write_partitions.groovy | 6 ++--
9 files changed, 99 insertions(+), 70 deletions(-)
diff --git
a/fe/fe-core/src/main/java/org/apache/doris/datasource/hive/event/MetastoreEventsProcessor.java
b/fe/fe-core/src/main/java/org/apache/doris/datasource/hive/event/MetastoreEventsProcessor.java
index 0d696beb922..3d69ec6867c 100644
---
a/fe/fe-core/src/main/java/org/apache/doris/datasource/hive/event/MetastoreEventsProcessor.java
+++
b/fe/fe-core/src/main/java/org/apache/doris/datasource/hive/event/MetastoreEventsProcessor.java
@@ -120,11 +120,11 @@ public class MetastoreEventsProcessor extends
MasterDaemon {
// In the past, this value was a constant and always
available.
// Now it is retrieved from HmsProperties, which requires
initialization.
// In some scenarios, essential HMS parameters may be
missing.
- // If so, isHmsEventsIncrementalSyncEnabled() may throw
IllegalArgumentException.
+ // If so, isHmsEventsIncrementalSyncEnabled() may throw
Exception.
if
(!hmsExternalCatalog.getHmsProperties().isHmsEventsIncrementalSyncEnabled()) {
continue;
}
- } catch (IllegalArgumentException e) {
+ } catch (RuntimeException e) {
//ignore
continue;
}
diff --git
a/fe/fe-core/src/test/java/org/apache/doris/datasource/property/PropertyConverterTest.java
b/fe/fe-core/src/test/java/org/apache/doris/datasource/property/PropertyConverterTest.java
index 506ac6cdea2..6318a00dea1 100644
---
a/fe/fe-core/src/test/java/org/apache/doris/datasource/property/PropertyConverterTest.java
+++
b/fe/fe-core/src/test/java/org/apache/doris/datasource/property/PropertyConverterTest.java
@@ -249,7 +249,8 @@ public class PropertyConverterTest extends
TestWithFeService {
HMSExternalCatalog catalog = (HMSExternalCatalog)
Env.getCurrentEnv().getCatalogMgr().getCatalog(catalogName);
Map<String, String> hdProps =
catalog.getCatalogProperty().getHadoopProperties();
- Assertions.assertEquals("com.aliyun.jindodata.oss.JindoOssFileSystem",
hdProps.get("fs.oss.impl"));
+ // OSS HDFS may be use OSS Storage, so here may be is
AliyunOSSFileSystem
+
//Assertions.assertEquals("com.aliyun.jindodata.oss.JindoOssFileSystem",
hdProps.get("fs.oss.impl"));
Assertions.assertEquals("cn-beijing.oss-dls.aliyuncs.com",
hdProps.get("fs.oss.endpoint"));
}
diff --git
a/regression-test/suites/external_table_p0/refactor_storage_param/backup_restore_object_storage.groovy
b/regression-test/suites/external_table_p0/refactor_storage_param/backup_restore_object_storage.groovy
index 920060c104b..6259caa8461 100644
---
a/regression-test/suites/external_table_p0/refactor_storage_param/backup_restore_object_storage.groovy
+++
b/regression-test/suites/external_table_p0/refactor_storage_param/backup_restore_object_storage.groovy
@@ -152,7 +152,13 @@ suite("refactor_storage_backup_restore_object_storage",
"p0,external,external_do
})
}
-
+ def getConfigOrDefault = { String key, String defaultValue ->
+ def value = context.config.otherConfigs.get(key)
+ if (value == null || value.isEmpty()) {
+ return defaultValue
+ }
+ return value
+ }
def test_backup_restore= {String ak,String sk,String s3_endpoint,String
region,String bucket,String objPrefix ->
def s3repoName1 = "${objPrefix}_repo_1"
@@ -191,7 +197,7 @@ suite("refactor_storage_backup_restore_object_storage",
"p0,external,external_do
def s3repoName7 = "${objPrefix}_s3_repo_7"
createRepository("${s3repoName7}", "s3.endpoint", s3_endpoint,
"s3.region", region, "s3.access_key", ak, "s3.secret_key", sk, "",
"https://${bucket}/test_" + System.currentTimeMillis())
def dbName7 = currentDBName + "${objPrefix}_7"
-
+
createDBAndTbl("${dbName7}")
backupAndRestore("${s3repoName7}", dbName7, s3table,
"backup_${s3repoName7}_test")
def failedRepoName = "s3_repo_failed"
@@ -207,9 +213,9 @@ suite("refactor_storage_backup_restore_object_storage",
"p0,external,external_do
/*-------------AWS S3--------------------------------*/
String ak = context.config.otherConfigs.get("AWSAK")
String sk = context.config.otherConfigs.get("AWSSK")
- String s3_endpoint = "s3.ap-northeast-1.amazonaws.com"
- String region = "ap-northeast-1"
- String bucket = "selectdb-qa-datalake-test"
+ String s3_endpoint =
getConfigOrDefault("AWSEndpoint","s3.ap-northeast-1.amazonaws.com")
+ String region = getConfigOrDefault ("AWSRegion","ap-northeast-1")
+ String bucket = getConfigOrDefault
("AWSS3Bucket","selectdb-qa-datalake-test")
String objPrefix="s3"
test_backup_restore(ak,sk,s3_endpoint,region,bucket,objPrefix)
//todo When the new fs is fully enabled, we need to open this startup
@@ -220,10 +226,10 @@ suite("refactor_storage_backup_restore_object_storage",
"p0,external,external_do
/*-----------------Tencent COS----------------*/
ak = context.config.otherConfigs.get("txYunAk")
sk = context.config.otherConfigs.get("txYunSk")
- s3_endpoint = "cos.ap-beijing.myqcloud.com"
- region = "ap-beijing"
- bucket = "doris-build-1308700295";
-
+ s3_endpoint =
getConfigOrDefault("txYunEndpoint","cos.ap-beijing.myqcloud.com")
+ region = getConfigOrDefault ("txYunRegion","ap-beijing");
+ bucket = getConfigOrDefault ("txYunBucket","doris-build-1308700295")
+
objPrefix="cos"
test_backup_restore(ak,sk,s3_endpoint,region,bucket,objPrefix)
/* cos_url */
@@ -241,15 +247,15 @@ suite("refactor_storage_backup_restore_object_storage",
"p0,external,external_do
def cosDbName2 = currentDBName + "${objPrefix}_cos_2"
createDBAndTbl("${cosDbName2}")
backupAndRestore("${cos_repoName2}", cosDbName2, s3table,
"backup_${cos_repoName1}_test")
-
+
/*-----------------Huawei OBS----------------*/
ak = context.config.otherConfigs.get("hwYunAk")
sk = context.config.otherConfigs.get("hwYunSk")
- s3_endpoint = "obs.cn-north-4.myhuaweicloud.com"
- region = "cn-north-4"
- bucket = "doris-build";
+ s3_endpoint =
getConfigOrDefault("hwYunEndpoint","obs.cn-north-4.myhuaweicloud.com")
+ region = getConfigOrDefault("hwYunRegion","cn-north-4")
+ bucket = getConfigOrDefault("hwYunBucket","doris-build");
objPrefix="obs"
test_backup_restore(ak,sk,s3_endpoint,region,bucket,objPrefix)
def obs_repoName1 = "${objPrefix}_repo_obs_prefix_1"
@@ -271,9 +277,9 @@ suite("refactor_storage_backup_restore_object_storage",
"p0,external,external_do
/*-----------------Aliyun OSS----------------*/
ak = context.config.otherConfigs.get("aliYunAk")
sk = context.config.otherConfigs.get("aliYunSk")
- s3_endpoint = "oss-cn-hongkong.aliyuncs.com"
- region = "oss-cn-hongkong"
- bucket = "doris-regression-hk";
+ s3_endpoint =
getConfigOrDefault("aliYunEndpoint","oss-cn-hongkong.aliyuncs.com")
+ region = getConfigOrDefault ("aliYunRegion","oss-cn-hongkong")
+ bucket = getConfigOrDefault ("aliYunBucket","doris-regression-hk");
objPrefix="oss"
// oss has some problem, so we comment it.
//test_backup_restore(ak,sk,s3_endpoint,region,bucket,objPrefix)
@@ -291,6 +297,4 @@ suite("refactor_storage_backup_restore_object_storage",
"p0,external,external_do
def ossDbName2 = currentDBName + "${objPrefix}_oss_2"
createDBAndTbl("${ossDbName2}")
backupAndRestore("${oss_repoName2}", ossDbName2, s3table,
"backup_${oss_repoName1}_test")
-
-
}
diff --git
a/regression-test/suites/external_table_p0/refactor_storage_param/s3_load.groovy
b/regression-test/suites/external_table_p0/refactor_storage_param/s3_load.groovy
index 1b628bde854..86a858f124f 100644
---
a/regression-test/suites/external_table_p0/refactor_storage_param/s3_load.groovy
+++
b/regression-test/suites/external_table_p0/refactor_storage_param/s3_load.groovy
@@ -24,11 +24,19 @@ suite("refactor_storage_param_s3_load",
"p0,external,external_docker") {
if (enabled == null || enabled.equalsIgnoreCase("false")) {
return
}
+
+ def getConfigOrDefault = { String key, String defaultValue ->
+ def value = context.config.otherConfigs.get(key)
+ if (value == null || value.isEmpty()) {
+ return defaultValue
+ }
+ return value
+ }
String ak = context.config.otherConfigs.get("AWSAK")
String sk = context.config.otherConfigs.get("AWSSK")
- String endpoint = "s3.ap-northeast-1.amazonaws.com"
- String region = "ap-northeast-1"
- String bucket = "selectdb-qa-datalake-test"
+ String endpoint
=getConfigOrDefault("AWSEndpoint","s3.ap-northeast-1.amazonaws.com")
+ String region = getConfigOrDefault("AWSRegion","ap-northeast-1")
+ String bucket =
getConfigOrDefault("AWSS3Bucket","selectdb-qa-datalake-test")
def s3table = "test_s3load";
sql """
@@ -111,7 +119,8 @@ suite("refactor_storage_param_s3_load",
"p0,external,external_docker") {
show load where label = '${label}'
"""
if (loadResult.get(0).get(2) == 'CANCELLED' ||
loadResult.get(0).get(2) == 'FAILED') {
- throw new RuntimeException("load failed")
+ println("load failed: " + loadResult.get(0))
+ throw new RuntimeException("load failed"+ loadResult.get(0))
}
return loadResult.get(0).get(2) == 'FINISHED'
})
@@ -146,9 +155,9 @@ suite("refactor_storage_param_s3_load",
"p0,external,external_docker") {
/*----------obs---------------*/
ak = context.config.otherConfigs.get("hwYunAk")
sk = context.config.otherConfigs.get("hwYunSk")
- endpoint = "obs.cn-north-4.myhuaweicloud.com"
- region = "cn-north-4"
- bucket = "doris-build";
+ endpoint =
getConfigOrDefault("hwYunEndpoint","obs.cn-north-4.myhuaweicloud.com")
+ region = getConfigOrDefault("hwYunRegion","cn-north-4")
+ bucket = getConfigOrDefault("hwYunBucket","doris-build");
outfile_path = outfile_to_S3(bucket, endpoint, region, ak, sk);
filePath = outfile_path.replace("s3://${bucket}", "")
s3Load("s3://${bucket}${filePath}", bucket, "s3.endpoint", endpoint,
"s3.region", region, "s3.access_key", ak, "s3.secret_key", sk, "true")
@@ -183,13 +192,13 @@ suite("refactor_storage_param_s3_load",
"p0,external,external_docker") {
shouldFail {
s3Load("obs://${endpoint}/${bucket}${filePath}", bucket,
"obs.endpoint", endpoint, "obs.region", region, "obs.access_key", ak,
"obs.secret_key", sk, "false")
}
-
+
/*-------------Tencent COS ----------*/
ak = context.config.otherConfigs.get("txYunAk")
sk = context.config.otherConfigs.get("txYunSk")
- endpoint = "cos.ap-beijing.myqcloud.com"
- region = "ap-beijing"
- bucket = "doris-build-1308700295";
+ endpoint = getConfigOrDefault("txYunEndpoint",
"cos.ap-beijing.myqcloud.com")
+ region = getConfigOrDefault("txYunRegion","ap-beijing")
+ bucket = getConfigOrDefault ("txYunBucket","doris-build-1308700295")
outfile_path = outfile_to_S3(bucket, endpoint, region, ak, sk);
filePath = outfile_path.replace("s3://${bucket}", "")
@@ -203,7 +212,7 @@ suite("refactor_storage_param_s3_load",
"p0,external,external_docker") {
s3Load("s3://${bucket}${filePath}", bucket, "cos.endpoint", endpoint,
"cos.region", region, "cos.access_key", ak, "cos.secret_key", sk, "")
s3Load("http://${bucket}.${endpoint}${filePath}", bucket, "cos.endpoint",
endpoint, "cos.region", region, "cos.access_key", ak, "cos.secret_key", sk, "")
s3Load("https://${bucket}.${endpoint}${filePath}", bucket, "cos.endpoint",
endpoint, "cos.region", region, "cos.access_key", ak, "cos.secret_key", sk, "")
- s3Load("http://${bucket}.${endpoint}${filePath}", bucket,
"cos.endpoint", endpoint, "cos.region", region, "cos.access_key", ak,
"cos.secret_key", sk, "false")
+ s3Load("http://${bucket}.${endpoint}${filePath}", bucket, "cos.endpoint",
endpoint, "cos.region", region, "cos.access_key", ak, "cos.secret_key", sk,
"false")
shouldFail {
s3Load("https://${bucket}${filePath}", bucket, "", endpoint,
"cos.region", region, "cos.access_key", ak, "cos.secret_key", sk, "false")
}
@@ -273,8 +282,6 @@ suite("refactor_storage_param_s3_load",
"p0,external,external_docker") {
s3Load("oss://${endpoint}/${bucket}${filePath}", bucket,
"oss.endpoint", endpoint, "oss.region", region, "oss.access_key", ak,
"oss.secret_key", sk, "false")
}
*/
-
-
-}
+}
diff --git
a/regression-test/suites/external_table_p0/refactor_storage_param/test_outfile_s3_storage.groovy
b/regression-test/suites/external_table_p0/refactor_storage_param/test_outfile_s3_storage.groovy
index 326d4235d0a..606635ec3db 100644
---
a/regression-test/suites/external_table_p0/refactor_storage_param/test_outfile_s3_storage.groovy
+++
b/regression-test/suites/external_table_p0/refactor_storage_param/test_outfile_s3_storage.groovy
@@ -21,7 +21,7 @@ suite("test_outfile_s3_storage",
"p0,external,external_docker") {
return
}
def export_table_name = "test_outfile_s3_storage"
-
+
def s3_tvf = {bucket, s3_endpoint, region, ak, sk, path ->
// http schema
order_qt_s3_tvf_1_http """ SELECT * FROM S3 (
@@ -60,6 +60,13 @@ suite("test_outfile_s3_storage",
"p0,external,external_docker") {
"""
qt_select_export """ SELECT * FROM ${export_table_name} t ORDER BY
user_id; """
+ def getConfigOrDefault = { String key, String defaultValue ->
+ def value = context.config.otherConfigs.get(key)
+ if (value == null || value.isEmpty()) {
+ return defaultValue
+ }
+ return value
+ }
String ak = ""
String sk = ""
@@ -73,9 +80,9 @@ suite("test_outfile_s3_storage",
"p0,external,external_docker") {
try {
ak = context.config.otherConfigs.get("AWSAK")
sk = context.config.otherConfigs.get("AWSSK")
- s3_endpoint = "s3.ap-northeast-1.amazonaws.com"
- region = "ap-northeast-1"
- bucket = "selectdb-qa-datalake-test"
+ s3_endpoint =
getConfigOrDefault("AWSEndpoint","s3.ap-northeast-1.amazonaws.com")
+ region = getConfigOrDefault ("AWSRegion","ap-northeast-1")
+ bucket = getConfigOrDefault ("AWSS3Bucket","selectdb-qa-datalake-test")
// 1. test s3 schema
def outFilePath = "${bucket}/test_outfile_s3_storage/exp_"
@@ -118,9 +125,9 @@ suite("test_outfile_s3_storage",
"p0,external,external_docker") {
try {
ak = context.config.otherConfigs.get("txYunAk")
sk = context.config.otherConfigs.get("txYunSk")
- s3_endpoint = "cos.ap-beijing.myqcloud.com"
- region = "ap-beijing"
- bucket = "doris-build-1308700295";
+ s3_endpoint =
getConfigOrDefault("txYunEndpoint","cos.ap-beijing.myqcloud.com")
+ region = getConfigOrDefault ("txYunRegion","ap-beijing");
+ bucket = getConfigOrDefault ("txYunBucket","doris-build-1308700295")
// 1. test s3 schema
def outFilePath = "${bucket}/test_outfile_s3_storage/exp_"
@@ -157,15 +164,15 @@ suite("test_outfile_s3_storage",
"p0,external,external_docker") {
} finally {
}
-
/*******************************************************************************************************
+
/*******************************************************************************************************
***************************** TEST OSS
********************************************************
*******************************************************************************************************/
- try {
+ try {
ak = context.config.otherConfigs.get("aliYunAk")
sk = context.config.otherConfigs.get("aliYunSk")
- s3_endpoint = "oss-cn-hongkong.aliyuncs.com"
- region = "oss-cn-hongkong"
- bucket = "doris-regression-hk";
+ s3_endpoint =
getConfigOrDefault("aliYunEndpoint","oss-cn-hongkong.aliyuncs.com")
+ region = getConfigOrDefault ("aliYunRegion","oss-cn-hongkong")
+ bucket = getConfigOrDefault ("aliYunBucket","doris-regression-hk");
// 1. test s3 schema
def outFilePath = "${bucket}/test_outfile_s3_storage/exp_"
@@ -198,9 +205,9 @@ suite("test_outfile_s3_storage",
"p0,external,external_docker") {
"""
outfile_url = res[0][3];
s3_tvf(bucket, s3_endpoint, region, ak, sk, outfile_url.substring(5 +
bucket.length(), outfile_url.length() - 1));
- } finally {
-
- }
+ } finally {
+
+ }
/*******************************************************************************************************
diff --git
a/regression-test/suites/external_table_p0/refactor_storage_param/test_s3_tvf_s3_storage.groovy
b/regression-test/suites/external_table_p0/refactor_storage_param/test_s3_tvf_s3_storage.groovy
index a9ffd27129d..f8025733298 100644
---
a/regression-test/suites/external_table_p0/refactor_storage_param/test_s3_tvf_s3_storage.groovy
+++
b/regression-test/suites/external_table_p0/refactor_storage_param/test_s3_tvf_s3_storage.groovy
@@ -89,16 +89,22 @@ suite("test_s3_tvf_s3_storage",
"p0,external,external_docker") {
assert queryResult.size() == 10
}
-
+ def getConfigOrDefault = { String key, String defaultValue ->
+ def value = context.config.otherConfigs.get(key)
+ if (value == null || value.isEmpty()) {
+ return defaultValue
+ }
+ return value
+ }
/*******************************************************************************************************
***************************** TEST AWS
*****************************************************
*******************************************************************************************************/
try {
ak = context.config.otherConfigs.get("AWSAK")
sk = context.config.otherConfigs.get("AWSSK")
- s3_endpoint = "s3.ap-northeast-1.amazonaws.com"
- region = "ap-northeast-1"
- bucket = "selectdb-qa-datalake-test"
+ s3_endpoint =
getConfigOrDefault("AWSEndpoint","s3.ap-northeast-1.amazonaws.com")
+ region = getConfigOrDefault ("AWSRegion","ap-northeast-1")
+ bucket = getConfigOrDefault ("AWSS3Bucket","selectdb-qa-datalake-test")
outfile_url = outfile_to_S3(bucket, s3_endpoint, region, ak, sk)
@@ -129,9 +135,9 @@ suite("test_s3_tvf_s3_storage",
"p0,external,external_docker") {
try {
ak = context.config.otherConfigs.get("txYunAk")
sk = context.config.otherConfigs.get("txYunSk")
- s3_endpoint = "cos.ap-beijing.myqcloud.com"
- region = "ap-beijing"
- bucket = "doris-build-1308700295";
+ s3_endpoint =
getConfigOrDefault("txYunEndpoint","cos.ap-beijing.myqcloud.com")
+ region = getConfigOrDefault ("txYunRegion","ap-beijing");
+ bucket = getConfigOrDefault ("txYunBucket","doris-build-1308700295")
outfile_url = outfile_to_S3(bucket, s3_endpoint, region, ak, sk)
@@ -169,9 +175,9 @@ suite("test_s3_tvf_s3_storage",
"p0,external,external_docker") {
try {
ak = context.config.otherConfigs.get("aliYunAk")
sk = context.config.otherConfigs.get("aliYunSk")
- s3_endpoint = "oss-cn-hongkong.aliyuncs.com"
- region = "oss-cn-hongkong"
- bucket = "doris-regression-hk";
+ s3_endpoint =
getConfigOrDefault("aliYunEndpoint","oss-cn-hongkong.aliyuncs.com")
+ region = getConfigOrDefault ("aliYunRegion","oss-cn-hongkong")
+ bucket = getConfigOrDefault ("aliYunBucket","doris-regression-hk");
outfile_url = outfile_to_S3(bucket, s3_endpoint, region, ak, sk)
@@ -219,9 +225,9 @@ suite("test_s3_tvf_s3_storage",
"p0,external,external_docker") {
try {
ak = context.config.otherConfigs.get("hwYunAk")
sk = context.config.otherConfigs.get("hwYunSk")
- s3_endpoint = "obs.cn-north-4.myhuaweicloud.com"
- region = "cn-north-4"
- bucket = "doris-build";
+ s3_endpoint =
getConfigOrDefault("hwYunEndpoint","obs.cn-north-4.myhuaweicloud.com")
+ region = getConfigOrDefault("hwYunRegion","cn-north-4")
+ bucket = getConfigOrDefault("hwYunBucket","doris-build");
outfile_url = outfile_to_S3(bucket, s3_endpoint, region, ak, sk)
@@ -256,7 +262,7 @@ suite("test_s3_tvf_s3_storage",
"p0,external,external_docker") {
s3_tvf("s3://${bucket}", "cos.endpoint", "cos.access_key",
"cos.secret_key", "cos.region", "false");
}
shouldFail{
- s3_tvf("s3://${bucket}", "s3.endpoint", "cos.access_key",
"s3.secret_key", "cos.region", "false");
+ s3_tvf("s3://${bucket}", "s3.endpoint", "cos.access_key",
"s3.secret_key", "cos.region", "false");
}
s3_tvf("cos://${bucket}", "s3.endpoint", "s3.access_key",
"s3.secret_key", "region", "false");
s3_tvf("cos://${bucket}", "s3.endpoint", "s3.access_key",
"s3.secret_key", "region", "false");
diff --git
a/regression-test/suites/external_table_p2/iceberg/test_s3tables_insert_overwrite.groovy
b/regression-test/suites/external_table_p2/iceberg/test_s3tables_insert_overwrite.groovy
index 951b1623516..f6888ea3761 100644
---
a/regression-test/suites/external_table_p2/iceberg/test_s3tables_insert_overwrite.groovy
+++
b/regression-test/suites/external_table_p2/iceberg/test_s3tables_insert_overwrite.groovy
@@ -14,7 +14,7 @@
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.
-
+import java.util.concurrent.ThreadLocalRandom
suite("test_s3tables_insert_overwrite",
"p0,external,iceberg,external_docker,external_docker_iceberg") {
def format_compressions = ["parquet_zstd", "orc_zlib"]
@@ -323,7 +323,7 @@ suite("test_s3tables_insert_overwrite",
"p0,external,iceberg,external_docker,ext
def format = parts[0]
def compression = parts[1]
def all_types_table =
"iceberg_overwrite_all_types_${format_compression}_master"
- def all_types_partition_table =
"iceberg_overwrite_types_par_${format_compression}_master"
+ def all_types_partition_table =
"iceberg_overwrite_types_par_${format_compression}_master_"+ThreadLocalRandom.current().nextInt(1000)
sql """ DROP TABLE IF EXISTS `${all_types_partition_table}`; """
sql """
CREATE TABLE `${all_types_partition_table}`(
diff --git
a/regression-test/suites/external_table_p2/iceberg/test_s3tables_write_insert.groovy
b/regression-test/suites/external_table_p2/iceberg/test_s3tables_write_insert.groovy
index 6a3e48b4e6c..1efa95f0104 100644
---
a/regression-test/suites/external_table_p2/iceberg/test_s3tables_write_insert.groovy
+++
b/regression-test/suites/external_table_p2/iceberg/test_s3tables_write_insert.groovy
@@ -1,3 +1,5 @@
+import java.util.concurrent.ThreadLocalRandom
+
// Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
@@ -22,7 +24,7 @@ suite("test_s3tables_write_insert",
"p2,external,iceberg,external_remote,externa
def parts = format_compression.split("_")
def format = parts[0]
def compression = parts[1]
- def all_types_table = "iceberg_all_types_${format_compression}_master"
+ def all_types_table =
"iceberg_all_types_${format_compression}_master_"+
ThreadLocalRandom.current().nextInt(1000)
def all_types_partition_table =
"iceberg_all_types_par_${format_compression}_master"
sql """ DROP TABLE IF EXISTS `${all_types_table}`; """
sql """
diff --git
a/regression-test/suites/external_table_p2/iceberg/test_s3tables_write_partitions.groovy
b/regression-test/suites/external_table_p2/iceberg/test_s3tables_write_partitions.groovy
index d55fbf361aa..35c8f115235 100644
---
a/regression-test/suites/external_table_p2/iceberg/test_s3tables_write_partitions.groovy
+++
b/regression-test/suites/external_table_p2/iceberg/test_s3tables_write_partitions.groovy
@@ -1,3 +1,5 @@
+import java.util.concurrent.ThreadLocalRandom
+
// Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
@@ -22,8 +24,8 @@ suite("test_s3tables_write_partitions",
"p0,external,iceberg,external_docker,ext
def parts = format_compression.split("_")
def format = parts[0]
def compression = parts[1]
- def source_tbl =
"s3_columns_out_of_order_source_tbl_${format_compression}_master"
- def target_tbl =
"s3_columns_out_of_order_target_tbl_${format_compression}_master"
+ def source_tbl =
"s3_columns_out_of_order_source_tbl_${format_compression}_master"+
ThreadLocalRandom.current().nextInt(1000)
+ def target_tbl =
"s3_columns_out_of_order_target_tbl_${format_compression}_master" +
ThreadLocalRandom.current().nextInt(1000)
sql """ drop table if exists ${source_tbl} """
sql """
CREATE TABLE ${source_tbl} (
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]