This is an automated email from the ASF dual-hosted git repository.
dataroaring pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/doris.git
The following commit(s) were added to refs/heads/branch-3.0 by this push:
new c051c9d7faa branch-3.0: [chore](regression-test) move
test_cold_data_compaction to `nonConcurrent` #50699 (#52066)
c051c9d7faa is described below
commit c051c9d7faa06b0f3a09fae72caa114ee116302e
Author: github-actions[bot]
<41898282+github-actions[bot]@users.noreply.github.com>
AuthorDate: Mon Jun 23 22:01:01 2025 +0800
branch-3.0: [chore](regression-test) move test_cold_data_compaction to
`nonConcurrent` #50699 (#52066)
Cherry-picked from #50699
Co-authored-by: yagagagaga <[email protected]>
---
.../cold_heat_separation/cold_data_compaction.groovy | 16 ++++++++++++----
1 file changed, 12 insertions(+), 4 deletions(-)
diff --git
a/regression-test/suites/cold_heat_separation/cold_data_compaction.groovy
b/regression-test/suites/cold_heat_separation/cold_data_compaction.groovy
index bf9e33e7528..c80e39b8a9d 100644
--- a/regression-test/suites/cold_heat_separation/cold_data_compaction.groovy
+++ b/regression-test/suites/cold_heat_separation/cold_data_compaction.groovy
@@ -18,7 +18,7 @@
import com.amazonaws.services.s3.model.ListObjectsRequest
import java.util.function.Supplier
-suite("test_cold_data_compaction") {
+suite("test_cold_data_compaction", "nonConcurrent") {
def retryUntilTimeout = { int timeoutSecond, Supplier<Boolean> closure ->
long start = System.currentTimeMillis()
while (true) {
@@ -87,7 +87,7 @@ suite("test_cold_data_compaction") {
"""
// wait until files upload to S3
- retryUntilTimeout(1800, {
+ retryUntilTimeout(3600, {
def res = sql_return_maparray "show data from t_recycle_in_s3"
String size = ""
String remoteSize = ""
@@ -112,9 +112,13 @@ suite("test_cold_data_compaction") {
// trigger cold data compaction
sql """alter table t_recycle_in_s3 set ("disable_auto_compaction" =
"false")"""
+ def v =
get_be_param("disable_auto_compaction").values().toArray()[0].toString()
+ if ("true" == v) {
+ set_be_param("disable_auto_compaction", "false")
+ }
// wait until compaction finish
- retryUntilTimeout(1800, {
+ retryUntilTimeout(3600, {
def filesAfterCompaction = getS3Client().listObjects(
new
ListObjectsRequest().withBucketName(getS3BucketName()).withPrefix(s3Prefix+
"/data/${tabletId}")).getObjectSummaries()
logger.info("t_recycle_in_s3's remote file number is
${filesAfterCompaction.size()}")
@@ -122,8 +126,12 @@ suite("test_cold_data_compaction") {
return filesAfterCompaction.size() == 2
})
+ if ("true" == v) {
+ set_be_param("disable_auto_compaction", "true")
+ }
+
sql "drop table t_recycle_in_s3 force"
- retryUntilTimeout(1800, {
+ retryUntilTimeout(3600, {
def filesAfterDrop = getS3Client().listObjects(
new
ListObjectsRequest().withBucketName(getS3BucketName()).withPrefix(s3Prefix+
"/data/${tabletId}")).getObjectSummaries()
logger.info("after drop t_recycle_in_s3, remote file number is
${filesAfterDrop.size()}")
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]