This is an automated email from the ASF dual-hosted git repository.

zhangduo pushed a commit to branch HBASE-29930
in repository https://gitbox.apache.org/repos/asf/hbase.git


The following commit(s) were added to refs/heads/HBASE-29930 by this push:
     new 10dd9f2d4a8 fix
10dd9f2d4a8 is described below

commit 10dd9f2d4a89f46a51441d2e132b05edc4f232a2
Author: Duo Zhang <[email protected]>
AuthorDate: Fri Feb 27 23:43:06 2026 +0800

    fix
---
 dev-support/integration-test.Jenkinsfile | 23 ++++++++++++-----------
 1 file changed, 12 insertions(+), 11 deletions(-)

diff --git a/dev-support/integration-test.Jenkinsfile 
b/dev-support/integration-test.Jenkinsfile
index 0e627900946..b4632270ff8 100644
--- a/dev-support/integration-test.Jenkinsfile
+++ b/dev-support/integration-test.Jenkinsfile
@@ -25,13 +25,10 @@ pipeline {
         label 'hbase'
       }
     }
-    stage ('scm-checkout') {
-      steps {
+    steps {
         dir('component') {
           checkout scm
         }
-      }
-    }
     parallel {
       stage ('hadoop 2 cache') {
         steps {
@@ -100,6 +97,7 @@ pipeline {
         } // steps
       } // stage ('hadoop 3 cache')
     } // parallel
+    } // steps
   } // stage ('thirdparty installs')
   // This is meant to mimic what a release manager will do to create RCs.
   // See http://hbase.apache.org/book.html#maven.release
@@ -171,6 +169,7 @@ pipeline {
       BASEDIR = "${env.WORKSPACE}/component"
       BRANCH = "${env.BRANCH_NAME}"
     }
+    steps {
     parallel {
       stage('hadoop 2 integration test') {
         agent {
@@ -321,9 +320,10 @@ pipeline {
             fi
             echo "(/) {color:green}+1 client integration test for 
${HADOOP_VERSION} with shaded hadoop client {color}" >> 
${OUTPUT_DIR}/commentfile
           '''
-        }
-      }
-    }
+        } // hadoop 3 integration test
+      } // hadoop 3 matrix
+    } // parallel
+    } // steps
   } // stage integration test
   }
   post {
@@ -349,10 +349,7 @@ pipeline {
             echo "No hbase-src.tar.gz, skipping"
           fi
         '''
-        archiveArtifacts artifacts: 'output-srctarball/*'
-        archiveArtifacts artifacts: 'output-srctarball/**/*'
-        archiveArtifacts artifacts: 'output-integration-*/*'
-        archiveArtifacts artifacts: 'output-integration-*/**/*'
+
         def results = []
         results.add('output-srctarball/commentfile')
         
results.add("output-integration-hadoop-${env.HADOOP_VERSION}/commentfile")
@@ -389,6 +386,10 @@ pipeline {
           jiraComment issueKey: currentIssue, body: comment
         }
       } // scripts
+        archiveArtifacts artifacts: 'output-srctarball/*'
+        archiveArtifacts artifacts: 'output-srctarball/**/*'
+        archiveArtifacts artifacts: 'output-integration-*/*'
+        archiveArtifacts artifacts: 'output-integration-*/**/*'
     } // always
   } // post
 }

Reply via email to