This is an automated email from the ASF dual-hosted git repository.
zhangduo pushed a commit to branch HBASE-29930
in repository https://gitbox.apache.org/repos/asf/hbase.git
The following commit(s) were added to refs/heads/HBASE-29930 by this push:
new b26c47d818c fix
b26c47d818c is described below
commit b26c47d818c1b0be742152873a5d4587c30d114b
Author: Duo Zhang <[email protected]>
AuthorDate: Sat Feb 28 19:32:08 2026 +0800
fix
---
.../integration-test/integration-test.Jenkinsfile | 47 +++++++++++-----------
1 file changed, 24 insertions(+), 23 deletions(-)
diff --git a/dev-support/integration-test/integration-test.Jenkinsfile
b/dev-support/integration-test/integration-test.Jenkinsfile
index 521d0ad011d..16defa5defc 100644
--- a/dev-support/integration-test/integration-test.Jenkinsfile
+++ b/dev-support/integration-test/integration-test.Jenkinsfile
@@ -111,11 +111,11 @@ pipeline {
values getHadoopVersions(env.HADOOP2_VERSIONS,
env.HADOOP3_VERSIONS)
}
}
+ environment {
+ BASEDIR = "${env.WORKSPACE}/component"
+ OUTPUT_DIR = "output-integration-hadoop-${env.HADOOP_VERSION}"
+ }
stages {
- environment {
- BASEDIR = "${env.WORKSPACE}/component"
- OUTPUT_DIR = "output-integration-hadoop-${env.HADOOP_VERSION}"
- }
stage('scm-checkout') {
steps {
sh '''#!/bin/bash -e
@@ -136,25 +136,26 @@ pipeline {
}
}
stage('download hadoop') {
- dir("downloads-hadoop") {
- sh '''#!/bin/bash -e
- echo "Make sure we have a directory for downloading
dependencies: $(pwd)"
- '''
- sh '''#!/bin/bash -e
-
- echo "Ensure we have a copy of Hadoop ${HADOOP_VERSION}"
-
"${WORKSPACE}/component/dev-support/jenkins-scripts/cache-apache-project-artifact.sh"
\
- --working-dir "${WORKSPACE}/downloads-hadoop" \
- --keys 'https://downloads.apache.org/hadoop/common/KEYS' \
- --verify-tar-gz \
- "${WORKSPACE}/hadoop-${HADOOP_VERSION}-bin.tar.gz" \
-
"hadoop/common/hadoop-${HADOOP_VERSION}/hadoop-${HADOOP_VERSION}.tar.gz"
- for stale in $(ls -1 "${WORKSPACE}"/hadoop-*.tar.gz | grep -v
${HADOOP_VERSION}); do
- echo "Delete stale hadoop cache ${stale}"
- rm -rf $stale
- done
- '''
- } // dir
+ steps {
+ dir("downloads-hadoop") {
+ sh '''#!/bin/bash -e
+ echo "Make sure we have a directory for downloading
dependencies: $(pwd)"
+ '''
+ sh '''#!/bin/bash -e
+ echo "Ensure we have a copy of Hadoop ${HADOOP_VERSION}"
+
"${WORKSPACE}/component/dev-support/jenkins-scripts/cache-apache-project-artifact.sh"
\
+ --working-dir "${WORKSPACE}/downloads-hadoop" \
+ --keys 'https://downloads.apache.org/hadoop/common/KEYS' \
+ --verify-tar-gz \
+ "${WORKSPACE}/hadoop-${HADOOP_VERSION}-bin.tar.gz" \
+
"hadoop/common/hadoop-${HADOOP_VERSION}/hadoop-${HADOOP_VERSION}.tar.gz"
+ for stale in $(ls -1 "${WORKSPACE}"/hadoop-*.tar.gz | grep
-v ${HADOOP_VERSION}); do
+ echo "Delete stale hadoop cache ${stale}"
+ rm -rf $stale
+ done
+ '''
+ } // dir
+ } // steps
} // download hadoop
stage('integration test ') {
steps {