This is an automated email from the ASF dual-hosted git repository.

zhangduo pushed a commit to branch HBASE-29930
in repository https://gitbox.apache.org/repos/asf/hbase.git


The following commit(s) were added to refs/heads/HBASE-29930 by this push:
     new fcd96d4167a fix
fcd96d4167a is described below

commit fcd96d4167a37ba3e85d796f894ddc162dac2cd7
Author: Duo Zhang <[email protected]>
AuthorDate: Sat Feb 28 00:37:44 2026 +0800

    fix
---
 dev-support/integration-test.Jenkinsfile | 409 ++++++++++++++++---------------
 1 file changed, 206 insertions(+), 203 deletions(-)

diff --git a/dev-support/integration-test.Jenkinsfile 
b/dev-support/integration-test.Jenkinsfile
index 922f56e10e1..9c4d0454623 100644
--- a/dev-support/integration-test.Jenkinsfile
+++ b/dev-support/integration-test.Jenkinsfile
@@ -18,8 +18,9 @@ pipeline {
   parameters {
     booleanParam(name: 'DEBUG', defaultValue: false, description: 'Produce a 
lot more meta-information.')
   }
+
   stages {
-    stage('thirdparty installs') {
+    stage('thirdparty installs - checkout') {
       agent {
         node {
           label 'hbase'
@@ -29,71 +30,75 @@ pipeline {
         dir('component') {
           checkout scm
         }
-        parallel {
-          stage('hadoop 2 cache') {
-            agent {
-              node {
-                label 'hbase'
-              }
+      }
+    }
+
+    stage('thirdparty installs - downloads') {
+      parallel {
+        stage('hadoop 2 cache') {
+          agent {
+            node {
+              label 'hbase'
             }
-            steps {
-              dir('downloads-hadoop-2') {
-                sh '''#!/usr/bin/env bash
-                  echo "Make sure we have a directory for downloading 
dependencies: $(pwd)"
-                '''
-              }
+          }
+          steps {
+            dir('downloads-hadoop-2') {
               sh '''#!/usr/bin/env bash
-                set -e
-                echo "Ensure we have a copy of Hadoop ${HADOOP2_VERSION}"
-                
"${WORKSPACE}/component/dev-support/jenkins-scripts/cache-apache-project-artifact.sh"
 \
-                  --working-dir "${WORKSPACE}/downloads-hadoop-2" \
-                  --keys 'https://downloads.apache.org/hadoop/common/KEYS' \
-                  --verify-tar-gz \
-                  "${WORKSPACE}/hadoop-${HADOOP2_VERSION}-bin.tar.gz" \
-                  
"hadoop/common/hadoop-${HADOOP2_VERSION}/hadoop-${HADOOP2_VERSION}.tar.gz"
-                for stale in $(ls -1 "${WORKSPACE}"/hadoop-2*.tar.gz | grep -v 
${HADOOP2_VERSION}); do
-                  echo "Delete stale hadoop 2 cache ${stale}"
-                  rm -rf $stale
-                done
+                echo "Make sure we have a directory for downloading 
dependencies: $(pwd)"
               '''
-              stash name: 'hadoop-2', includes: 
"hadoop-${HADOOP2_VERSION}-bin.tar.gz"
             }
+            sh '''#!/usr/bin/env bash
+              set -e
+              echo "Ensure we have a copy of Hadoop ${HADOOP2_VERSION}"
+              
"${WORKSPACE}/component/dev-support/jenkins-scripts/cache-apache-project-artifact.sh"
 \
+                --working-dir "${WORKSPACE}/downloads-hadoop-2" \
+                --keys 'https://downloads.apache.org/hadoop/common/KEYS' \
+                --verify-tar-gz \
+                "${WORKSPACE}/hadoop-${HADOOP2_VERSION}-bin.tar.gz" \
+                
"hadoop/common/hadoop-${HADOOP2_VERSION}/hadoop-${HADOOP2_VERSION}.tar.gz"
+              for stale in $(ls -1 "${WORKSPACE}"/hadoop-2*.tar.gz | grep -v 
${HADOOP2_VERSION}); do
+                echo "Delete stale hadoop 2 cache ${stale}"
+                rm -rf $stale
+              done
+            '''
+            stash name: 'hadoop-2', includes: 
"hadoop-${HADOOP2_VERSION}-bin.tar.gz"
           }
-          stage('hadoop 3 cache') {
-            agent {
-              node {
-                label 'hbase'
-              }
+        }
+
+        stage('hadoop 3 cache') {
+          agent {
+            node {
+              label 'hbase'
             }
-            steps {
-              script {
-                hadoop3_versions = env.HADOOP3_VERSIONS.split(",");
-                env.HADOOP3_VERSIONS_REGEX = "[" + hadoop3_versions.join("|") 
+ "]";
-                for (hadoop3_version in hadoop3_versions) {
-                  env.HADOOP3_VERSION = hadoop3_version;
-                  echo "env.HADOOP3_VERSION" + env.hadoop3_version;
-                  stage('Hadoop 3 cache inner stage') {
-                    dir("downloads-hadoop-${HADOOP3_VERSION}") {
-                      sh '''#!/usr/bin/env bash
-                        echo "Make sure we have a directory for downloading 
dependencies: $(pwd)"
-                      '''
-                    }
+          }
+          steps {
+            script {
+              hadoop3_versions = env.HADOOP3_VERSIONS.split(",");
+              env.HADOOP3_VERSIONS_REGEX = "[" + hadoop3_versions.join("|") + 
"]";
+              for (hadoop3_version in hadoop3_versions) {
+                env.HADOOP3_VERSION = hadoop3_version;
+                echo "env.HADOOP3_VERSION" + env.hadoop3_version;
+                stage('Hadoop 3 cache inner stage') {
+                  dir("downloads-hadoop-${HADOOP3_VERSION}") {
                     sh '''#!/usr/bin/env bash
-                      set -e
-                      echo "Ensure we have a copy of Hadoop ${HADOOP3_VERSION}"
-                      
"${WORKSPACE}/component/dev-support/jenkins-scripts/cache-apache-project-artifact.sh"
 \
-                        --working-dir 
"${WORKSPACE}/downloads-hadoop-${HADOOP3_VERSION}" \
-                        --keys 
'https://downloads.apache.org/hadoop/common/KEYS' \
-                        --verify-tar-gz \
-                        "${WORKSPACE}/hadoop-${HADOOP3_VERSION}-bin.tar.gz" \
-                        
"hadoop/common/hadoop-${HADOOP3_VERSION}/hadoop-${HADOOP3_VERSION}.tar.gz"
-                      for stale in $(ls -1 "${WORKSPACE}"/hadoop-3*.tar.gz | 
grep -v ${HADOOP3_VERSION}); do
-                        echo "Delete stale hadoop 3 cache ${stale}"
-                        rm -rf $stale
-                      done
+                      echo "Make sure we have a directory for downloading 
dependencies: $(pwd)"
                     '''
-                    stash name: "hadoop-${HADOOP3_VERSION}", includes: 
"hadoop-${HADOOP3_VERSION}-bin.tar.gz"
                   }
+                  sh '''#!/usr/bin/env bash
+                    set -e
+                    echo "Ensure we have a copy of Hadoop ${HADOOP3_VERSION}"
+                    
"${WORKSPACE}/component/dev-support/jenkins-scripts/cache-apache-project-artifact.sh"
 \
+                      --working-dir 
"${WORKSPACE}/downloads-hadoop-${HADOOP3_VERSION}" \
+                      --keys 'https://downloads.apache.org/hadoop/common/KEYS' 
\
+                      --verify-tar-gz \
+                      "${WORKSPACE}/hadoop-${HADOOP3_VERSION}-bin.tar.gz" \
+                      
"hadoop/common/hadoop-${HADOOP3_VERSION}/hadoop-${HADOOP3_VERSION}.tar.gz"
+                    for stale in $(ls -1 "${WORKSPACE}"/hadoop-3*.tar.gz | 
grep -v ${HADOOP3_VERSION}); do
+                      echo "Delete stale hadoop 3 cache ${stale}"
+                      rm -rf $stale
+                    done
+                  '''
+                  stash name: "hadoop-${HADOOP3_VERSION}", includes: 
"hadoop-${HADOOP3_VERSION}-bin.tar.gz"
                 }
               }
             }
@@ -170,159 +175,157 @@ pipeline {
         BASEDIR = "${env.WORKSPACE}/component"
         BRANCH = "${env.BRANCH_NAME}"
       }
-      steps {
-        parallel {
-          stage('hadoop 2 integration test') {
-            agent {
-              node {
-                label 'hbase'
-              }
-            }
-            environment {
-              OUTPUT_DIR = "output-integration-hadoop-${env.HADOOP2_VERSION}"
+      parallel {
+        stage('hadoop 2 integration test') {
+          agent {
+            node {
+              label 'hbase'
             }
-            steps {
-              sh '''#!/bin/bash -e
-                echo "Setting up directories"
-                rm -rf "${OUTPUT_DIR}"
-                echo "(x) {color:red}-1 client integration test{color}\n-- 
Something went wrong with this stage, [check relevant console 
output|${BUILD_URL}/console]." >${OUTPUT_DIR}/commentfile
-                rm -rf "hbase-install"
-                rm -rf "hbase-client"
-                rm -rf "hbase-hadoop3-install"
-                rm -rf "hbase-hadoop3-client"
-                rm -rf hadoop-2*.tar.gz
-              '''
-              unstash 'hadoop-2'
-              unstash 'hbase-install'
-              sh '''#!/bin/bash -xe
-                if [[ "${BRANCH_NAME}" == *"branch-2"* ]]; then
-                  echo "Attempting to run an instance on top of Hadoop 2."
-                  hadoop_artifact=$(ls -1 "${WORKSPACE}"/hadoop-2*.tar.gz | 
head -n 1)
-                  tar --strip-components=1 -xzf "${hadoop_artifact}" -C 
"hadoop-2"
-                  install_artifact=$(ls -1 
"${WORKSPACE}"/unpacked_src_tarball/hbase-assembly/target/hbase-*-bin.tar.gz | 
grep -v client-bin | grep -v hadoop3)
-                  tar --strip-component=1 -xzf "${install_artifact}" -C 
"hbase-install"
-                  client_artifact=$(ls -1 
"${WORKSPACE}"/unpacked_src_tarball/hbase-assembly/target/hbase-*-client-bin.tar.gz
 | grep -v hadoop3)
-                  tar --strip-component=1 -xzf "${client_artifact}" -C 
"hbase-client"
-                  docker build -t hbase-integration-test -f 
"${BASEDIR}/dev-support/docker/Dockerfile" .
-                  docker run --rm -v "${WORKSPACE}":/hbase -v 
/etc/passwd:/etc/passwd:ro -v /etc/group:/etc/group:ro \
-                    -u `id -u`:`id -g` -e JAVA_HOME="/usr/lib/jvm/java-8" 
--workdir=/hbase hbase-integration-test \
-                    
component/dev-support/hbase_nightly_pseudo-distributed-test.sh \
-                    --single-process \
-                    --working-dir ${OUTPUT_DIR}/hadoop-2 \
-                    --hbase-client-install "hbase-client" \
-                    hbase-install \
-                    hadoop-2/bin/hadoop \
-                    hadoop-2/share/hadoop/yarn/timelineservice \
-                    
hadoop-2/share/hadoop/yarn/test/hadoop-yarn-server-tests-*-tests.jar \
-                    
hadoop-2/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-*-tests.jar \
-                    hadoop-2/bin/mapred \
-                    >${OUTPUT_DIR}/hadoop-2.log 2>&1
-                  if [ $? -ne 0 ]; then
-                    echo "(x) {color:red}-1 client integration 
test{color}\n--Failed when running client tests on top of Hadoop 2. [see log 
for details|${BUILD_URL}/artifact/${OUTPUT_DIR}/hadoop-2.log]. (note that this 
means we didn't run on Hadoop 3)" >${OUTPUT_DIR}/commentfile
-                    exit 2
-                  fi
-                  echo "(/) {color:green}+1 client integration test for HBase 
2 {color}" >${OUTPUT_DIR}/commentfile
-                else
-                  echo "Skipping to run against Hadoop 2 for branch 
${BRANCH_NAME}"
+          }
+          environment {
+            OUTPUT_DIR = "output-integration-hadoop-${env.HADOOP2_VERSION}"
+          }
+          steps {
+            sh '''#!/bin/bash -e
+              echo "Setting up directories"
+              rm -rf "${OUTPUT_DIR}"
+              echo "(x) {color:red}-1 client integration test{color}\n-- 
Something went wrong with this stage, [check relevant console 
output|${BUILD_URL}/console]." >${OUTPUT_DIR}/commentfile
+              rm -rf "hbase-install"
+              rm -rf "hbase-client"
+              rm -rf "hbase-hadoop3-install"
+              rm -rf "hbase-hadoop3-client"
+              rm -rf hadoop-2*.tar.gz
+            '''
+            unstash 'hadoop-2'
+            unstash 'hbase-install'
+            sh '''#!/bin/bash -xe
+              if [[ "${BRANCH_NAME}" == *"branch-2"* ]]; then
+                echo "Attempting to run an instance on top of Hadoop 2."
+                hadoop_artifact=$(ls -1 "${WORKSPACE}"/hadoop-2*.tar.gz | head 
-n 1)
+                tar --strip-components=1 -xzf "${hadoop_artifact}" -C 
"hadoop-2"
+                install_artifact=$(ls -1 
"${WORKSPACE}"/unpacked_src_tarball/hbase-assembly/target/hbase-*-bin.tar.gz | 
grep -v client-bin | grep -v hadoop3)
+                tar --strip-component=1 -xzf "${install_artifact}" -C 
"hbase-install"
+                client_artifact=$(ls -1 
"${WORKSPACE}"/unpacked_src_tarball/hbase-assembly/target/hbase-*-client-bin.tar.gz
 | grep -v hadoop3)
+                tar --strip-component=1 -xzf "${client_artifact}" -C 
"hbase-client"
+                docker build -t hbase-integration-test -f 
"${BASEDIR}/dev-support/docker/Dockerfile" .
+                docker run --rm -v "${WORKSPACE}":/hbase -v 
/etc/passwd:/etc/passwd:ro -v /etc/group:/etc/group:ro \
+                  -u `id -u`:`id -g` -e JAVA_HOME="/usr/lib/jvm/java-8" 
--workdir=/hbase hbase-integration-test \
+                  
component/dev-support/hbase_nightly_pseudo-distributed-test.sh \
+                  --single-process \
+                  --working-dir ${OUTPUT_DIR}/hadoop-2 \
+                  --hbase-client-install "hbase-client" \
+                  hbase-install \
+                  hadoop-2/bin/hadoop \
+                  hadoop-2/share/hadoop/yarn/timelineservice \
+                  
hadoop-2/share/hadoop/yarn/test/hadoop-yarn-server-tests-*-tests.jar \
+                  
hadoop-2/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-*-tests.jar \
+                  hadoop-2/bin/mapred \
+                  >${OUTPUT_DIR}/hadoop-2.log 2>&1
+                if [ $? -ne 0 ]; then
+                  echo "(x) {color:red}-1 client integration 
test{color}\n--Failed when running client tests on top of Hadoop 2. [see log 
for details|${BUILD_URL}/artifact/${OUTPUT_DIR}/hadoop-2.log]. (note that this 
means we didn't run on Hadoop 3)" >${OUTPUT_DIR}/commentfile
+                  exit 2
                 fi
-              '''
-            }
+                echo "(/) {color:green}+1 client integration test for HBase 2 
{color}" >${OUTPUT_DIR}/commentfile
+              else
+                echo "Skipping to run against Hadoop 2 for branch 
${BRANCH_NAME}"
+              fi
+            '''
           }
+        }
 
-          stage('hadoop 3 integration tests') {
-            matrix {
-              axes {
-                axis {
-                  name 'HADOOP_VERSION'
-                  values getHadoopVersions(env.HADOOP3_VERSIONS)
-                }
+        stage('hadoop 3 integration tests') {
+          matrix {
+            axes {
+              axis {
+                name 'HADOOP_VERSION'
+                values getHadoopVersions(env.HADOOP3_VERSIONS)
               }
-              agent {
-                node {
-                  label 'hbase'
-                }
+            }
+            agent {
+              node {
+                label 'hbase'
               }
-              stages {
-                stage('hadoop 3 integration test') {
-                  environment {
-                    OUTPUT_DIR = 
"output-integration-hadoop-${env.HADOOP_VERSION}"
-                  }
-                  steps {
-                    dir('component') {
-                      checkout scm
-                    }
-                    sh '''#!/bin/bash -e
-                      echo "Setting up directories"
-                      rm -rf "${OUTPUT_DIR}"
-                      echo "(x) {color:red}-1 client integration 
test{color}\n-- Something went wrong with this stage, [check relevant console 
output|${BUILD_URL}/console]." >${OUTPUT_DIR}/commentfile
-                      rm -rf "hbase-install"
-                      rm -rf "hbase-client"
-                      rm -rf "hbase-hadoop3-install"
-                      rm -rf "hbase-hadoop3-client"
-                      rm -rf hadoop-3*.tar.gz
-                    '''
-                    unstash "hadoop-${HADOOP_VERSION}"
-                    unstash 'hbase-install'
-                    sh '''#!/bin/bash -e
-                      echo "Attempting to use run an instance on top of Hadoop 
${HADOOP_VERSION}."
-                      rm -rf hadoop-3/*
-                      rm -rf "${OUTPUT_DIR}/non-shaded" && mkdir 
"${OUTPUT_DIR}/non-shaded"
-                      rm -rf "${OUTPUT_DIR}/shaded" && mkdir 
"${OUTPUT_DIR}/shaded"
-                      artifact=$(ls -1 
"${WORKSPACE}"/hadoop-${HADOOP_VERSION}-bin.tar.gz | head -n 1)
-                      tar --strip-components=1 -xzf "${artifact}" -C "hadoop-3"
-                      ${BASEDIR}/dev-support/patch-hadoop3.sh hadoop-3
-                      hbase_install_dir="hbase-install"
-                      hbase_client_dir="hbase-client"
-                      if [ -d "hbase-hadoop3-install" ]; then
-                        echo "run hadoop3 client integration test against 
hbase hadoop3 binaries"
-                        hbase_install_dir="hbase-hadoop3-install"
-                        hbase_client_dir="hbase-hadoop3-client"
-                      fi
-                      docker build -t hbase-integration-test -f 
"${BASEDIR}/dev-support/docker/Dockerfile" .
-                      docker run --rm -v "${WORKSPACE}":/hbase -v 
/etc/passwd:/etc/passwd:ro -v /etc/group:/etc/group:ro \
-                        -u `id -u`:`id -g` -e JAVA_HOME="/usr/lib/jvm/java-17" 
\
-                        -e HADOOP_OPTS="--add-opens 
java.base/java.lang=ALL-UNNAMED" \
-                        --workdir=/hbase hbase-integration-test \
-                        
component/dev-support/hbase_nightly_pseudo-distributed-test.sh \
-                        --single-process \
-                        --working-dir ${OUTPUT_DIR}/non-shaded \
-                        --hbase-client-install ${hbase_client_dir} \
-                        ${hbase_install_dir} \
-                        hadoop-3/bin/hadoop \
-                        hadoop-3/share/hadoop/yarn/timelineservice \
-                        
hadoop-3/share/hadoop/yarn/test/hadoop-yarn-server-tests-*-tests.jar \
-                        
hadoop-3/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-*-tests.jar \
-                        hadoop-3/bin/mapred \
-                        >${OUTPUT_DIR}/hadoop.log 2>&1
-                      if [ $? -ne 0 ]; then
-                        echo "(x) {color:red}-1 client integration 
test{color}\n--Failed when running client tests on top of Hadoop 
${HADOOP_VERSION}. [see log for 
details|${BUILD_URL}/artifact/${OUTPUT_DIR}/hadoop.log]. (note that this means 
we didn't check the Hadoop ${HADOOP_VERSION} shaded client)" > 
${OUTPUT_DIR}/commentfile
-                        exit 2
-                      fi
-                      echo "(/) {color:green}+1 client integration test for 
${HADOOP_VERSION} {color}" >> ${OUTPUT_DIR}/commentfile
-                      echo "Attempting to run an instance on top of Hadoop 
${HADOOP_VERSION}, relying on the Hadoop client artifacts for the example 
client program."
-                      docker run --rm -v "${WORKSPACE}":/hbase -v 
/etc/passwd:/etc/passwd:ro -v /etc/group:/etc/group:ro \
-                        -u `id -u`:`id -g` -e JAVA_HOME="/usr/lib/jvm/java-17" 
\
-                        -e HADOOP_OPTS="--add-opens 
java.base/java.lang=ALL-UNNAMED" \
-                        --workdir=/hbase hbase-integration-test \
-                        
component/dev-support/hbase_nightly_pseudo-distributed-test.sh \
-                        --single-process \
-                        --hadoop-client-classpath 
hadoop-3/share/hadoop/client/hadoop-client-api-*.jar:hadoop-3/share/hadoop/client/hadoop-client-runtime-*.jar
 \
-                        --working-dir ${OUTPUT_DIR}/shade \
-                        --hbase-client-install ${hbase_client_dir} \
-                        ${hbase_install_dir} \
-                        hadoop-3/bin/hadoop \
-                        hadoop-3/share/hadoop/yarn/timelineservice \
-                        
hadoop-3/share/hadoop/yarn/test/hadoop-yarn-server-tests-*-tests.jar \
-                        
hadoop-3/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-*-tests.jar \
-                        hadoop-3/bin/mapred \
-                        >${OUTPUT_DIR}/hadoop-shaded.log 2>&1
-                      if [ $? -ne 0 ]; then
-                        echo "(x) {color:red}-1 client integration 
test{color}\n--Failed when running client tests on top of Hadoop 
${HADOOP_VERSION} using Hadoop's shaded client. [see log for 
details|${BUILD_URL}/artifact/${OUTPUT_DIR}/hadoop-shaded.log]." >> 
${OUTPUT_DIR}/commentfile
-                        exit 2
-                      fi
-                      echo "(/) {color:green}+1 client integration test for 
${HADOOP_VERSION} with shaded hadoop client {color}" >> 
${OUTPUT_DIR}/commentfile
-                    '''
+            }
+            stages {
+              stage('hadoop 3 integration test') {
+                environment {
+                  OUTPUT_DIR = 
"output-integration-hadoop-${env.HADOOP_VERSION}"
+                }
+                steps {
+                  dir('component') {
+                    checkout scm
                   }
+                  sh '''#!/bin/bash -e
+                    echo "Setting up directories"
+                    rm -rf "${OUTPUT_DIR}"
+                    echo "(x) {color:red}-1 client integration test{color}\n-- 
Something went wrong with this stage, [check relevant console 
output|${BUILD_URL}/console]." >${OUTPUT_DIR}/commentfile
+                    rm -rf "hbase-install"
+                    rm -rf "hbase-client"
+                    rm -rf "hbase-hadoop3-install"
+                    rm -rf "hbase-hadoop3-client"
+                    rm -rf hadoop-3*.tar.gz
+                  '''
+                  unstash "hadoop-${HADOOP_VERSION}"
+                  unstash 'hbase-install'
+                  sh '''#!/bin/bash -e
+                    echo "Attempting to use run an instance on top of Hadoop 
${HADOOP_VERSION}."
+                    rm -rf hadoop-3/*
+                    rm -rf "${OUTPUT_DIR}/non-shaded" && mkdir 
"${OUTPUT_DIR}/non-shaded"
+                    rm -rf "${OUTPUT_DIR}/shaded" && mkdir 
"${OUTPUT_DIR}/shaded"
+                    artifact=$(ls -1 
"${WORKSPACE}"/hadoop-${HADOOP_VERSION}-bin.tar.gz | head -n 1)
+                    tar --strip-components=1 -xzf "${artifact}" -C "hadoop-3"
+                    ${BASEDIR}/dev-support/patch-hadoop3.sh hadoop-3
+                    hbase_install_dir="hbase-install"
+                    hbase_client_dir="hbase-client"
+                    if [ -d "hbase-hadoop3-install" ]; then
+                      echo "run hadoop3 client integration test against hbase 
hadoop3 binaries"
+                      hbase_install_dir="hbase-hadoop3-install"
+                      hbase_client_dir="hbase-hadoop3-client"
+                    fi
+                    docker build -t hbase-integration-test -f 
"${BASEDIR}/dev-support/docker/Dockerfile" .
+                    docker run --rm -v "${WORKSPACE}":/hbase -v 
/etc/passwd:/etc/passwd:ro -v /etc/group:/etc/group:ro \
+                      -u `id -u`:`id -g` -e JAVA_HOME="/usr/lib/jvm/java-17" \
+                      -e HADOOP_OPTS="--add-opens 
java.base/java.lang=ALL-UNNAMED" \
+                      --workdir=/hbase hbase-integration-test \
+                      
component/dev-support/hbase_nightly_pseudo-distributed-test.sh \
+                      --single-process \
+                      --working-dir ${OUTPUT_DIR}/non-shaded \
+                      --hbase-client-install ${hbase_client_dir} \
+                      ${hbase_install_dir} \
+                      hadoop-3/bin/hadoop \
+                      hadoop-3/share/hadoop/yarn/timelineservice \
+                      
hadoop-3/share/hadoop/yarn/test/hadoop-yarn-server-tests-*-tests.jar \
+                      
hadoop-3/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-*-tests.jar \
+                      hadoop-3/bin/mapred \
+                      >${OUTPUT_DIR}/hadoop.log 2>&1
+                    if [ $? -ne 0 ]; then
+                      echo "(x) {color:red}-1 client integration 
test{color}\n--Failed when running client tests on top of Hadoop 
${HADOOP_VERSION}. [see log for 
details|${BUILD_URL}/artifact/${OUTPUT_DIR}/hadoop.log]. (note that this means 
we didn't check the Hadoop ${HADOOP_VERSION} shaded client)" > 
${OUTPUT_DIR}/commentfile
+                      exit 2
+                    fi
+                    echo "(/) {color:green}+1 client integration test for 
${HADOOP_VERSION} {color}" >> ${OUTPUT_DIR}/commentfile
+                    echo "Attempting to run an instance on top of Hadoop 
${HADOOP_VERSION}, relying on the Hadoop client artifacts for the example 
client program."
+                    docker run --rm -v "${WORKSPACE}":/hbase -v 
/etc/passwd:/etc/passwd:ro -v /etc/group:/etc/group:ro \
+                      -u `id -u`:`id -g` -e JAVA_HOME="/usr/lib/jvm/java-17" \
+                      -e HADOOP_OPTS="--add-opens 
java.base/java.lang=ALL-UNNAMED" \
+                      --workdir=/hbase hbase-integration-test \
+                      
component/dev-support/hbase_nightly_pseudo-distributed-test.sh \
+                      --single-process \
+                      --hadoop-client-classpath 
hadoop-3/share/hadoop/client/hadoop-client-api-*.jar:hadoop-3/share/hadoop/client/hadoop-client-runtime-*.jar
 \
+                      --working-dir ${OUTPUT_DIR}/shade \
+                      --hbase-client-install ${hbase_client_dir} \
+                      ${hbase_install_dir} \
+                      hadoop-3/bin/hadoop \
+                      hadoop-3/share/hadoop/yarn/timelineservice \
+                      
hadoop-3/share/hadoop/yarn/test/hadoop-yarn-server-tests-*-tests.jar \
+                      
hadoop-3/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-*-tests.jar \
+                      hadoop-3/bin/mapred \
+                      >${OUTPUT_DIR}/hadoop-shaded.log 2>&1
+                    if [ $? -ne 0 ]; then
+                      echo "(x) {color:red}-1 client integration 
test{color}\n--Failed when running client tests on top of Hadoop 
${HADOOP_VERSION} using Hadoop's shaded client. [see log for 
details|${BUILD_URL}/artifact/${OUTPUT_DIR}/hadoop-shaded.log]." >> 
${OUTPUT_DIR}/commentfile
+                      exit 2
+                    fi
+                    echo "(/) {color:green}+1 client integration test for 
${HADOOP_VERSION} with shaded hadoop client {color}" >> 
${OUTPUT_DIR}/commentfile
+                  '''
                 }
               }
             }

Reply via email to