This is an automated email from the ASF dual-hosted git repository.

jackylee pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-gluten.git


The following commit(s) were added to refs/heads/main by this push:
     new 9c15d2e1e [GLUTEN-5335][VL] Use common name for both celeborn and 
uniffle (#5385)
9c15d2e1e is described below

commit 9c15d2e1e2f5cac4cf65c169c099f81c1ba1bc2e
Author: summaryzb <[email protected]>
AuthorDate: Tue Apr 16 10:56:38 2024 +0800

    [GLUTEN-5335][VL] Use common name for both celeborn and uniffle (#5385)
---
 .github/workflows/build_bundle_package.yml         |  6 +++---
 .github/workflows/velox_docker.yml                 | 24 +++++++++++-----------
 cpp/core/CMakeLists.txt                            |  2 +-
 cpp/core/jni/JniCommon.h                           | 23 ++++++++++-----------
 cpp/core/jni/JniWrapper.cc                         | 15 +++++++-------
 cpp/core/shuffle/Options.h                         |  2 +-
 ...ornPartitionWriter.cc => RssPartitionWriter.cc} | 21 +++++++++----------
 ...ebornPartitionWriter.h => RssPartitionWriter.h} | 10 ++++-----
 cpp/velox/benchmarks/GenericBenchmark.cc           |  8 ++++----
 cpp/velox/tests/VeloxShuffleWriterTest.cc          |  4 ++--
 cpp/velox/utils/tests/VeloxShuffleWriterTestBase.h |  4 ++--
 dev/buildbundle-veloxbe.sh                         |  8 ++++----
 dev/package-vcpkg.sh                               |  8 ++++----
 dev/package.sh                                     |  8 ++++----
 docs/developers/HowTo.md                           |  4 ++--
 docs/developers/MicroBenchmarks.md                 |  4 ++--
 docs/developers/NewToGluten.md                     |  4 ++--
 docs/get-started/Velox.md                          |  6 +++---
 ep/build-clickhouse/src/package.sh                 |  4 ++--
 .../gluten/uniffle/UniffleShuffleManager.java      |  2 +-
 package/pom.xml                                    |  4 ++--
 pom.xml                                            |  4 ++--
 .../scala/org/apache/gluten/GlutenConfig.scala     |  2 +-
 tools/gluten-it/package/pom.xml                    |  4 ++--
 tools/gluten-te/centos/dockerfile-build            |  3 ++-
 tools/gluten-te/ubuntu/dockerfile-build            |  3 ++-
 .../examples/buildhere-veloxbe-dev/scripts/all.sh  |  3 ++-
 .../ubuntu/examples/buildhere-veloxbe/run.sh       |  3 ++-
 28 files changed, 97 insertions(+), 96 deletions(-)

diff --git a/.github/workflows/build_bundle_package.yml 
b/.github/workflows/build_bundle_package.yml
index 4fc45f098..01ddd6f43 100644
--- a/.github/workflows/build_bundle_package.yml
+++ b/.github/workflows/build_bundle_package.yml
@@ -79,7 +79,7 @@ jobs:
       - name: Build for Spark ${{ github.event.inputs.spark }}
         run: |
           cd $GITHUB_WORKSPACE/ && \
-          mvn clean install -P${{ github.event.inputs.spark }} 
-Dhadoop.version=${{ github.event.inputs.hadoop }} -Pbackends-velox -Prss 
-DskipTests -Dmaven.source.skip
+          mvn clean install -P${{ github.event.inputs.spark }} 
-Dhadoop.version=${{ github.event.inputs.hadoop }} -Pbackends-velox -Pceleborn 
-DskipTests -Dmaven.source.skip
       - name: Upload bundle package
         uses: actions/upload-artifact@v2
         with:
@@ -110,7 +110,7 @@ jobs:
           cd $GITHUB_WORKSPACE/ && \
           export MAVEN_HOME=/usr/lib/maven && \
           export PATH=${PATH}:${MAVEN_HOME}/bin && \
-          mvn clean install -P${{ github.event.inputs.spark }} 
-Dhadoop.version=${{ github.event.inputs.hadoop }} -Pbackends-velox -Prss 
-DskipTests -Dmaven.source.skip
+          mvn clean install -P${{ github.event.inputs.spark }} 
-Dhadoop.version=${{ github.event.inputs.hadoop }} -Pbackends-velox -Pceleborn 
-DskipTests -Dmaven.source.skip
       - name: Upload bundle package
         uses: actions/upload-artifact@v2
         with:
@@ -145,7 +145,7 @@ jobs:
           cd $GITHUB_WORKSPACE/ && \
           export MAVEN_HOME=/usr/lib/maven && \
           export PATH=${PATH}:${MAVEN_HOME}/bin && \
-          mvn clean install -P${{ github.event.inputs.spark }} 
-Dhadoop.version=${{ github.event.inputs.hadoop }} -Pbackends-velox -Prss 
-DskipTests -Dmaven.source.skip
+          mvn clean install -P${{ github.event.inputs.spark }} 
-Dhadoop.version=${{ github.event.inputs.hadoop }} -Pbackends-velox -Pceleborn 
-DskipTests -Dmaven.source.skip
       - name: Upload bundle package
         uses: actions/upload-artifact@v2
         with:
diff --git a/.github/workflows/velox_docker.yml 
b/.github/workflows/velox_docker.yml
index 8515d79d2..271daf679 100644
--- a/.github/workflows/velox_docker.yml
+++ b/.github/workflows/velox_docker.yml
@@ -369,7 +369,7 @@ jobs:
           cd $GITHUB_WORKSPACE/ && \
           export MAVEN_HOME=/usr/lib/maven && \
           export PATH=${PATH}:${MAVEN_HOME}/bin && \
-          mvn clean install -P${{ matrix.spark }} -Pbackends-velox 
-Prss-uniffle -DskipTests
+          mvn clean install -P${{ matrix.spark }} -Pbackends-velox -Puniffle 
-DskipTests
       - name: TPC-H SF1.0 && TPC-DS SF1.0 Parquet local spark3.2 with uniffle 
0.8.0
         run: |
           export MAVEN_HOME=/usr/lib/maven && \
@@ -395,7 +395,7 @@ jobs:
           bash -c "echo -e 'rss.coordinator.shuffle.nodes.max 
1\nrss.rpc.server.port 19999' > ./conf/coordinator.conf" && \
           bash -c "echo -e 'rss.server.app.expired.withoutHeartbeat 
7200000\nrss.server.heartbeat.delay 3000\nrss.rpc.server.port 
19997\nrss.jetty.http.port 19996\nrss.server.netty.port 
19995\nrss.storage.basePath /opt/uniffle/shuffle_data\nrss.storage.type 
MEMORY_LOCALFILE\nrss.coordinator.quorum 
localhost:19999\nrss.server.flush.thread.alive 
10\nrss.server.single.buffer.flush.threshold 64m' > ./conf/server.conf" && \
           bash ./bin/start-coordinator.sh && bash ./bin/start-shuffle-server.sh
-          cd $GITHUB_WORKSPACE/tools/gluten-it && mvn clean install 
-Pspark-3.2,rss-uniffle && \
+          cd $GITHUB_WORKSPACE/tools/gluten-it && mvn clean install 
-Pspark-3.2 -Puniffle && \
           GLUTEN_IT_JVM_ARGS=-Xmx5G sbin/gluten-it.sh queries-compare \
             --local --preset=velox-with-uniffle --benchmark-type=h 
--error-on-memleak --off-heap-size=10g -s=1.0 --threads=16 --iterations=1
 
@@ -421,7 +421,7 @@ jobs:
       - name: Build for Spark ${{ matrix.spark }}
         run: |
           cd $GITHUB_WORKSPACE/
-          mvn clean install -P${{ matrix.spark }} -Pbackends-velox,rss 
-DskipTests
+          mvn clean install -P${{ matrix.spark }} -Pbackends-velox -Pceleborn 
-DskipTests
       - name: TPC-H SF1.0 && TPC-DS SF1.0 Parquet local spark3.2 with ${{ 
matrix.celeborn }}
         run: |
           export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
@@ -437,7 +437,7 @@ jobs:
           bash -c "echo -e 
'CELEBORN_MASTER_MEMORY=4g\nCELEBORN_WORKER_MEMORY=4g\nCELEBORN_WORKER_OFFHEAP_MEMORY=8g'
 > ./conf/celeborn-env.sh" && \
           bash -c "echo -e 'celeborn.worker.commitFiles.threads 
128\nceleborn.worker.sortPartition.threads 64' > ./conf/celeborn-defaults.conf" 
&& \
           bash ./sbin/start-master.sh && bash ./sbin/start-worker.sh && \
-          cd $GITHUB_WORKSPACE/tools/gluten-it && mvn clean install 
-Pspark-3.2,rss ${EXTRA_PROFILE} && \
+          cd $GITHUB_WORKSPACE/tools/gluten-it && mvn clean install 
-Pspark-3.2 -Pceleborn ${EXTRA_PROFILE} && \
           GLUTEN_IT_JVM_ARGS=-Xmx5G sbin/gluten-it.sh queries-compare \
             --local --preset=velox-with-celeborn --benchmark-type=h 
--error-on-memleak --off-heap-size=10g -s=1.0 --threads=8 --iterations=1 && \
           GLUTEN_IT_JVM_ARGS=-Xmx5G sbin/gluten-it.sh queries-compare \
@@ -488,7 +488,7 @@ jobs:
           export SPARK_SCALA_VERSION=2.12
           export MAVEN_HOME=/usr/lib/maven 
           export PATH=${PATH}:${MAVEN_HOME}/bin
-          mvn -ntp clean install -Pspark-3.2 -Pspark-ut -Pbackends-velox -Prss 
-Piceberg -Pdelta 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark32/spark_home/" 
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
 && \
+          mvn -ntp clean install -Pspark-3.2 -Pspark-ut -Pbackends-velox 
-Pceleborn -Piceberg -Pdelta 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark32/spark_home/" 
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
 && \
           mvn -ntp test -Pspark-3.2 -Pbackends-velox -Piceberg -Pdelta 
-DtagsToExclude=None -DtagsToInclude=org.apache.gluten.tags.UDFTest
       - name: Upload golden files
         if: failure()
@@ -536,7 +536,7 @@ jobs:
           cd $GITHUB_WORKSPACE/ && \
           export MAVEN_HOME=/usr/lib/maven 
           export PATH=${PATH}:${MAVEN_HOME}/bin
-          mvn -ntp clean install -Pspark-3.2 -Pspark-ut -Pbackends-velox -Prss 
-Piceberg -Pdelta 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark32/spark_home/" 
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
+          mvn -ntp clean install -Pspark-3.2 -Pspark-ut -Pbackends-velox 
-Pceleborn -Piceberg -Pdelta 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark32/spark_home/" 
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
 
   run-spark-test-spark33:
     runs-on: ubuntu-20.04
@@ -579,7 +579,7 @@ jobs:
           export SPARK_SCALA_VERSION=2.12 && \
           export MAVEN_HOME=/usr/lib/maven 
           export PATH=${PATH}:${MAVEN_HOME}/bin
-          mvn -ntp clean install -Pspark-3.3 -Pbackends-velox -Prss -Piceberg 
-Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark33/spark_home/" 
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
 && \
+          mvn -ntp clean install -Pspark-3.3 -Pbackends-velox -Pceleborn 
-Piceberg -Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark33/spark_home/" 
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
 && \
           mvn -ntp test -Pspark-3.3 -Pbackends-velox -Piceberg -Pdelta 
-DtagsToExclude=None -DtagsToInclude=org.apache.gluten.tags.UDFTest
       - name: Upload golden files
         if: failure()
@@ -623,7 +623,7 @@ jobs:
           cd $GITHUB_WORKSPACE/ && \
           export MAVEN_HOME=/usr/lib/maven 
           export PATH=${PATH}:${MAVEN_HOME}/bin
-          mvn -ntp clean install -Pspark-3.3 -Pbackends-velox -Prss -Piceberg 
-Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark33/spark_home/" 
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
+          mvn -ntp clean install -Pspark-3.3 -Pbackends-velox -Pceleborn 
-Piceberg -Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark33/spark_home/" 
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
 
   run-spark-test-spark34:
     runs-on: ubuntu-20.04
@@ -666,7 +666,7 @@ jobs:
           export SPARK_SCALA_VERSION=2.12 && \
           export MAVEN_HOME=/usr/lib/maven 
           export PATH=${PATH}:${MAVEN_HOME}/bin
-          mvn -ntp clean install -Pspark-3.4 -Pbackends-velox -Prss -Piceberg 
-Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark34/spark_home/" 
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
 && \
+          mvn -ntp clean install -Pspark-3.4 -Pbackends-velox -Pceleborn 
-Piceberg -Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark34/spark_home/" 
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
 && \
           mvn -ntp test -Pspark-3.4 -Pbackends-velox -Piceberg -Pdelta 
-DtagsToExclude=None -DtagsToInclude=org.apache.gluten.tags.UDFTest
       - name: Upload golden files
         if: failure()
@@ -710,7 +710,7 @@ jobs:
           cd $GITHUB_WORKSPACE/
           export MAVEN_HOME=/usr/lib/maven
           export PATH=${PATH}:${MAVEN_HOME}/bin
-          mvn -ntp clean install -Pspark-3.4 -Pbackends-velox -Prss -Piceberg 
-Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark34/spark_home/" 
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
+          mvn -ntp clean install -Pspark-3.4 -Pbackends-velox -Pceleborn 
-Piceberg -Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark34/spark_home/" 
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
 
   run-spark-test-spark35:
     runs-on: ubuntu-20.04
@@ -753,7 +753,7 @@ jobs:
           export SPARK_SCALA_VERSION=2.12 && \
           export MAVEN_HOME=/usr/lib/maven 
           export PATH=${PATH}:${MAVEN_HOME}/bin
-          mvn -ntp clean install -Pspark-3.5 -Pbackends-velox -Prss -Piceberg 
-Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark35/spark_home/" 
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
 && \
+          mvn -ntp clean install -Pspark-3.5 -Pbackends-velox -Pceleborn 
-Piceberg -Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark35/spark_home/" 
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
 && \
           mvn -ntp test -Pspark-3.5 -Pbackends-velox -Piceberg -Pdelta 
-DtagsToExclude=None -DtagsToInclude=org.apache.gluten.tags.UDFTest
       - name: Upload golden files
         if: failure()
@@ -802,4 +802,4 @@ jobs:
           cd $GITHUB_WORKSPACE/
           export MAVEN_HOME=/usr/lib/maven
           export PATH=${PATH}:${MAVEN_HOME}/bin
-          mvn -ntp clean install -Pspark-3.5 -Pbackends-velox -Prss -Piceberg 
-Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark35/spark_home/" 
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
\ No newline at end of file
+          mvn -ntp clean install -Pspark-3.5 -Pbackends-velox -Pceleborn 
-Piceberg -Pdelta -Pspark-ut 
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark35/spark_home/" 
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
\ No newline at end of file
diff --git a/cpp/core/CMakeLists.txt b/cpp/core/CMakeLists.txt
index 7e3e1ca00..c369a5bc3 100644
--- a/cpp/core/CMakeLists.txt
+++ b/cpp/core/CMakeLists.txt
@@ -194,7 +194,7 @@ set(SPARK_COLUMNAR_PLUGIN_SRCS
         shuffle/Partitioning.cc
         shuffle/Payload.cc
         shuffle/rss/RemotePartitionWriter.cc
-        shuffle/rss/CelebornPartitionWriter.cc
+        shuffle/rss/RssPartitionWriter.cc
         shuffle/RoundRobinPartitioner.cc
         shuffle/ShuffleMemoryPool.cc
         shuffle/ShuffleReader.cc
diff --git a/cpp/core/jni/JniCommon.h b/cpp/core/jni/JniCommon.h
index ad21d442c..bda5fc1df 100644
--- a/cpp/core/jni/JniCommon.h
+++ b/cpp/core/jni/JniCommon.h
@@ -411,28 +411,28 @@ class BacktraceAllocationListener final : public 
gluten::AllocationListener {
   std::atomic_int64_t backtraceBytes_{1L << 30};
 };
 
-class CelebornClient : public RssClient {
+class JavaRssClient : public RssClient {
  public:
-  CelebornClient(JavaVM* vm, jobject javaCelebornShuffleWriter, jmethodID 
javaCelebornPushPartitionDataMethod)
-      : vm_(vm), 
javaCelebornPushPartitionData_(javaCelebornPushPartitionDataMethod) {
+  JavaRssClient(JavaVM* vm, jobject javaRssShuffleWriter, jmethodID 
javaPushPartitionDataMethod)
+      : vm_(vm), javaPushPartitionData_(javaPushPartitionDataMethod) {
     JNIEnv* env;
     if (vm_->GetEnv(reinterpret_cast<void**>(&env), jniVersion) != JNI_OK) {
       throw gluten::GlutenException("JNIEnv was not attached to current 
thread");
     }
 
-    javaCelebornShuffleWriter_ = env->NewGlobalRef(javaCelebornShuffleWriter);
+    javaRssShuffleWriter_ = env->NewGlobalRef(javaRssShuffleWriter);
     array_ = env->NewByteArray(1024 * 1024);
     array_ = static_cast<jbyteArray>(env->NewGlobalRef(array_));
   }
 
-  ~CelebornClient() {
+  ~JavaRssClient() {
     JNIEnv* env;
     if (vm_->GetEnv(reinterpret_cast<void**>(&env), jniVersion) != JNI_OK) {
-      LOG(WARNING) << "CelebornClient#~CelebornClient(): "
+      LOG(WARNING) << "JavaRssClient#~JavaRssClient(): "
                    << "JNIEnv was not attached to current thread";
       return;
     }
-    env->DeleteGlobalRef(javaCelebornShuffleWriter_);
+    env->DeleteGlobalRef(javaRssShuffleWriter_);
     jbyte* byteArray = env->GetByteArrayElements(array_, NULL);
     env->ReleaseByteArrayElements(array_, byteArray, JNI_ABORT);
     env->DeleteGlobalRef(array_);
@@ -452,17 +452,16 @@ class CelebornClient : public RssClient {
       array_ = static_cast<jbyteArray>(env->NewGlobalRef(array_));
     }
     env->SetByteArrayRegion(array_, 0, size, reinterpret_cast<jbyte*>(bytes));
-    jint celebornBytesSize =
-        env->CallIntMethod(javaCelebornShuffleWriter_, 
javaCelebornPushPartitionData_, partitionId, array_, size);
+    jint javaBytesSize = env->CallIntMethod(javaRssShuffleWriter_, 
javaPushPartitionData_, partitionId, array_, size);
     checkException(env);
-    return static_cast<int32_t>(celebornBytesSize);
+    return static_cast<int32_t>(javaBytesSize);
   }
 
   void stop() override {}
 
  private:
   JavaVM* vm_;
-  jobject javaCelebornShuffleWriter_;
-  jmethodID javaCelebornPushPartitionData_;
+  jobject javaRssShuffleWriter_;
+  jmethodID javaPushPartitionData_;
   jbyteArray array_;
 };
diff --git a/cpp/core/jni/JniWrapper.cc b/cpp/core/jni/JniWrapper.cc
index 6f5b2e332..748eb5831 100644
--- a/cpp/core/jni/JniWrapper.cc
+++ b/cpp/core/jni/JniWrapper.cc
@@ -34,7 +34,7 @@
 #include "shuffle/ShuffleReader.h"
 #include "shuffle/ShuffleWriter.h"
 #include "shuffle/Utils.h"
-#include "shuffle/rss/CelebornPartitionWriter.h"
+#include "shuffle/rss/RssPartitionWriter.h"
 #include "utils/ArrowStatus.h"
 #include "utils/StringUtil.h"
 
@@ -908,9 +908,9 @@ JNIEXPORT jlong JNICALL 
Java_org_apache_gluten_vectorized_ShuffleWriterJniWrappe
     if (env->GetJavaVM(&vm) != JNI_OK) {
       throw gluten::GlutenException("Unable to get JavaVM instance");
     }
-    std::shared_ptr<CelebornClient> celebornClient =
-        std::make_shared<CelebornClient>(vm, partitionPusher, 
celebornPushPartitionDataMethod);
-    partitionWriter = std::make_unique<CelebornPartitionWriter>(
+    std::shared_ptr<JavaRssClient> celebornClient =
+        std::make_shared<JavaRssClient>(vm, partitionPusher, 
celebornPushPartitionDataMethod);
+    partitionWriter = std::make_unique<RssPartitionWriter>(
         numPartitions,
         std::move(partitionWriterOptions),
         memoryManager->getArrowMemoryPool(),
@@ -924,10 +924,9 @@ JNIEXPORT jlong JNICALL 
Java_org_apache_gluten_vectorized_ShuffleWriterJniWrappe
     if (env->GetJavaVM(&vm) != JNI_OK) {
       throw gluten::GlutenException("Unable to get JavaVM instance");
     }
-    // rename CelebornClient RssClient
-    std::shared_ptr<CelebornClient> uniffleClient =
-        std::make_shared<CelebornClient>(vm, partitionPusher, 
unifflePushPartitionDataMethod);
-    partitionWriter = std::make_unique<CelebornPartitionWriter>(
+    std::shared_ptr<JavaRssClient> uniffleClient =
+        std::make_shared<JavaRssClient>(vm, partitionPusher, 
unifflePushPartitionDataMethod);
+    partitionWriter = std::make_unique<RssPartitionWriter>(
         numPartitions,
         std::move(partitionWriterOptions),
         memoryManager->getArrowMemoryPool(),
diff --git a/cpp/core/shuffle/Options.h b/cpp/core/shuffle/Options.h
index 6a793daa3..d8fe1c802 100644
--- a/cpp/core/shuffle/Options.h
+++ b/cpp/core/shuffle/Options.h
@@ -33,7 +33,7 @@ static constexpr double kDefaultBufferReallocThreshold = 0.25;
 static constexpr double kDefaultMergeBufferThreshold = 0.25;
 static constexpr bool kEnableBufferedWrite = true;
 
-enum PartitionWriterType { kLocal, kCeleborn, kUniffle };
+enum PartitionWriterType { kLocal, kRss };
 
 struct ShuffleReaderOptions {
   arrow::Compression::type compressionType = 
arrow::Compression::type::LZ4_FRAME;
diff --git a/cpp/core/shuffle/rss/CelebornPartitionWriter.cc 
b/cpp/core/shuffle/rss/RssPartitionWriter.cc
similarity index 78%
rename from cpp/core/shuffle/rss/CelebornPartitionWriter.cc
rename to cpp/core/shuffle/rss/RssPartitionWriter.cc
index b1ac1a1e3..72bcf1d3a 100644
--- a/cpp/core/shuffle/rss/CelebornPartitionWriter.cc
+++ b/cpp/core/shuffle/rss/RssPartitionWriter.cc
@@ -19,20 +19,20 @@
 
 #include "shuffle/Payload.h"
 #include "shuffle/Utils.h"
-#include "shuffle/rss/CelebornPartitionWriter.h"
+#include "shuffle/rss/RssPartitionWriter.h"
 #include "utils/Timer.h"
 
 namespace gluten {
 
-void CelebornPartitionWriter::init() {
+void RssPartitionWriter::init() {
   bytesEvicted_.resize(numPartitions_, 0);
   rawPartitionLengths_.resize(numPartitions_, 0);
 }
 
-arrow::Status CelebornPartitionWriter::stop(ShuffleWriterMetrics* metrics) {
+arrow::Status RssPartitionWriter::stop(ShuffleWriterMetrics* metrics) {
   // Push data and collect metrics.
   auto totalBytesEvicted = std::accumulate(bytesEvicted_.begin(), 
bytesEvicted_.end(), 0LL);
-  celebornClient_->stop();
+  rssClient_->stop();
   // Populate metrics.
   metrics->totalCompressTime += compressTime_;
   metrics->totalEvictTime += spillTime_;
@@ -44,12 +44,12 @@ arrow::Status 
CelebornPartitionWriter::stop(ShuffleWriterMetrics* metrics) {
   return arrow::Status::OK();
 }
 
-arrow::Status CelebornPartitionWriter::reclaimFixedSize(int64_t size, int64_t* 
actual) {
+arrow::Status RssPartitionWriter::reclaimFixedSize(int64_t size, int64_t* 
actual) {
   *actual = 0;
   return arrow::Status::OK();
 }
 
-arrow::Status CelebornPartitionWriter::evict(
+arrow::Status RssPartitionWriter::evict(
     uint32_t partitionId,
     std::unique_ptr<InMemoryPayload> inMemoryPayload,
     Evict::type evictType,
@@ -64,14 +64,13 @@ arrow::Status CelebornPartitionWriter::evict(
   ARROW_ASSIGN_OR_RAISE(
       auto payload, inMemoryPayload->toBlockPayload(payloadType, 
payloadPool_.get(), codec_ ? codec_.get() : nullptr));
   // Copy payload to arrow buffered os.
-  ARROW_ASSIGN_OR_RAISE(
-      auto celebornBufferOs, 
arrow::io::BufferOutputStream::Create(options_.pushBufferMaxSize, pool_));
-  RETURN_NOT_OK(payload->serialize(celebornBufferOs.get()));
+  ARROW_ASSIGN_OR_RAISE(auto rssBufferOs, 
arrow::io::BufferOutputStream::Create(options_.pushBufferMaxSize, pool_));
+  RETURN_NOT_OK(payload->serialize(rssBufferOs.get()));
   payload = nullptr; // Invalidate payload immediately.
 
   // Push.
-  ARROW_ASSIGN_OR_RAISE(auto buffer, celebornBufferOs->Finish());
-  bytesEvicted_[partitionId] += celebornClient_->pushPartitionData(
+  ARROW_ASSIGN_OR_RAISE(auto buffer, rssBufferOs->Finish());
+  bytesEvicted_[partitionId] += rssClient_->pushPartitionData(
       partitionId, 
reinterpret_cast<char*>(const_cast<uint8_t*>(buffer->data())), buffer->size());
   return arrow::Status::OK();
 }
diff --git a/cpp/core/shuffle/rss/CelebornPartitionWriter.h 
b/cpp/core/shuffle/rss/RssPartitionWriter.h
similarity index 87%
rename from cpp/core/shuffle/rss/CelebornPartitionWriter.h
rename to cpp/core/shuffle/rss/RssPartitionWriter.h
index 612979c02..ef43017fc 100644
--- a/cpp/core/shuffle/rss/CelebornPartitionWriter.h
+++ b/cpp/core/shuffle/rss/RssPartitionWriter.h
@@ -26,14 +26,14 @@
 
 namespace gluten {
 
-class CelebornPartitionWriter final : public RemotePartitionWriter {
+class RssPartitionWriter final : public RemotePartitionWriter {
  public:
-  CelebornPartitionWriter(
+  RssPartitionWriter(
       uint32_t numPartitions,
       PartitionWriterOptions options,
       arrow::MemoryPool* pool,
-      std::shared_ptr<RssClient> celebornClient)
-      : RemotePartitionWriter(numPartitions, std::move(options), pool), 
celebornClient_(celebornClient) {
+      std::shared_ptr<RssClient> rssClient)
+      : RemotePartitionWriter(numPartitions, std::move(options), pool), 
rssClient_(rssClient) {
     init();
   }
 
@@ -51,7 +51,7 @@ class CelebornPartitionWriter final : public 
RemotePartitionWriter {
  private:
   void init();
 
-  std::shared_ptr<RssClient> celebornClient_;
+  std::shared_ptr<RssClient> rssClient_;
 
   std::vector<int64_t> bytesEvicted_;
   std::vector<int64_t> rawPartitionLengths_;
diff --git a/cpp/velox/benchmarks/GenericBenchmark.cc 
b/cpp/velox/benchmarks/GenericBenchmark.cc
index f2e275911..ef88da2c3 100644
--- a/cpp/velox/benchmarks/GenericBenchmark.cc
+++ b/cpp/velox/benchmarks/GenericBenchmark.cc
@@ -31,7 +31,7 @@
 #include "config/GlutenConfig.h"
 #include "shuffle/LocalPartitionWriter.h"
 #include "shuffle/VeloxShuffleWriter.h"
-#include "shuffle/rss/CelebornPartitionWriter.h"
+#include "shuffle/rss/RssPartitionWriter.h"
 #include "utils/StringUtil.h"
 #include "utils/VeloxArrowUtils.h"
 #include "utils/exception.h"
@@ -46,7 +46,7 @@ DEFINE_bool(print_result, true, "Print result for execution");
 DEFINE_string(save_output, "", "Path to parquet file for saving the task 
output iterator");
 DEFINE_bool(with_shuffle, false, "Add shuffle split at end.");
 DEFINE_string(partitioning, "rr", "Short partitioning name. Valid options are 
rr, hash, range, single");
-DEFINE_bool(celeborn, false, "Mocking celeborn shuffle.");
+DEFINE_bool(rss, false, "Mocking rss.");
 DEFINE_bool(zstd, false, "Use ZSTD as shuffle compression codec");
 DEFINE_bool(qat_gzip, false, "Use QAT GZIP as shuffle compression codec");
 DEFINE_bool(qat_zstd, false, "Use QAT ZSTD as shuffle compression codec");
@@ -90,9 +90,9 @@ std::shared_ptr<VeloxShuffleWriter> createShuffleWriter(
   }
 
   std::unique_ptr<PartitionWriter> partitionWriter;
-  if (FLAGS_celeborn) {
+  if (FLAGS_rss) {
     auto rssClient = std::make_unique<LocalRssClient>(dataFile);
-    partitionWriter = std::make_unique<CelebornPartitionWriter>(
+    partitionWriter = std::make_unique<RssPartitionWriter>(
         FLAGS_shuffle_partitions,
         std::move(partitionWriterOptions),
         memoryManager->getArrowMemoryPool(),
diff --git a/cpp/velox/tests/VeloxShuffleWriterTest.cc 
b/cpp/velox/tests/VeloxShuffleWriterTest.cc
index 3f370207d..2a8f12afb 100644
--- a/cpp/velox/tests/VeloxShuffleWriterTest.cc
+++ b/cpp/velox/tests/VeloxShuffleWriterTest.cc
@@ -20,7 +20,7 @@
 
 #include "shuffle/LocalPartitionWriter.h"
 #include "shuffle/VeloxShuffleWriter.h"
-#include "shuffle/rss/CelebornPartitionWriter.h"
+#include "shuffle/rss/RssPartitionWriter.h"
 #include "utils/TestUtils.h"
 #include "utils/VeloxArrowUtils.h"
 #include "utils/tests/MemoryPoolUtils.h"
@@ -73,7 +73,7 @@ std::vector<ShuffleTestParams> createShuffleTestParams() {
         params.push_back(
             ShuffleTestParams{PartitionWriterType::kLocal, compression, 
compressionThreshold, mergeBufferSize});
       }
-      params.push_back(ShuffleTestParams{PartitionWriterType::kCeleborn, 
compression, compressionThreshold, 0});
+      params.push_back(ShuffleTestParams{PartitionWriterType::kRss, 
compression, compressionThreshold, 0});
     }
   }
 
diff --git a/cpp/velox/utils/tests/VeloxShuffleWriterTestBase.h 
b/cpp/velox/utils/tests/VeloxShuffleWriterTestBase.h
index bef131981..972c0cb25 100644
--- a/cpp/velox/utils/tests/VeloxShuffleWriterTestBase.h
+++ b/cpp/velox/utils/tests/VeloxShuffleWriterTestBase.h
@@ -53,9 +53,9 @@ std::unique_ptr<PartitionWriter> createPartitionWriter(
     const std::vector<std::string>& localDirs,
     const PartitionWriterOptions& options,
     arrow::MemoryPool* pool) {
-  if (partitionWriterType == PartitionWriterType::kCeleborn) {
+  if (partitionWriterType == PartitionWriterType::kRss) {
     auto rssClient = std::make_unique<LocalRssClient>(dataFile);
-    return std::make_unique<CelebornPartitionWriter>(numPartitions, options, 
pool, std::move(rssClient));
+    return std::make_unique<RssPartitionWriter>(numPartitions, options, pool, 
std::move(rssClient));
   }
   return std::make_unique<LocalPartitionWriter>(numPartitions, options, pool, 
dataFile, localDirs);
 }
diff --git a/dev/buildbundle-veloxbe.sh b/dev/buildbundle-veloxbe.sh
index 1c77b52a6..10c8e61e2 100755
--- a/dev/buildbundle-veloxbe.sh
+++ b/dev/buildbundle-veloxbe.sh
@@ -4,7 +4,7 @@ BASEDIR=$(dirname $0)
 source "$BASEDIR/builddeps-veloxbe.sh"
 
 cd $GLUTEN_DIR
-mvn clean package -Pbackends-velox -Prss -Pspark-3.2 -DskipTests
-mvn clean package -Pbackends-velox -Prss -Pspark-3.3 -DskipTests
-mvn clean package -Pbackends-velox -Prss -Pspark-3.4 -DskipTests
-mvn clean package -Pbackends-velox -Prss -Pspark-3.5 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.2 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.3 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.4 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.5 -DskipTests
diff --git a/dev/package-vcpkg.sh b/dev/package-vcpkg.sh
index 5c3e95759..3f14f1c90 100755
--- a/dev/package-vcpkg.sh
+++ b/dev/package-vcpkg.sh
@@ -8,7 +8,7 @@ GLUTEN_DIR="$CURRENT_DIR/.."
 cd "$GLUTEN_DIR"
 source ./dev/vcpkg/env.sh
 ./dev/buildbundle-veloxbe.sh --build_tests=ON --build_benchmarks=ON 
--enable_s3=ON  --enable_hdfs=ON
-mvn clean package -Pbackends-velox -Prss -Pspark-3.2 -DskipTests
-mvn clean package -Pbackends-velox -Prss -Pspark-3.3 -DskipTests
-mvn clean package -Pbackends-velox -Prss -Pspark-3.4 -DskipTests
-mvn clean package -Pbackends-velox -Prss -Pspark-3.5 -DskipTests
\ No newline at end of file
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.2 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.3 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.4 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.5 -DskipTests
\ No newline at end of file
diff --git a/dev/package.sh b/dev/package.sh
index 2e33fcd85..1b9ca85e9 100755
--- a/dev/package.sh
+++ b/dev/package.sh
@@ -11,10 +11,10 @@ ARCH=`uname -m`
 
 # compile gluten jar
 $GLUTEN_DIR/dev/builddeps-veloxbe.sh --build_tests=ON --build_benchmarks=ON 
--enable_s3=ON --enable_hdfs=ON
-mvn clean package -Pbackends-velox -Prss -Pspark-3.2 -DskipTests
-mvn clean package -Pbackends-velox -Prss -Pspark-3.3 -DskipTests
-mvn clean package -Pbackends-velox -Prss -Pspark-3.4 -DskipTests
-mvn clean package -Pbackends-velox -Prss -Pspark-3.5 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.2 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.3 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.4 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.5 -DskipTests
 
 mkdir -p $THIRDPARTY_LIB
 function process_setup_ubuntu_2004 {
diff --git a/docs/developers/HowTo.md b/docs/developers/HowTo.md
index ff036a060..a13bf02eb 100644
--- a/docs/developers/HowTo.md
+++ b/docs/developers/HowTo.md
@@ -54,8 +54,8 @@ ${GLUTEN_HOME}/dev/builddeps-veloxbe.sh --build_tests=ON 
--build_benchmarks=ON -
 
 ```
 cd ${GLUTEN_HOME}
-mvn clean package -Pspark-3.2 -Pbackends-velox -Prss
-mvn test -Pspark-3.2 -Pbackends-velox -Prss -pl backends-velox \
+mvn clean package -Pspark-3.2 -Pbackends-velox -Pceleborn -Puniffle
+mvn test -Pspark-3.2 -Pbackends-velox -Pceleborn -pl backends-velox \
 -am -DtagsToInclude="org.apache.gluten.tags.GenerateExample" \
 -Dtest=none -DfailIfNoTests=false \
 -Dexec.skip
diff --git a/docs/developers/MicroBenchmarks.md 
b/docs/developers/MicroBenchmarks.md
index f6c95d593..dc0c3b2a0 100644
--- a/docs/developers/MicroBenchmarks.md
+++ b/docs/developers/MicroBenchmarks.md
@@ -36,8 +36,8 @@ generate example input files:
 cd /path/to/gluten/
 ./dev/buildbundle-veloxbe.sh --build_tests=ON --build_benchmarks=ON
 
-# Run test to generate input data files. If you are using spark 3.3, replace 
-Pspark-3.2 with -Pspark-3.3
-mvn test -Pspark-3.2 -Pbackends-velox -Prss -pl backends-velox -am \
+# Run test to generate input data files. If you are using spark 3.3, replace 
-Pspark-3.2 with -Pspark-3.3, If you are using uniffle, replace -Pceleborn with 
-Puniffle
+mvn test -Pspark-3.2 -Pbackends-velox -Pcelenborn -pl backends-velox -am \
 -DtagsToInclude="org.apache.gluten.tags.GenerateExample" -Dtest=none 
-DfailIfNoTests=false -Dexec.skip
 ```
 
diff --git a/docs/developers/NewToGluten.md b/docs/developers/NewToGluten.md
index abaabc699..b3f05a64b 100644
--- a/docs/developers/NewToGluten.md
+++ b/docs/developers/NewToGluten.md
@@ -120,8 +120,8 @@ To generate a fix for Java/Scala code style, you can run 
one or more of the belo
 
 For Velox backend:
 ```
-mvn spotless:apply -Pbackends-velox -Prss -Pspark-3.2 -Pspark-ut -DskipTests
-mvn spotless:apply -Pbackends-velox -Prss -Pspark-3.3 -Pspark-ut -DskipTests
+mvn spotless:apply -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.2 
-Pspark-ut -DskipTests
+mvn spotless:apply -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.3 
-Pspark-ut -DskipTests
 ```
 For Clickhouse backend:
 ```
diff --git a/docs/get-started/Velox.md b/docs/get-started/Velox.md
index 81693a134..426b56c6b 100644
--- a/docs/get-started/Velox.md
+++ b/docs/get-started/Velox.md
@@ -101,9 +101,9 @@ cd /path/to/gluten/cpp
 ## compile Gluten java module and create package jar
 cd /path/to/gluten
 # For spark3.2.x
-mvn clean package -Pbackends-velox -Prss -Pspark-3.2 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.2 -DskipTests
 # For spark3.3.x
-mvn clean package -Pbackends-velox -Prss -Pspark-3.3 -DskipTests
+mvn clean package -Pbackends-velox -Pceleborn -Puniffle -Pspark-3.3 -DskipTests
 ```
 
 notes:The compilation of `Velox` using the script of `build_velox.sh` may fail 
caused by `oom`, you can prevent this failure by using the user command of 
`export NUM_THREADS=4` before executing the above scripts.
@@ -214,7 +214,7 @@ First refer to this URL(https://github.com/apache/celeborn) 
to setup a celeborn
 When compiling the Gluten Java module, it's required to enable `rss` profile, 
as follows:
 
 ```
-mvn clean package -Pbackends-velox -Pspark-3.3 -Prss -DskipTests
+mvn clean package -Pbackends-velox -Pspark-3.3 -Pceleborn -DskipTests
 ```
 
 Then add the Gluten and Spark Celeborn Client packages to your Spark 
application's classpath(usually add them into `$SPARK_HOME/jars`).
diff --git a/ep/build-clickhouse/src/package.sh 
b/ep/build-clickhouse/src/package.sh
index a1cf9733c..3741e9ae0 100755
--- a/ep/build-clickhouse/src/package.sh
+++ b/ep/build-clickhouse/src/package.sh
@@ -79,7 +79,7 @@ cp "${GLUTEN_SOURCE}"/LICENSE 
"${GLUTEN_SOURCE}"/dist/"${PACKAGE_NAME}"
 cp "${GLUTEN_SOURCE}"/README.md "${GLUTEN_SOURCE}"/dist/"${PACKAGE_NAME}"
 
 # build gluten with spark32
-mvn clean install -Pbackends-clickhouse -Pspark-3.2 -Prss -DskipTests 
-Dcheckstyle.skip
+mvn clean install -Pbackends-clickhouse -Pspark-3.2 -Pceleborn -Puniffle 
-DskipTests -Dcheckstyle.skip
 cp 
"${GLUTEN_SOURCE}"/backends-clickhouse/target/gluten-*-spark-3.2-jar-with-dependencies.jar
 "${PACKAGE_DIR_PATH}"/jars/spark32/gluten.jar
 cp 
"${GLUTEN_SOURCE}"/gluten-celeborn/clickhouse/target/gluten-celeborn-clickhouse-${PROJECT_VERSION}-jar-with-dependencies.jar
 "${PACKAGE_DIR_PATH}"/jars/spark32
 delta_version_32=$(mvn -q -Dexec.executable="echo" 
-Dexec.args='${delta.version}' -Pspark-3.2 --non-recursive exec:exec)
@@ -87,7 +87,7 @@ wget 
https://repo1.maven.org/maven2/io/delta/delta-core_2.12/${delta_version_32}
 wget 
https://repo1.maven.org/maven2/io/delta/delta-storage/${delta_version_32}/delta-storage-${delta_version_32}.jar
 -P "${PACKAGE_DIR_PATH}"/jars/spark32
 
 # build gluten with spark33
-mvn clean install -Pbackends-clickhouse -Pspark-3.3 -Prss -DskipTests 
-Dcheckstyle.skip
+mvn clean install -Pbackends-clickhouse -Pspark-3.3 -Pceleborn -Puniffle 
-DskipTests -Dcheckstyle.skip
 cp 
"${GLUTEN_SOURCE}"/backends-clickhouse/target/gluten-*-spark-3.3-jar-with-dependencies.jar
 "${PACKAGE_DIR_PATH}"/jars/spark33/gluten.jar
 cp 
"${GLUTEN_SOURCE}"/gluten-celeborn/clickhouse/target/gluten-celeborn-clickhouse-${PROJECT_VERSION}-jar-with-dependencies.jar
 "${PACKAGE_DIR_PATH}"/jars/spark33
 delta_version_33=$(mvn -q -Dexec.executable="echo" 
-Dexec.args='${delta.version}' -Pspark-3.3 --non-recursive exec:exec)
diff --git 
a/gluten-uniffle/velox/src/main/java/org/apache/spark/shuffle/gluten/uniffle/UniffleShuffleManager.java
 
b/gluten-uniffle/velox/src/main/java/org/apache/spark/shuffle/gluten/uniffle/UniffleShuffleManager.java
index 9ae62f8b8..8e0f4f085 100644
--- 
a/gluten-uniffle/velox/src/main/java/org/apache/spark/shuffle/gluten/uniffle/UniffleShuffleManager.java
+++ 
b/gluten-uniffle/velox/src/main/java/org/apache/spark/shuffle/gluten/uniffle/UniffleShuffleManager.java
@@ -42,7 +42,7 @@ public class UniffleShuffleManager extends RssShuffleManager {
 
   public UniffleShuffleManager(SparkConf conf, boolean isDriver) {
     super(conf, isDriver);
-    conf.set(RssSparkConfig.SPARK_RSS_CONFIG_PREFIX + 
RssSparkConfig.RSS_ROW_BASED, "false");
+    conf.set(RssSparkConfig.SPARK_RSS_CONFIG_PREFIX + 
RssSparkConfig.RSS_ROW_BASED.key(), "false");
   }
 
   @Override
diff --git a/package/pom.xml b/package/pom.xml
index 913621ee1..52fac53e5 100644
--- a/package/pom.xml
+++ b/package/pom.xml
@@ -71,7 +71,7 @@
       </dependencies>
     </profile>
     <profile>
-      <id>rss</id>
+      <id>celeborn</id>
       <dependencies>
         <dependency>
           <groupId>org.apache.gluten</groupId>
@@ -81,7 +81,7 @@
       </dependencies>
     </profile>
     <profile>
-      <id>rss-uniffle</id>
+      <id>uniffle</id>
       <dependencies>
         <dependency>
           <groupId>org.apache.gluten</groupId>
diff --git a/pom.xml b/pom.xml
index 0707d348d..6ba36121d 100644
--- a/pom.xml
+++ b/pom.xml
@@ -235,7 +235,7 @@
       </properties>
     </profile>
     <profile>
-      <id>rss</id>
+      <id>celeborn</id>
       <activation>
         <activeByDefault>false</activeByDefault>
       </activation>
@@ -244,7 +244,7 @@
       </modules>
     </profile>
     <profile>
-      <id>rss-uniffle</id>
+      <id>uniffle</id>
       <activation>
         <activeByDefault>false</activeByDefault>
       </activation>
diff --git a/shims/common/src/main/scala/org/apache/gluten/GlutenConfig.scala 
b/shims/common/src/main/scala/org/apache/gluten/GlutenConfig.scala
index cfaf54cdd..60ff95a7e 100644
--- a/shims/common/src/main/scala/org/apache/gluten/GlutenConfig.scala
+++ b/shims/common/src/main/scala/org/apache/gluten/GlutenConfig.scala
@@ -124,7 +124,7 @@ class GlutenConfig(conf: SQLConf) extends Logging {
   def isUseUniffleShuffleManager: Boolean =
     conf
       .getConfString("spark.shuffle.manager", "sort")
-      .contains("RssShuffleManager")
+      .contains("UniffleShuffleManager")
 
   def enableColumnarShuffle: Boolean = conf.getConf(COLUMNAR_SHUFFLE_ENABLED)
 
diff --git a/tools/gluten-it/package/pom.xml b/tools/gluten-it/package/pom.xml
index 3beef2df7..1f86ee723 100644
--- a/tools/gluten-it/package/pom.xml
+++ b/tools/gluten-it/package/pom.xml
@@ -95,7 +95,7 @@
 
   <profiles>
     <profile>
-      <id>rss</id>
+      <id>celeborn</id>
       <activation>
         <activeByDefault>false</activeByDefault>
       </activation>
@@ -119,7 +119,7 @@
       </dependencies>
     </profile>
     <profile>
-      <id>rss-uniffle</id>
+      <id>uniffle</id>
       <activation>
         <activeByDefault>false</activeByDefault>
       </activation>
diff --git a/tools/gluten-te/centos/dockerfile-build 
b/tools/gluten-te/centos/dockerfile-build
index 02bc8b6a9..fa496613b 100644
--- a/tools/gluten-te/centos/dockerfile-build
+++ b/tools/gluten-te/centos/dockerfile-build
@@ -65,7 +65,8 @@ RUN if [ "$BUILD_BACKEND_TYPE" == "velox" ]; \
                            --build_type=$GLUTEN_BUILD_TYPE 
--enable_ep_cache=ON"; \
       EXTRA_MAVEN_OPTIONS="-Pspark-3.2 \
                            -Pbackends-velox \
-                           -Prss \
+                           -Pceleborn \
+                           -Puniffle \
                            -Piceberg \
                            -Pdelta \
                            -DskipTests \
diff --git a/tools/gluten-te/ubuntu/dockerfile-build 
b/tools/gluten-te/ubuntu/dockerfile-build
index 606e14715..59736c986 100644
--- a/tools/gluten-te/ubuntu/dockerfile-build
+++ b/tools/gluten-te/ubuntu/dockerfile-build
@@ -65,7 +65,8 @@ RUN if [ "$BUILD_BACKEND_TYPE" == "velox" ]; \
                            --build_type=$GLUTEN_BUILD_TYPE 
--enable_ep_cache=ON"; \
       EXTRA_MAVEN_OPTIONS="-Pspark-3.2 \
                            -Pbackends-velox \
-                           -Prss \
+                           -Pceleborn \
+                           -Puniffle \
                            -Piceberg \
                            -Pdelta \
                            -DskipTests \
diff --git 
a/tools/gluten-te/ubuntu/examples/buildhere-veloxbe-dev/scripts/all.sh 
b/tools/gluten-te/ubuntu/examples/buildhere-veloxbe-dev/scripts/all.sh
index 0e1634b3f..f795b5ba9 100755
--- a/tools/gluten-te/ubuntu/examples/buildhere-veloxbe-dev/scripts/all.sh
+++ b/tools/gluten-te/ubuntu/examples/buildhere-veloxbe-dev/scripts/all.sh
@@ -19,7 +19,8 @@ set -ex
 # Build Gluten
 EXTRA_MAVEN_OPTIONS="-Pspark-3.2 \
                      -Pbackends-velox \
-                     -Prss \
+                     -Pceleborn \
+                     -Puniffle \
                      -DskipTests \
                      -Dscalastyle.skip=true \
                      -Dcheckstyle.skip=true"
diff --git a/tools/gluten-te/ubuntu/examples/buildhere-veloxbe/run.sh 
b/tools/gluten-te/ubuntu/examples/buildhere-veloxbe/run.sh
index ec4a1d628..a7ba6c526 100755
--- a/tools/gluten-te/ubuntu/examples/buildhere-veloxbe/run.sh
+++ b/tools/gluten-te/ubuntu/examples/buildhere-veloxbe/run.sh
@@ -20,7 +20,8 @@ BASEDIR=$(dirname $0)
 
 EXTRA_MAVEN_OPTIONS="-Pspark-3.2 \
                      -Pbackends-velox \
-                     -Prss \
+                     -Pceleborn \
+                     -Puniffle \
                      -DskipTests \
                      -Dscalastyle.skip=true \
                      -Dcheckstyle.skip=true"


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]


Reply via email to