This is an automated email from the ASF dual-hosted git repository.
yao pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-kyuubi.git
The following commit(s) were added to refs/heads/master by this push:
new 74abe85 [KYUUBI #1950] Remove ambiguous SPARK_HADOOP_VERSION
74abe85 is described below
commit 74abe8510c3a01283e52d8ba59a7fe350e42434e
Author: Kent Yao <[email protected]>
AuthorDate: Tue Feb 22 11:26:38 2022 +0800
[KYUUBI #1950] Remove ambiguous SPARK_HADOOP_VERSION
<!--
Thanks for sending a pull request!
Here are some tips for you:
1. If this is your first time, please read our contributor guidelines:
https://kyuubi.readthedocs.io/en/latest/community/contributions.html
2. If the PR is related to an issue in
https://github.com/apache/incubator-kyuubi/issues, add '[KYUUBI #XXXX]' in your
PR title, e.g., '[KYUUBI #XXXX] Your PR title ...'.
3. If the PR is unfinished, add '[WIP]' in your PR title, e.g.,
'[WIP][KYUUBI #XXXX] Your PR title ...'.
-->
### _Why are the changes needed?_
<!--
Please clarify why the changes are needed. For instance,
1. If you add a feature, you can talk about the use case of it.
2. If you fix a bug, you can clarify why it is a bug.
-->
The original idea of SPARK_HADOOP_VERSION is used to concat spark release
names only, now we need to remove it as
- SPARK_HADOOP_VERSION is misunderstood by developers and misused somewhere
like the one of kyuubi compiled
- multi-engine support now
- the release names of spark(or something else) are very easy to get
through code with different environments, prod/test/dev
- A `mvn` job is bundled with `bin/load-kyuubi-env.sh` which is truly
worrisome
- SPARK_HADOOP_VERSION on spark side hass broken already for spark 3.2
which actually bundled with hadoop 3.3, see
https://github.com/apache/spark-website/pull/361#discussion_r730716668
### _How was this patch tested?_
- [ ] Add some test cases that check the changes thoroughly including
negative and positive cases if possible
- [ ] Add screenshots for manual tests if appropriate
- [x] [Run
test](https://kyuubi.apache.org/docs/latest/develop_tools/testing.html#running-tests)
locally before make a pull request
Closes #1950 from yaooqinn/hadoop.
Closes #1950
b47be7c6 [Kent Yao] Remove ambiguous SPARK_HADOOP_VERSION
3b33ee56 [Kent Yao] Remove ambiguous SPARK_HADOOP_VERSION
Authored-by: Kent Yao <[email protected]>
Signed-off-by: Kent Yao <[email protected]>
---
bin/load-kyuubi-env.sh | 30 ++++--------------------------
build/dist | 16 +++++-----------
pom.xml | 17 +----------------
3 files changed, 10 insertions(+), 53 deletions(-)
diff --git a/bin/load-kyuubi-env.sh b/bin/load-kyuubi-env.sh
index 050fac3..7d65cca 100755
--- a/bin/load-kyuubi-env.sh
+++ b/bin/load-kyuubi-env.sh
@@ -71,32 +71,11 @@ fi
export KYUUBI_SCALA_VERSION="${KYUUBI_SCALA_VERSION:-"2.12"}"
if [[ -f ${KYUUBI_HOME}/RELEASE ]]; then
- KYUUBI_VERSION="$(grep "Kyuubi " "$KYUUBI_HOME/RELEASE" | awk -F ' ' '{print
$2}')"
- FLINK_VERSION_BUILD="$(grep "Flink " "$KYUUBI_HOME/RELEASE" | awk -F ' '
'{print $2}')"
- SPARK_VERSION_BUILD="$(grep "Spark " "$KYUUBI_HOME/RELEASE" | awk -F ' '
'{print $2}' | grep -v 'Hadoop')"
- HADOOP_VERSION_BUILD="$(grep "Spark Hadoop " "$KYUUBI_HOME/RELEASE" | awk -F
' ' '{print $3}')"
- FLINK_BUILTIN="${KYUUBI_HOME}/externals/flink-$FLINK_VERSION_BUILD"
-
SPARK_BUILTIN="${KYUUBI_HOME}/externals/spark-$SPARK_VERSION_BUILD-bin-hadoop${HADOOP_VERSION_BUILD:0:3}"
+ FLINK_BUILTIN="$(find "$KYUUBI_HOME/externals" -name 'flink-*' -type d |
head -n 1)"
+ SPARK_BUILTIN="$(find "$KYUUBI_HOME/externals" -name 'spark-*' -type d |
head -n 1)"
else
- MVN="${MVN:-"${KYUUBI_HOME}/build/mvn"}"
- KYUUBI_VERSION=$("$MVN" help:evaluate -Dexpression=project.version
2>/dev/null\
- | grep -v "INFO"\
- | grep -v "WARNING"\
- | tail -n 1)
- FLINK_VERSION_BUILD=$("$MVN" help:evaluate -Dexpression=flink.version
2>/dev/null\
- | grep -v "INFO"\
- | grep -v "WARNING"\
- | tail -n 1)
- SPARK_VERSION_BUILD=$("$MVN" help:evaluate -Dexpression=spark.version
2>/dev/null\
- | grep -v "INFO"\
- | grep -v "WARNING"\
- | tail -n 1)
- HADOOP_VERSION_BUILD=$("$MVN" help:evaluate
-Dexpression=hadoop.binary.version 2>/dev/null\
- | grep -v "INFO"\
- | grep -v "WARNING"\
- | tail -n 1)
-
FLINK_BUILTIN="${KYUUBI_HOME}/externals/kyuubi-download/target/flink-$FLINK_VERSION_BUILD"
-
SPARK_BUILTIN="${KYUUBI_HOME}/externals/kyuubi-download/target/spark-$SPARK_VERSION_BUILD-bin-hadoop${HADOOP_VERSION_BUILD}"
+ FLINK_BUILTIN="$(find "$KYUUBI_HOME/externals/kyuubi-download/target" -name
'flink-*' -type d | head -n 1)"
+ SPARK_BUILTIN="$(find "$KYUUBI_HOME/externals/kyuubi-download/target" -name
'spark-*' -type d | head -n 1)"
fi
export FLINK_HOME="${FLINK_HOME:-"${FLINK_BUILTIN}"}"
@@ -109,7 +88,6 @@ if [ $silent -eq 0 ]; then
echo "JAVA_HOME: ${JAVA_HOME}"
echo "KYUUBI_HOME: ${KYUUBI_HOME}"
- echo "KYUUBI_VERSION: ${KYUUBI_VERSION}"
echo "KYUUBI_CONF_DIR: ${KYUUBI_CONF_DIR}"
echo "KYUUBI_LOG_DIR: ${KYUUBI_LOG_DIR}"
echo "KYUUBI_PID_DIR: ${KYUUBI_PID_DIR}"
diff --git a/build/dist b/build/dist
index f4d0d42..6874ecc 100755
--- a/build/dist
+++ b/build/dist
@@ -153,11 +153,6 @@ SPARK_VERSION=$("$MVN" help:evaluate
-Dexpression=spark.version $@ 2>/dev/null\
| grep -v "WARNING"\
| tail -n 1)
-SPARK_HADOOP_VERSION=$("$MVN" help:evaluate -Dexpression=hadoop.binary.version
$@ 2>/dev/null\
- | grep -v "INFO"\
- | grep -v "WARNING"\
- | tail -n 1)
-
HADOOP_VERSION=$("$MVN" help:evaluate -Dexpression=hadoop.version $@
2>/dev/null\
| grep -v "INFO"\
| grep -v "WARNING"\
@@ -175,7 +170,7 @@ if [[ "$NAME" == "none" ]]; then
if [[ "$SPARK_PROVIDED" == "true" ]]; then
SUFFIX=""
else
- SUFFIX="-spark-${SPARK_VERSION:0:3}-hadoop${SPARK_HADOOP_VERSION}"
+ SUFFIX="-spark-${SPARK_VERSION:0:3}"
fi
fi
@@ -219,7 +214,6 @@ echo "Java $JAVA_VERSION" >> "$DISTDIR/RELEASE"
echo "Scala $SCALA_VERSION" >> "$DISTDIR/RELEASE"
echo "Flink $FLINK_VERSION" >> "$DISTDIR/RELEASE"
echo "Spark $SPARK_VERSION" >> "$DISTDIR/RELEASE"
-echo "Spark Hadoop $SPARK_HADOOP_VERSION" >> "$DISTDIR/RELEASE"
echo "Kyuubi Hadoop $HADOOP_VERSION" >> "$DISTDIR/RELEASE"
echo "Hive $HIVE_VERSION" >> "$DISTDIR/RELEASE"
echo "Build flags: $@" >> "$DISTDIR/RELEASE"
@@ -275,14 +269,14 @@ done
if [[ "$FLINK_PROVIDED" != "true" ]]; then
# Copy flink binary dist
- cp -r "$KYUUBI_HOME/externals/kyuubi-download/target/flink-$FLINK_VERSION/" \
- "$DISTDIR/externals/flink-$FLINK_VERSION/"
+ FLINK_BUILTIN="$(find "$KYUUBI_HOME/externals/kyuubi-download/target" -name
'flink-*' -type d)"
+ cp -r "$FLINK_BUILTIN" "$DISTDIR/externals/"
fi
if [[ "$SPARK_PROVIDED" != "true" ]]; then
# Copy spark binary dist
- cp -r
"$KYUUBI_HOME/externals/kyuubi-download/target/spark-$SPARK_VERSION-bin-hadoop${SPARK_HADOOP_VERSION}$HIVE_VERSION_SUFFIX/"
\
-
"$DISTDIR/externals/spark-$SPARK_VERSION-bin-hadoop${SPARK_HADOOP_VERSION}$HIVE_VERSION_SUFFIX/"
+ SPARK_BUILTIN="$(find "$KYUUBI_HOME/externals/kyuubi-download/target" -name
'spark-*' -type d)"
+ cp -r "$SPARK_BUILTIN" "$DISTDIR/externals/"
fi
# Copy license files
diff --git a/pom.xml b/pom.xml
index adae5e3..ecbe31b 100644
--- a/pom.xml
+++ b/pom.xml
@@ -110,7 +110,6 @@
<guava.version>30.1-jre</guava.version>
<guava.failureaccess.version>1.0.1</guava.failureaccess.version>
<hadoop.version>3.3.1</hadoop.version>
- <hadoop.binary.version>3.2</hadoop.binary.version>
<hive.version>2.3.9</hive.version>
<hudi.version>0.10.0</hudi.version>
<iceberg.name>iceberg-spark-runtime-3.2_${scala.binary.version}</iceberg.name>
@@ -138,7 +137,7 @@
`delta.version`, `iceberg.name`
-->
<spark.version>3.2.1</spark.version>
-
<spark.archive.name>spark-${spark.version}-bin-hadoop${hadoop.binary.version}.tgz</spark.archive.name>
+
<spark.archive.name>spark-${spark.version}-bin-hadoop3.2.tgz</spark.archive.name>
<spark.archive.mirror>https://archive.apache.org/dist/spark/spark-${spark.version}</spark.archive.mirror>
<spark.archive.download.skip>false</spark.archive.download.skip>
<swagger.version>2.1.11</swagger.version>
@@ -1902,20 +1901,6 @@
</profile>
<profile>
- <id>spark-hadoop-2.7</id>
- <properties>
- <hadoop.binary.version>2.7</hadoop.binary.version>
- </properties>
- </profile>
-
- <profile>
- <id>spark-hadoop-3.2</id>
- <properties>
- <hadoop.binary.version>3.2</hadoop.binary.version>
- </properties>
- </profile>
-
- <profile>
<id>spark-provided</id>
<properties>
<spark.archive.download.skip>true</spark.archive.download.skip>