yihua commented on code in PR #11130:
URL: https://github.com/apache/hudi/pull/11130#discussion_r1587002563


##########
packaging/bundle-validation/base/Dockerfile:
##########
@@ -51,9 +52,16 @@ RUN wget 
https://archive.apache.org/dist/flink/flink-$FLINK_VERSION/flink-$FLINK
     && rm $WORKDIR/flink-$FLINK_VERSION-bin-scala_2.12.tgz
 ENV FLINK_HOME=$WORKDIR/flink-$FLINK_VERSION
 
-RUN wget 
https://archive.apache.org/dist/spark/spark-$SPARK_VERSION/spark-$SPARK_VERSION-bin-hadoop$SPARK_HADOOP_VERSION.tgz
 -P "$WORKDIR" \
-    && tar -xf 
$WORKDIR/spark-$SPARK_VERSION-bin-hadoop$SPARK_HADOOP_VERSION.tgz -C $WORKDIR/ \
-    && rm $WORKDIR/spark-$SPARK_VERSION-bin-hadoop$SPARK_HADOOP_VERSION.tgz
+RUN if [ "$SCALA_VERSION" = "2.13" ]; then \

Review Comment:
   That's the default value: `ARG SPARK_VERSION=3.1.3`.  When we build the 
docker image, we override the arguments as needed, e.g., see 
`packaging/bundle-validation/base/build_flink1180hive313spark350scala213.sh`:
   ```
   docker build \
    --build-arg HIVE_VERSION=3.1.3 \
    --build-arg FLINK_VERSION=1.18.0 \
    --build-arg SPARK_VERSION=3.5.0 \
    --build-arg SPARK_HADOOP_VERSION=3 \
    --build-arg HADOOP_VERSION=3.3.5 \
    --build-arg SCALA_VERSION=2.13 \
    -t hudi-ci-bundle-validation-base:flink1180hive313spark350scala213 .
   ```



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to