snuyanzin commented on code in PR #24285:
URL: https://github.com/apache/flink/pull/24285#discussion_r1534233653
##########
flink-end-to-end-tests/run-nightly-tests.sh:
##########
@@ -125,30 +125,28 @@ function run_group_1 {
# Docker / Container / Kubernetes tests
################################################################################
- if [[ ${PROFILE} != *"enable-adaptive-scheduler"* ]]; then
- run_test "Wordcount on Docker test (custom fs plugin)"
"$END_TO_END_DIR/test-scripts/test_docker_embedded_job.sh dummy-fs"
-
- run_test "Run Kubernetes test"
"$END_TO_END_DIR/test-scripts/test_kubernetes_embedded_job.sh"
- run_test "Run kubernetes session test (default input)"
"$END_TO_END_DIR/test-scripts/test_kubernetes_session.sh"
- run_test "Run kubernetes session test (custom fs plugin)"
"$END_TO_END_DIR/test-scripts/test_kubernetes_session.sh dummy-fs"
- run_test "Run kubernetes application test"
"$END_TO_END_DIR/test-scripts/test_kubernetes_application.sh"
- run_test "Run kubernetes application HA test"
"$END_TO_END_DIR/test-scripts/test_kubernetes_application_ha.sh"
- run_test "Run Kubernetes IT test"
"$END_TO_END_DIR/test-scripts/test_kubernetes_itcases.sh"
-
- run_test "Running Flink over NAT end-to-end test"
"$END_TO_END_DIR/test-scripts/test_nat.sh" "skip_check_exceptions"
-
- if [[ `uname -i` != 'aarch64' ]]; then
- # Skip PyFlink e2e test, because MiniConda and Pyarrow which
Pyflink depends doesn't support aarch64 currently.
- run_test "Run kubernetes pyflink application test"
"$END_TO_END_DIR/test-scripts/test_kubernetes_pyflink_application.sh"
-
- # Hadoop YARN deosn't support aarch64 at this moment. See:
https://issues.apache.org/jira/browse/HADOOP-16723
- # These tests are known to fail on JDK11. See FLINK-13719
- if [[ ${PROFILE} != *"jdk11"* ]]; then
- run_test "Running Kerberized YARN per-job on Docker test
(default input)" "$END_TO_END_DIR/test-scripts/test_yarn_job_kerberos_docker.sh"
- run_test "Running Kerberized YARN per-job on Docker test
(custom fs plugin)"
"$END_TO_END_DIR/test-scripts/test_yarn_job_kerberos_docker.sh dummy-fs"
- run_test "Running Kerberized YARN application on Docker test
(default input)"
"$END_TO_END_DIR/test-scripts/test_yarn_application_kerberos_docker.sh"
- run_test "Running Kerberized YARN application on Docker test
(custom fs plugin)"
"$END_TO_END_DIR/test-scripts/test_yarn_application_kerberos_docker.sh dummy-fs"
- fi
+ run_test "Wordcount on Docker test (custom fs plugin)"
"$END_TO_END_DIR/test-scripts/test_docker_embedded_job.sh dummy-fs"
+
+ run_test "Run Kubernetes test"
"$END_TO_END_DIR/test-scripts/test_kubernetes_embedded_job.sh"
+ run_test "Run kubernetes session test (default input)"
"$END_TO_END_DIR/test-scripts/test_kubernetes_session.sh"
+ run_test "Run kubernetes session test (custom fs plugin)"
"$END_TO_END_DIR/test-scripts/test_kubernetes_session.sh dummy-fs"
+ run_test "Run kubernetes application test"
"$END_TO_END_DIR/test-scripts/test_kubernetes_application.sh"
+ run_test "Run kubernetes application HA test"
"$END_TO_END_DIR/test-scripts/test_kubernetes_application_ha.sh"
+ run_test "Run Kubernetes IT test"
"$END_TO_END_DIR/test-scripts/test_kubernetes_itcases.sh"
+
+ run_test "Running Flink over NAT end-to-end test"
"$END_TO_END_DIR/test-scripts/test_nat.sh" "skip_check_exceptions"
+
+ if [[ `uname -i` != 'aarch64' ]]; then
+ # Skip PyFlink e2e test, because MiniConda and Pyarrow which Pyflink
depends doesn't support aarch64 currently.
+ run_test "Run kubernetes pyflink application test"
"$END_TO_END_DIR/test-scripts/test_kubernetes_pyflink_application.sh"
+
+ # Hadoop YARN deosn't support aarch64 at this moment. See:
https://issues.apache.org/jira/browse/HADOOP-16723
Review Comment:
```suggestion
# Hadoop YARN doesn't support aarch64 at this moment. See:
https://issues.apache.org/jira/browse/HADOOP-16723
```
nit
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]