This is an automated email from the ASF dual-hosted git repository.

mapohl pushed a commit to branch release-1.18
in repository https://gitbox.apache.org/repos/asf/flink.git


The following commit(s) were added to refs/heads/release-1.18 by this push:
     new a27e03650d9 [FLINK-33915][ci] Adds workflow release-1.18 nightly CI 
workflow
a27e03650d9 is described below

commit a27e03650d9e03c9317e19051f8aa4871f199f4a
Author: Matthias Pohl <matthias.p...@aiven.io>
AuthorDate: Fri Dec 15 10:46:13 2023 +0100

    [FLINK-33915][ci] Adds workflow release-1.18 nightly CI workflow
---
 .github/workflows/nightly.yml | 85 +++++++++++++++++++++++++++++++++++++++++++
 1 file changed, 85 insertions(+)

diff --git a/.github/workflows/nightly.yml b/.github/workflows/nightly.yml
new file mode 100644
index 00000000000..c02da2e4954
--- /dev/null
+++ b/.github/workflows/nightly.yml
@@ -0,0 +1,85 @@
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license agreements.  See the NOTICE file distributed with
+# this work for additional information regarding copyright ownership.
+# The ASF licenses this file to You under the Apache License, Version 2.0
+# (the "License"); you may not use this file except in compliance with
+# the License.  You may obtain a copy of the License at
+#
+#    http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+# This workflow is meant as an extended CI run that includes certain features 
that shall be tested
+# and JDK versions that are supported but not considered default.
+
+name: "Nightly (beta)"
+
+on:
+  workflow_dispatch:
+
+permissions: read-all
+
+jobs:
+  pre-compile-checks:
+    name: "Pre-compile Checks"
+    uses: ./.github/workflows/template.pre-compile-checks.yml
+
+  java8:
+    name: "Java 8"
+    uses: ./.github/workflows/template.flink-ci.yml
+    with:
+      workflow-caller-id: java8
+      environment: 'PROFILE="-Dinclude_hadoop_aws"'
+      jdk_version: 8
+    secrets:
+      s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }}
+      s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }}
+      s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }}
+  java11:
+    name: "Java 11"
+    uses: ./.github/workflows/template.flink-ci.yml
+    with:
+      workflow-caller-id: java11
+      environment: 'PROFILE="-Dinclude_hadoop_aws -Djdk11 -Pjava11-target"'
+      jdk_version: 11
+    secrets:
+      s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }}
+      s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }}
+      s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }}
+  java17:
+    name: "Java 17"
+    uses: ./.github/workflows/template.flink-ci.yml
+    with:
+      workflow-caller-id: java17
+      environment: 'PROFILE="-Dinclude_hadoop_aws -Djdk11 -Djdk17 
-Pjava17-target"'
+      jdk_version: 17
+    secrets:
+      s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }}
+      s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }}
+      s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }}
+  hadoop313:
+    name: "Hadoop 3.1.3"
+    uses: ./.github/workflows/template.flink-ci.yml
+    with:
+      workflow-caller-id: hadoop313
+      environment: 'PROFILE="-Dflink.hadoop.version=3.2.3 
-Phadoop3-tests,hive3"'
+      jdk_version: 8
+    secrets:
+      s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }}
+      s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }}
+      s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }}
+  adaptive-scheduler:
+    name: "AdaptiveScheduler"
+    uses: ./.github/workflows/template.flink-ci.yml
+    with:
+      workflow-caller-id: adaptive-scheduler
+      environment: 'PROFILE="-Penable-adaptive-scheduler"'
+      jdk_version: 8
+    secrets:
+      s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }}
+      s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }}
+      s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }}

Reply via email to