This is an automated email from the ASF dual-hosted git repository.
potiuk pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/airflow-site.git
The following commit(s) were added to refs/heads/master by this push:
new 1e72444 Add docker-compose.yaml for Airflow 2.0.0 (#377)
1e72444 is described below
commit 1e72444ec3747d4b7de8702521eeb61485e80b69
Author: Kamil BreguĊa <[email protected]>
AuthorDate: Thu Feb 4 13:55:18 2021 +0100
Add docker-compose.yaml for Airflow 2.0.0 (#377)
---
.../apache-airflow/2.0.0/docker-compose.yaml | 145 +++++++++++++++++++++
1 file changed, 145 insertions(+)
diff --git a/docs-archive/apache-airflow/2.0.0/docker-compose.yaml
b/docs-archive/apache-airflow/2.0.0/docker-compose.yaml
new file mode 100644
index 0000000..2cb405c
--- /dev/null
+++ b/docs-archive/apache-airflow/2.0.0/docker-compose.yaml
@@ -0,0 +1,145 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements. See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership. The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License. You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied. See the License for the
+# specific language governing permissions and limitations
+# under the License.
+#
+
+# Basic Airflow cluster configuration for CeleryExecutor with Redis and
PostgreSQL.
+#
+# WARNING: This configuration is for local development. Do not use it in a
production deployment.
+#
+# This configuration supports basic configuration using environment variables
or an .env file
+# The following variables are supported:
+#
+# AIRFLOW_IMAGE_NAME - Docker image name used to run Airflow.
+# Default: apache/airflow:master-python3.8
+# AIRFLOW_UID - User ID in Airflow containers
+# Default: 50000
+# AIRFLOW_GID - Group ID in Airflow containers
+# Default: 50000
+# _AIRFLOW_WWW_USER_USERNAME - Username for the administrator account.
+# Default: airflow
+# _AIRFLOW_WWW_USER_PASSWORD - Password for the administrator account.
+# Default: airflow
+#
+# Feel free to modify this file to suit your needs.
+---
+version: '3'
+x-airflow-common:
+ &airflow-common
+ image: ${AIRFLOW_IMAGE_NAME:-apache/airflow:2.0.0-python3.8}
+ environment:
+ &airflow-common-env
+ AIRFLOW__CORE__EXECUTOR: CeleryExecutor
+ AIRFLOW__CORE__SQL_ALCHEMY_CONN:
postgresql+psycopg2://airflow:airflow@postgres/airflow
+ AIRFLOW__CELERY__RESULT_BACKEND:
db+postgresql://airflow:airflow@postgres/airflow
+ AIRFLOW__CELERY__BROKER_URL: redis://:@redis:6379/0
+ AIRFLOW__CORE__FERNET_KEY: ''
+ AIRFLOW__CORE__DAGS_ARE_PAUSED_AT_CREATION: 'true'
+ AIRFLOW__CORE__LOAD_EXAMPLES: 'true'
+ volumes:
+ - ./dags:/opt/airflow/dags
+ - ./logs:/opt/airflow/logs
+ - ./plugins:/opt/airflow/plugins
+ user: "${AIRFLOW_UID:-50000}:${AIRFLOW_GID:-50000}"
+ depends_on:
+ redis:
+ condition: service_healthy
+ postgres:
+ condition: service_healthy
+
+services:
+ postgres:
+ image: postgres:13
+ environment:
+ POSTGRES_USER: airflow
+ POSTGRES_PASSWORD: airflow
+ POSTGRES_DB: airflow
+ volumes:
+ - postgres-db-volume:/var/lib/postgresql/data
+ healthcheck:
+ test: ["CMD", "pg_isready", "-U", "airflow"]
+ interval: 5s
+ retries: 5
+ restart: always
+
+ redis:
+ image: redis:latest
+ ports:
+ - 6379:6379
+ healthcheck:
+ test: ["CMD", "redis-cli", "ping"]
+ interval: 5s
+ timeout: 30s
+ retries: 50
+ restart: always
+
+ airflow-webserver:
+ <<: *airflow-common
+ command: webserver
+ ports:
+ - 8080:8080
+ healthcheck:
+ test: ["CMD", "curl", "--fail", "http://localhost:8080/health"]
+ interval: 10s
+ timeout: 10s
+ retries: 5
+ restart: always
+
+ airflow-scheduler:
+ <<: *airflow-common
+ command: scheduler
+ restart: always
+
+ airflow-worker:
+ <<: *airflow-common
+ command: celery worker
+ restart: always
+
+ airflow-init:
+ <<: *airflow-common
+ entrypoint: /bin/bash
+ command:
+ - -c
+ - airflow users list || (
+ airflow db init &&
+ airflow users create
+ --role Admin
+ --username "$${_AIRFLOW_WWW_USER_USERNAME}"
+ --password "$${_AIRFLOW_WWW_USER_PASSWORD}"
+ --email [email protected]
+ --firstname airflow
+ --lastname airflow
+ )
+ environment:
+ <<: *airflow-common-env
+ _AIRFLOW_WWW_USER_USERNAME: ${_AIRFLOW_WWW_USER_USERNAME:-airflow}
+ _AIRFLOW_WWW_USER_PASSWORD: ${_AIRFLOW_WWW_USER_PASSWORD:-airflow}
+ restart: on-failure
+
+ flower:
+ <<: *airflow-common
+ command: celery flower
+ ports:
+ - 5555:5555
+ healthcheck:
+ test: ["CMD", "curl", "--fail", "http://localhost:5555/"]
+ interval: 10s
+ timeout: 10s
+ retries: 5
+ restart: always
+
+volumes:
+ postgres-db-volume: