Team, Any update on this?
On Fri, Jan 25, 2019 at 8:20 PM Priyanka Singh (Reco 2.0) < [email protected]> wrote: > I am trying to setup Airflow as a cluster. Celery executor is unable to > connect to rabbitmq while executing jobs Here are the configurations: > > Machine 1: webserver and scheduler > Machine 2: webserver > Machine 3: worker > Machine 4: Rabbitmq > > Airflow version: v1.8.0 > Celery version: 4.1.0 > Flower UI version: 0.9.1 > > > > **airflow.cfg** > airflow_home = ~/airflow > dags_folder = ~/airflow/dags > base_log_folder = ~/airflow/logs > executor = CeleryExecutor > sql_alchemy_conn = mysql://reco_airflow:[email protected]:3306/airflow > sql_alchemy_pool_size = 5 > dag_concurrency = 16 > dags_are_paused_at_creation = False > plugins_folder = ~/airflow/plugins > # Secret key to save connection passwords in the db > api_client = airflow.api.client.local_client > endpoint_url = http://10.34.110.227:8080 > base_url = http://10.34.110.227:8080 > web_server_host = 0.0.0.0 > web_server_port = 8080 > workers = 4 > broker_url = amqp://guest:[email protected]:5672// > celery_result_backend = > db+mysql://reco_airflow:[email protected]:3306/airflow > flower_host = 10.34.110.227 > flower_port = 5555 > default_queue = queue > > Dags are running successfully using celery executor. But I can't see any > connection with rabbitmq(it is always idle). Also, if I try to use flower, > the UI opens but keeps on loading and stop responding after ~5 sec. No > error is coming in logs. Am I missing something in the configuration? > > > Here is the Stackoverflow link: > > > https://stackoverflow.com/questions/54367536/issues-in-running-airflow-as-cluster-with-celery-executors >
