[
https://issues.apache.org/jira/browse/AIRFLOW-5355?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
t oo closed AIRFLOW-5355.
-------------------------
Resolution: Fixed
> 1.10.4 upgrade issues - No module named kubernetes (but i'm using
> localexecutor)
> --------------------------------------------------------------------------------
>
> Key: AIRFLOW-5355
> URL: https://issues.apache.org/jira/browse/AIRFLOW-5355
> Project: Apache Airflow
> Issue Type: Bug
> Components: ui
> Affects Versions: 1.10.4
> Reporter: t oo
> Priority: Major
>
> i upgraded from 1.10.3 to 1.10.4 just now but in the main page of the ui, if
> i click refresh button next to my dag it shows error 'Broken DAG: [x.py] No
> module named kubernetes'
>
> I have debug log enabled but have no idea why it is trying to find kubernetes.
>
>
> my install steps:
> pip install cryptography mysqlclient ldap3 gunicorn[gevent]
> pip install kubernetes
> pip install apache-airflow-1.10.4-bin.tar.gz
> pip install apache-airflow-1.10.4-bin.tar.gz[kubernetes]
> airflow initdb
> airflow upgradedb
>
> my only dag has:
>
> import datetime as dt
> import glob
> import json
> import logging
> import os
> import subprocess
> import re
> from airflow import DAG
> from airflow.contrib.operators.spark_submit_operator import
> SparkSubmitOperator
> from airflow.operators.python_operator import PythonOperator
> from airflow.operators.dummy_operator import DummyOperator
> from airflow.operators.python_operator import BranchPythonOperator
> from airflow.hooks.base_hook import BaseHook
>
>
> airflow 1.10.4, localexecutor, python2, t3.large EC2, spark standalone
> scheduler
>
> it worked fine in 1.10.3
>
> some things that may be relevant:
> # when i query dag table in mysql metastore db table, some dags have
> {{last_expired populated with a timestamp but some have }}{{last_expired}}{{
> that is empty. }}
> # {{i am doing blue-green deploy so i have one ec2 running 1.10.3 and one
> ec2 running 1.10.4 but both ec2s are talking to a common mysql metastore}}
>
> UPDATE As an awful workaround i commented out all references to kube*/pod in
> many .py files from site-packages!
>
>
> some other small issues:
> a) tutorial.py dag is in the ui, how to remove? when i click delete it says
> tutorial.py not found UPDATE: delete dag from cli removed this
> b) [2019-08-30 04:12:57,714] \{scheduler_job.py:924} WARNING - Tasks using
> non-existent pool '' will not be scheduled is in logs. UPDATE: due to
> pool=None in spark_submit, once that was removed it could run.
> c) this is in logs: UPDATE -
> [https://github.com/apache/airflow/pull/5330#issuecomment-526919369] mentions
> expected
> airflow-scheduler.log-[2019-08-30 09:05:38,451] \{settings.py:327} DEBUG -
> Failed to import airflow_local_settings.
> airflow-scheduler.log-Traceback (most recent call last):
> airflow-scheduler.log- File
> "/home/ec2-user/venv/local/lib/python2.7/site-packages/airflow/settings.py",
> line 315, in import_local_settings
> airflow-scheduler.log- import airflow_local_settings
> airflow-scheduler.log:ImportError: No module named airflow_local_settings
> airflow-scheduler.log-[2019-08-30 09:05:38,452] \{logging_config.py:59}
> DEBUG - Unable to load custom logging, using default config instead
--
This message was sent by Atlassian Jira
(v8.3.4#803005)