Ben Storrie created AIRFLOW-4761:
------------------------------------
Summary: Airflow Task Clear function throws error
Key: AIRFLOW-4761
URL: https://issues.apache.org/jira/browse/AIRFLOW-4761
Project: Apache Airflow
Issue Type: Bug
Components: DAG, DagRun
Affects Versions: 1.10.3
Environment: CentOS 7, Python 2.7.10
Reporter: Ben Storrie
When using the airflow webserver to clear a task inside a dagrun, an error is
thrown on certain types of tasks:
{code:java}
Traceback (most recent call last):
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask/app.py",
line 2311, in wsgi_app
response = self.full_dispatch_request()
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask/app.py",
line 1834, in full_dispatch_request
rv = self.handle_user_exception(e)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask/app.py",
line 1737, in handle_user_exception
reraise(exc_type, exc_value, tb)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask/app.py",
line 1832, in full_dispatch_request
rv = self.dispatch_request()
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask/app.py",
line 1818, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask_admin/base.py",
line 69, in inner
return self._run_view(f, *args, **kwargs)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask_admin/base.py",
line 368, in _run_view
return fn(self, *args, **kwargs)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask_login/utils.py",
line 261, in decorated_view
return func(*args, **kwargs)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/www/utils.py",
line 275, in wrapper
return f(*args, **kwargs)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/www/utils.py",
line 322, in wrapper
return f(*args, **kwargs)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/www/views.py",
line 1202, in clear
include_upstream=upstream)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/models/__init__.py",
line 3830, in sub_dag
dag = copy.deepcopy(self)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py",
line 174, in deepcopy
y = copier(memo)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/models/__init__.py",
line 3815, in __deepcopy__
setattr(result, k, copy.deepcopy(v, memo))
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py",
line 163, in deepcopy
y = copier(x, memo)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py",
line 257, in _deepcopy_dict
y[deepcopy(key, memo)] = deepcopy(value, memo)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py",
line 174, in deepcopy
y = copier(memo)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/models/__init__.py",
line 2492, in __deepcopy__
setattr(result, k, copy.copy(v))
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py",
line 96, in copy
return _reconstruct(x, rv, 0)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py",
line 329, in _reconstruct
y = callable(*args)
File
"/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy_reg.py",
line 93, in __newobj__
return cls.__new__(cls, *args)
TypeError: instancemethod expected at least 2 arguments, got 0{code}
I had expected AIRFLOW-2060 being resolved to resolve this on upgrade to 1.10.3:
{code:java}
(my-hadoop-airflow) [user@hostname ~]$ pip freeze | grep pendulum
pendulum==1.4.4{code}
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)