[jira] [Closed] (AIRFLOW-1407) Enable tests in Spark submit operator
[ https://issues.apache.org/jira/browse/AIRFLOW-1407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fokko Driesprong closed AIRFLOW-1407. - Resolution: Duplicate > Enable tests in Spark submit operator > - > > Key: AIRFLOW-1407 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1407 > Project: Apache Airflow > Issue Type: Bug >Reporter: Fokko Driesprong > > Currently the tests are disabled at the spark-submit operator for Python 3. > This is definitely something that we don't want. I would like to enable the > tests and fix the underlying issue. -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Commented] (AIRFLOW-1407) Enable tests in Spark submit operator
[ https://issues.apache.org/jira/browse/AIRFLOW-1407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16084475#comment-16084475 ] Fokko Driesprong commented on AIRFLOW-1407: --- Ok, missed that one! Good work > Enable tests in Spark submit operator > - > > Key: AIRFLOW-1407 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1407 > Project: Apache Airflow > Issue Type: Bug >Reporter: Fokko Driesprong > > Currently the tests are disabled at the spark-submit operator for Python 3. > This is definitely something that we don't want. I would like to enable the > tests and fix the underlying issue. -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Closed] (AIRFLOW-1402) Cleanup SafeConfigParser DeprecationWarning
[ https://issues.apache.org/jira/browse/AIRFLOW-1402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Riccomini closed AIRFLOW-1402. Resolution: Fixed Fix Version/s: 1.9.0 > Cleanup SafeConfigParser DeprecationWarning > --- > > Key: AIRFLOW-1402 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1402 > Project: Apache Airflow > Issue Type: Bug >Reporter: Ron Fung >Assignee: Ron Fung >Priority: Minor > Fix For: 1.9.0 > > > {code} DeprecationWarning: The SafeConfigParser class has been renamed to > ConfigParser in Python 3.2. This alias will be removed in future versions. > Use ConfigParser directly instead. {code} -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Updated] (AIRFLOW-1408) Add CreditCards.com
[ https://issues.apache.org/jira/browse/AIRFLOW-1408?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vikas Malhotra updated AIRFLOW-1408: Description: Add to README.md: ``` [CreditCards.com](https://www.creditcards.com/)[[[~vikas.malhotra]](https://github.com/vmAggies), (https://github.com/jay-wallaby)] ``` was: Add to README.md: ``` [iHeartRadio](http://www.iheart.com/)[[@yiwang](https://github.com/yiwang)] ``` > Add CreditCards.com > --- > > Key: AIRFLOW-1408 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1408 > Project: Apache Airflow > Issue Type: Wish > Components: docs >Reporter: Vikas Malhotra >Priority: Trivial > Labels: beginner, document, newbie > > Add to README.md: > ``` > > [CreditCards.com](https://www.creditcards.com/)[[[~vikas.malhotra]](https://github.com/vmAggies), > (https://github.com/jay-wallaby)] > ``` -- This message was sent by Atlassian JIRA (v6.4.14#64029)
incubator-airflow git commit: [AIRFLOW-1402] Cleanup SafeConfigParser DeprecationWarning
Repository: incubator-airflow Updated Branches: refs/heads/master f5ffb9deb -> 38c86bbbc [AIRFLOW-1402] Cleanup SafeConfigParser DeprecationWarning Closes #2435 from ronfung/work Project: http://git-wip-us.apache.org/repos/asf/incubator-airflow/repo Commit: http://git-wip-us.apache.org/repos/asf/incubator-airflow/commit/38c86bbb Tree: http://git-wip-us.apache.org/repos/asf/incubator-airflow/tree/38c86bbb Diff: http://git-wip-us.apache.org/repos/asf/incubator-airflow/diff/38c86bbb Branch: refs/heads/master Commit: 38c86bbbcbc85ec470f31a33f2a480ea45510a45 Parents: f5ffb9d Author: Ron FungAuthored: Wed Jul 12 10:09:24 2017 -0700 Committer: Chris Riccomini Committed: Wed Jul 12 10:09:24 2017 -0700 -- airflow/configuration.py | 5 - 1 file changed, 4 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/incubator-airflow/blob/38c86bbb/airflow/configuration.py -- diff --git a/airflow/configuration.py b/airflow/configuration.py index e325067..460d975 100644 --- a/airflow/configuration.py +++ b/airflow/configuration.py @@ -42,7 +42,10 @@ warnings.filterwarnings( warnings.filterwarnings( action='default', category=PendingDeprecationWarning, module='airflow') -ConfigParser = configparser.SafeConfigParser +if six.PY3: +ConfigParser = configparser.ConfigParser +else: +ConfigParser = configparser.SafeConfigParser def generate_fernet_key():
[jira] [Commented] (AIRFLOW-1402) Cleanup SafeConfigParser DeprecationWarning
[ https://issues.apache.org/jira/browse/AIRFLOW-1402?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16084317#comment-16084317 ] ASF subversion and git services commented on AIRFLOW-1402: -- Commit 38c86bbbcbc85ec470f31a33f2a480ea45510a45 in incubator-airflow's branch refs/heads/master from [~ronfung] [ https://git-wip-us.apache.org/repos/asf?p=incubator-airflow.git;h=38c86bb ] [AIRFLOW-1402] Cleanup SafeConfigParser DeprecationWarning Closes #2435 from ronfung/work > Cleanup SafeConfigParser DeprecationWarning > --- > > Key: AIRFLOW-1402 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1402 > Project: Apache Airflow > Issue Type: Bug >Reporter: Ron Fung >Assignee: Ron Fung >Priority: Minor > > {code} DeprecationWarning: The SafeConfigParser class has been renamed to > ConfigParser in Python 3.2. This alias will be removed in future versions. > Use ConfigParser directly instead. {code} -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (AIRFLOW-1408) Add CreditCards.com
Vikas Malhotra created AIRFLOW-1408: --- Summary: Add CreditCards.com Key: AIRFLOW-1408 URL: https://issues.apache.org/jira/browse/AIRFLOW-1408 Project: Apache Airflow Issue Type: Wish Components: docs Reporter: Vikas Malhotra Priority: Trivial Add to README.md: ``` [iHeartRadio](http://www.iheart.com/)[[@yiwang](https://github.com/yiwang)] ``` -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Commented] (AIRFLOW-1407) Enable tests in Spark submit operator
[ https://issues.apache.org/jira/browse/AIRFLOW-1407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16084147#comment-16084147 ] Ash Berlin-Taylor commented on AIRFLOW-1407: Duplicate of AIRFLOW-1393, and I've made a start fixing the tests (py3 binary vs string differences in the data tests are feeding) > Enable tests in Spark submit operator > - > > Key: AIRFLOW-1407 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1407 > Project: Apache Airflow > Issue Type: Bug >Reporter: Fokko Driesprong > > Currently the tests are disabled at the spark-submit operator for Python 3. > This is definitely something that we don't want. I would like to enable the > tests and fix the underlying issue. -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (AIRFLOW-1407) Enable tests in Spark submit operator
Fokko Driesprong created AIRFLOW-1407: - Summary: Enable tests in Spark submit operator Key: AIRFLOW-1407 URL: https://issues.apache.org/jira/browse/AIRFLOW-1407 Project: Apache Airflow Issue Type: Bug Reporter: Fokko Driesprong Currently the tests are disabled at the spark-submit operator for Python 3. This is definitely something that we don't want. I would like to enable the tests and fix the underlying issue. -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Created] (AIRFLOW-1406) ignore_depends_on_past property ignored by backfill
Tobias Feldhaus created AIRFLOW-1406: Summary: ignore_depends_on_past property ignored by backfill Key: AIRFLOW-1406 URL: https://issues.apache.org/jira/browse/AIRFLOW-1406 Project: Apache Airflow Issue Type: Bug Components: backfill Affects Versions: Airflow 1.8 Reporter: Tobias Feldhaus Trying to run the following command does not ignore the first depends_on_past dependency of the DAG that has depends_on_past set to true: {code}airflow backfill -t search_log_sensor -s 2017-07-10 -e 2017-07-11 -I google_pipelines{code} neither does this one: {code}airflow backfill -t search_log_sensor -s 2017-07-10 -e 2017-07-11 --ignore_depends_on_past google_pipelines {code} both result in: {code} BackfillJob is deadlocked.Some of the deadlocked tasks were unable to run because of "depends_on_past" relationships. Try running the backfill with the option "ignore_first_depends_on_past=True" or passing "-I" at the command line. {code} Trying to run with {code} --ignore_depends_on_past=True {code} yields: {code} airflow backfill: error: argument -I/--ignore_first_depends_on_past: ignored explicit argument 'True' {code} -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Commented] (AIRFLOW-1400) catchup=False caused exception
[ https://issues.apache.org/jira/browse/AIRFLOW-1400?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16083708#comment-16083708 ] Xi Wang commented on AIRFLOW-1400: -- [~dxhuang] It's @once. is that the issue? thx. > catchup=False caused exception > -- > > Key: AIRFLOW-1400 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1400 > Project: Apache Airflow > Issue Type: Bug > Components: scheduler >Affects Versions: Airflow 1.8 >Reporter: Xi Wang > > When I set of the task with catchup=False, it threw a error as > follow(logs/scheduler/my_dag_name): > [2017-07-10 15:13:12,534] {jobs.py:354} DagFileProcessor373 ERROR - Got an > exception! Propagating... > Traceback (most recent call last): > File "/usr/lib/python2.7/site-packages/airflow/jobs.py", line 346, in helper > pickle_dags) > File "/usr/lib/python2.7/site-packages/airflow/utils/db.py", line 53, in > wrapper > result = func(*args, **kwargs) > File "/usr/lib/python2.7/site-packages/airflow/jobs.py", line 1581, in > process_file > self._process_dags(dagbag, dags, ti_keys_to_schedule) > File "/usr/lib/python2.7/site-packages/airflow/jobs.py", line 1171, in > _process_dags > dag_run = self.create_dag_run(dag) > File "/usr/lib/python2.7/site-packages/airflow/utils/db.py", line 53, in > wrapper > result = func(*args, **kwargs) > File "/usr/lib/python2.7/site-packages/airflow/jobs.py", line 776, in > create_dag_run > if next_start <= now: > TypeError: can't compare datetime.datetime to NoneType > It seems next_start was not defined properly, > (https://github.com/apache/incubator-airflow/blob/master/airflow/jobs.py#L777) > Any help is appreciated. -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Resolved] (AIRFLOW-1326) Contrib Spark Submit Hook should be able to produce arguments containing spaces.
[ https://issues.apache.org/jira/browse/AIRFLOW-1326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bolke de Bruin resolved AIRFLOW-1326. - Resolution: Fixed Fix Version/s: (was: Airflow 1.8) (was: Airflow 2.0) 1.8.3 Issue resolved by pull request #2382 [https://github.com/apache/incubator-airflow/pull/2382] > Contrib Spark Submit Hook should be able to produce arguments containing > spaces. > > > Key: AIRFLOW-1326 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1326 > Project: Apache Airflow > Issue Type: Bug > Components: contrib, hooks >Affects Versions: Airflow 2.0, Airflow 1.8 >Reporter: Ash Berlin-Taylor >Assignee: Vianney FOUCAULT > Fix For: 1.8.3 > > > AIRFLOW-1184 introduced a bug making it impossible to include an argument > with an embedded space in the SparkSubmitOperator. (I believe the original > bug report should have been closed as wontfix/not-a-bug as it was user error > in how the SparkSubmitOperator is being called) > i.e. the correct way to achieve what was wanted in the bug is like this: > {code} > 'application_args': [ > '-f', ' foo', > '--bar', 'bar', > '--start', '{{ macros.ds_add(ds, -1)}}', > '--end', '{{ ds }}' > ] > {code} > This was possibly compounded as the tests were showing something that would > never "work" -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Resolved] (AIRFLOW-1184) Contrib Spark Submit Hook does not split argument and argument value
[ https://issues.apache.org/jira/browse/AIRFLOW-1184?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bolke de Bruin resolved AIRFLOW-1184. - Resolution: Fixed Fix Version/s: (was: Airflow 1.8) (was: Airflow 2.0) 1.8.3 Issue resolved by pull request #2382 [https://github.com/apache/incubator-airflow/pull/2382] > Contrib Spark Submit Hook does not split argument and argument value > > > Key: AIRFLOW-1184 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1184 > Project: Apache Airflow > Issue Type: Bug > Components: contrib, hooks >Affects Versions: Airflow 2.0, Airflow 1.8 >Reporter: Vianney FOUCAULT >Assignee: Vianney FOUCAULT > Fix For: 1.8.3 > > > Python Popen expect a list as command. Spark submit too, as: > * ['--option value'] > is not the same as > * ['--option', 'value'] > in regards of spark. eg spark logs : (yarn logs) > Error: Unknown option --end 2017-05-08 > Error: Unknown option --begin 2017-05-07 > Error: Unknown option --db_name mydb > Error: Missing option --begin > Error: Missing option --end > Error: Missing option --db_name -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Commented] (AIRFLOW-1326) Contrib Spark Submit Hook should be able to produce arguments containing spaces.
[ https://issues.apache.org/jira/browse/AIRFLOW-1326?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16083694#comment-16083694 ] ASF subversion and git services commented on AIRFLOW-1326: -- Commit f5ffb9deb51e896cdda8992c8b28102f7d650662 in incubator-airflow's branch refs/heads/master from [~ashb] [ https://git-wip-us.apache.org/repos/asf?p=incubator-airflow.git;h=f5ffb9d ] [AIRFLOW-1326][[AIRFLOW-1326][AIRFLOW-1184] Don't split argument array -- it's already an array.[ Closes #2382 from ashb/spark-submit-operator- preserve-spaces > Contrib Spark Submit Hook should be able to produce arguments containing > spaces. > > > Key: AIRFLOW-1326 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1326 > Project: Apache Airflow > Issue Type: Bug > Components: contrib, hooks >Affects Versions: Airflow 2.0, Airflow 1.8 >Reporter: Ash Berlin-Taylor >Assignee: Vianney FOUCAULT > Fix For: Airflow 2.0, Airflow 1.8 > > > AIRFLOW-1184 introduced a bug making it impossible to include an argument > with an embedded space in the SparkSubmitOperator. (I believe the original > bug report should have been closed as wontfix/not-a-bug as it was user error > in how the SparkSubmitOperator is being called) > i.e. the correct way to achieve what was wanted in the bug is like this: > {code} > 'application_args': [ > '-f', ' foo', > '--bar', 'bar', > '--start', '{{ macros.ds_add(ds, -1)}}', > '--end', '{{ ds }}' > ] > {code} > This was possibly compounded as the tests were showing something that would > never "work" -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Commented] (AIRFLOW-1184) Contrib Spark Submit Hook does not split argument and argument value
[ https://issues.apache.org/jira/browse/AIRFLOW-1184?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16083696#comment-16083696 ] ASF subversion and git services commented on AIRFLOW-1184: -- Commit f5ffb9deb51e896cdda8992c8b28102f7d650662 in incubator-airflow's branch refs/heads/master from [~ashb] [ https://git-wip-us.apache.org/repos/asf?p=incubator-airflow.git;h=f5ffb9d ] [AIRFLOW-1326][[AIRFLOW-1326][AIRFLOW-1184] Don't split argument array -- it's already an array.[ Closes #2382 from ashb/spark-submit-operator- preserve-spaces > Contrib Spark Submit Hook does not split argument and argument value > > > Key: AIRFLOW-1184 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1184 > Project: Apache Airflow > Issue Type: Bug > Components: contrib, hooks >Affects Versions: Airflow 2.0, Airflow 1.8 >Reporter: Vianney FOUCAULT >Assignee: Vianney FOUCAULT > Fix For: Airflow 2.0, Airflow 1.8 > > > Python Popen expect a list as command. Spark submit too, as: > * ['--option value'] > is not the same as > * ['--option', 'value'] > in regards of spark. eg spark logs : (yarn logs) > Error: Unknown option --end 2017-05-08 > Error: Unknown option --begin 2017-05-07 > Error: Unknown option --db_name mydb > Error: Missing option --begin > Error: Missing option --end > Error: Missing option --db_name -- This message was sent by Atlassian JIRA (v6.4.14#64029)
[jira] [Commented] (AIRFLOW-1326) Contrib Spark Submit Hook should be able to produce arguments containing spaces.
[ https://issues.apache.org/jira/browse/AIRFLOW-1326?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16083695#comment-16083695 ] ASF subversion and git services commented on AIRFLOW-1326: -- Commit f5ffb9deb51e896cdda8992c8b28102f7d650662 in incubator-airflow's branch refs/heads/master from [~ashb] [ https://git-wip-us.apache.org/repos/asf?p=incubator-airflow.git;h=f5ffb9d ] [AIRFLOW-1326][[AIRFLOW-1326][AIRFLOW-1184] Don't split argument array -- it's already an array.[ Closes #2382 from ashb/spark-submit-operator- preserve-spaces > Contrib Spark Submit Hook should be able to produce arguments containing > spaces. > > > Key: AIRFLOW-1326 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1326 > Project: Apache Airflow > Issue Type: Bug > Components: contrib, hooks >Affects Versions: Airflow 2.0, Airflow 1.8 >Reporter: Ash Berlin-Taylor >Assignee: Vianney FOUCAULT > Fix For: Airflow 2.0, Airflow 1.8 > > > AIRFLOW-1184 introduced a bug making it impossible to include an argument > with an embedded space in the SparkSubmitOperator. (I believe the original > bug report should have been closed as wontfix/not-a-bug as it was user error > in how the SparkSubmitOperator is being called) > i.e. the correct way to achieve what was wanted in the bug is like this: > {code} > 'application_args': [ > '-f', ' foo', > '--bar', 'bar', > '--start', '{{ macros.ds_add(ds, -1)}}', > '--end', '{{ ds }}' > ] > {code} > This was possibly compounded as the tests were showing something that would > never "work" -- This message was sent by Atlassian JIRA (v6.4.14#64029)
incubator-airflow git commit: [AIRFLOW-1326][[AIRFLOW-1326][AIRFLOW-1184] Don't split argument array -- it's already an array.[
Repository: incubator-airflow Updated Branches: refs/heads/master 71594751c -> f5ffb9deb [AIRFLOW-1326][[AIRFLOW-1326][AIRFLOW-1184] Don't split argument array -- it's already an array.[ Closes #2382 from ashb/spark-submit-operator- preserve-spaces Project: http://git-wip-us.apache.org/repos/asf/incubator-airflow/repo Commit: http://git-wip-us.apache.org/repos/asf/incubator-airflow/commit/f5ffb9de Tree: http://git-wip-us.apache.org/repos/asf/incubator-airflow/tree/f5ffb9de Diff: http://git-wip-us.apache.org/repos/asf/incubator-airflow/diff/f5ffb9de Branch: refs/heads/master Commit: f5ffb9deb51e896cdda8992c8b28102f7d650662 Parents: 7159475 Author: Ash Berlin-TaylorAuthored: Wed Jul 12 11:27:39 2017 +0200 Committer: Bolke de Bruin Committed: Wed Jul 12 11:27:39 2017 +0200 -- airflow/contrib/hooks/spark_submit_hook.py | 8 ++--- tests/contrib/hooks/test_spark_submit_hook.py | 6 ++-- .../operators/test_spark_submit_operator.py | 31 +++- 3 files changed, 23 insertions(+), 22 deletions(-) -- http://git-wip-us.apache.org/repos/asf/incubator-airflow/blob/f5ffb9de/airflow/contrib/hooks/spark_submit_hook.py -- diff --git a/airflow/contrib/hooks/spark_submit_hook.py b/airflow/contrib/hooks/spark_submit_hook.py index f3d6e34..88d547b 100644 --- a/airflow/contrib/hooks/spark_submit_hook.py +++ b/airflow/contrib/hooks/spark_submit_hook.py @@ -194,12 +194,8 @@ class SparkSubmitHook(BaseHook): # Append any application arguments if self._application_args: -for arg in self._application_args: -if len(arg.split()) > 1: -for splitted_option in arg.split(): -connection_cmd += [splitted_option] -else: -connection_cmd += [arg] +connection_cmd += self._application_args + logging.debug("Spark-Submit cmd: {}".format(connection_cmd)) return connection_cmd http://git-wip-us.apache.org/repos/asf/incubator-airflow/blob/f5ffb9de/tests/contrib/hooks/test_spark_submit_hook.py -- diff --git a/tests/contrib/hooks/test_spark_submit_hook.py b/tests/contrib/hooks/test_spark_submit_hook.py index ef0857b..98e959b 100644 --- a/tests/contrib/hooks/test_spark_submit_hook.py +++ b/tests/contrib/hooks/test_spark_submit_hook.py @@ -45,8 +45,9 @@ class TestSparkSubmitHook(unittest.TestCase): 'driver_memory': '3g', 'java_class': 'com.foo.bar.AppMain', 'application_args': [ -'-f foo', -'--bar bar', +'-f', 'foo', +'--bar', 'bar', +'--with-spaces', 'args should keep embdedded spaces', 'baz' ] } @@ -129,6 +130,7 @@ class TestSparkSubmitHook(unittest.TestCase): 'test_application.py', '-f', 'foo', '--bar', 'bar', +'--with-spaces', 'args should keep embdedded spaces', 'baz' ] self.assertEquals(expected_build_cmd, cmd) http://git-wip-us.apache.org/repos/asf/incubator-airflow/blob/f5ffb9de/tests/contrib/operators/test_spark_submit_operator.py -- diff --git a/tests/contrib/operators/test_spark_submit_operator.py b/tests/contrib/operators/test_spark_submit_operator.py index 09c5a93..1d41941 100644 --- a/tests/contrib/operators/test_spark_submit_operator.py +++ b/tests/contrib/operators/test_spark_submit_operator.py @@ -46,10 +46,11 @@ class TestSparkSubmitOperator(unittest.TestCase): 'driver_memory': '3g', 'java_class': 'com.foo.bar.AppMain', 'application_args': [ -'-f foo', -'--bar bar', -'--start {{ macros.ds_add(ds, -1)}}', -'--end {{ ds }}' +'-f', 'foo', +'--bar', 'bar', +'--start', '{{ macros.ds_add(ds, -1)}}', +'--end', '{{ ds }}', +'--with-spaces', 'args should keep embdedded spaces', ] } @@ -95,10 +96,11 @@ class TestSparkSubmitOperator(unittest.TestCase): 'driver_memory': '3g', 'java_class': 'com.foo.bar.AppMain', 'application_args': [ -'-f foo', -'--bar bar', -'--start {{ macros.ds_add(ds, -1)}}', -'--end {{ ds }}' +'-f', 'foo', +'--bar', 'bar', +'--start', '{{ macros.ds_add(ds, -1)}}', +'--end', '{{ ds }}', +'--with-spaces', 'args should keep embdedded spaces', ] } @@ -130,14 +132,15 @@ class
[jira] [Created] (AIRFLOW-1405) Airflow v 1.8.1 unable to properly initialize with MySQL
Aakash Bhardwaj created AIRFLOW-1405: Summary: Airflow v 1.8.1 unable to properly initialize with MySQL Key: AIRFLOW-1405 URL: https://issues.apache.org/jira/browse/AIRFLOW-1405 Project: Apache Airflow Issue Type: Bug Components: db Affects Versions: 1.8.1 Environment: CentOS7 Reporter: Aakash Bhardwaj Fix For: 1.8.1 Attachments: error_log.txt While working on a CentOS7 system, I was trying to configure Airflow version 1.8.1 to run with MySql in the backend. I have installed Airflow in a Virtual Environment, and the MySQL has a database named airflow (default). But on running the command - {code:shell} airflow initdb {code} the following error is reported {noformat} [2017-07-12 13:22:36,558] {__init__.py:57} INFO - Using executor LocalExecutor DB: mysql://airflow:***@localhost/airflow [2017-07-12 13:22:37,218] {db.py:287} INFO - Creating tables INFO [alembic.runtime.migration] Context impl MySQLImpl. INFO [alembic.runtime.migration] Will assume non-transactional DDL. INFO [alembic.runtime.migration] Running upgrade f2ca10b85618 -> 4addfa1236f1, Add fractional seconds to mysql tables Traceback (most recent call last): File "/opt/airflow_virtual_environment/airflow_venv/bin/airflow", line 28, in args.func(args) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/bin/cli.py", line 951, in initdb db_utils.initdb() File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/utils/db.py", line 106, in initdb upgradedb() File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/utils/db.py", line 294, in upgradedb command.upgrade(config, 'heads') File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/command.py", line 174, in upgrade script.run_env() File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/script/base.py", line 416, in run_env util.load_python_file(self.dir, 'env.py') File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/util/pyfiles.py", line 93, in load_python_file module = load_module_py(module_id, path) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/util/compat.py", line 79, in load_module_py mod = imp.load_source(module_id, path, fp) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/migrations/env.py", line 86, in run_migrations_online() File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/migrations/env.py", line 81, in run_migrations_online context.run_migrations() File "", line 8, in run_migrations File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/runtime/environment.py", line 807, in run_migrations self.get_context().run_migrations(**kw) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/runtime/migration.py", line 321, in run_migrations step.migration_fn(**kw) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/migrations/versions/4addfa1236f1_add_fractional_seconds_to_mysql_tables.py", line 36, in upgrade op.alter_column(table_name='dag', column_name='last_scheduler_run', type_=mysql.DATETIME(fsp=6)) File "", line 8, in alter_column File "", line 3, in alter_column File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/operations/ops.py", line 1420, in alter_column return operations.invoke(alt) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/operations/base.py", line 318, in invoke return fn(self, operation) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/operations/toimpl.py", line 53, in alter_column **operation.kw File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/ddl/mysql.py", line 67, in alter_column else existing_autoincrement File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/ddl/impl.py", line 118, in _exec return conn.execute(construct, *multiparams, **params) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 945, in execute return meth(self, multiparams, params) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/sqlalchemy/sql/ddl.py", line 68, in _execute_on_connection return connection._execute_ddl(self, multiparams, params) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 1002, in
[jira] [Updated] (AIRFLOW-1405) Airflow v 1.8.1 unable to properly initialize with MySQL
[ https://issues.apache.org/jira/browse/AIRFLOW-1405?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Aakash Bhardwaj updated AIRFLOW-1405: - Description: While working on a CentOS7 system, I was trying to configure Airflow version 1.8.1 to run with MySql in the backend. I have installed Airflow in a Virtual Environment, and the MySQL has a database named airflow (default). But on running the command - {code:none} airflow initdb {code} the following error is reported {noformat} [2017-07-12 13:22:36,558] {__init__.py:57} INFO - Using executor LocalExecutor DB: mysql://airflow:***@localhost/airflow [2017-07-12 13:22:37,218] {db.py:287} INFO - Creating tables INFO [alembic.runtime.migration] Context impl MySQLImpl. INFO [alembic.runtime.migration] Will assume non-transactional DDL. INFO [alembic.runtime.migration] Running upgrade f2ca10b85618 -> 4addfa1236f1, Add fractional seconds to mysql tables Traceback (most recent call last): File "/opt/airflow_virtual_environment/airflow_venv/bin/airflow", line 28, in args.func(args) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/bin/cli.py", line 951, in initdb db_utils.initdb() File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/utils/db.py", line 106, in initdb upgradedb() File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/utils/db.py", line 294, in upgradedb command.upgrade(config, 'heads') File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/command.py", line 174, in upgrade script.run_env() File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/script/base.py", line 416, in run_env util.load_python_file(self.dir, 'env.py') File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/util/pyfiles.py", line 93, in load_python_file module = load_module_py(module_id, path) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/util/compat.py", line 79, in load_module_py mod = imp.load_source(module_id, path, fp) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/migrations/env.py", line 86, in run_migrations_online() File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/migrations/env.py", line 81, in run_migrations_online context.run_migrations() File "", line 8, in run_migrations File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/runtime/environment.py", line 807, in run_migrations self.get_context().run_migrations(**kw) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/runtime/migration.py", line 321, in run_migrations step.migration_fn(**kw) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/airflow/migrations/versions/4addfa1236f1_add_fractional_seconds_to_mysql_tables.py", line 36, in upgrade op.alter_column(table_name='dag', column_name='last_scheduler_run', type_=mysql.DATETIME(fsp=6)) File "", line 8, in alter_column File "", line 3, in alter_column File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/operations/ops.py", line 1420, in alter_column return operations.invoke(alt) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/operations/base.py", line 318, in invoke return fn(self, operation) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/operations/toimpl.py", line 53, in alter_column **operation.kw File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/ddl/mysql.py", line 67, in alter_column else existing_autoincrement File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/alembic/ddl/impl.py", line 118, in _exec return conn.execute(construct, *multiparams, **params) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 945, in execute return meth(self, multiparams, params) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/sqlalchemy/sql/ddl.py", line 68, in _execute_on_connection return connection._execute_ddl(self, multiparams, params) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 1002, in _execute_ddl compiled File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 1189, in _execute_context context) File "/opt/airflow_virtual_environment/airflow_venv/lib/python2.7/site-packages/sqlalchemy/engine/base.py",
[jira] [Created] (AIRFLOW-1404) Big Query Operator add max_billing_tier, create_disposition and flatten_results
Alex Aslan created AIRFLOW-1404: --- Summary: Big Query Operator add max_billing_tier, create_disposition and flatten_results Key: AIRFLOW-1404 URL: https://issues.apache.org/jira/browse/AIRFLOW-1404 Project: Apache Airflow Issue Type: Wish Components: operators Affects Versions: Airflow 1.8 Reporter: Alex Aslan Priority: Trivial I would like to have the option to set explicitly the params max_billing_tier, flatten_results and create_disposition in the contrib big query operator -- This message was sent by Atlassian JIRA (v6.4.14#64029)