[jira] [Created] (AIRFLOW-3560) Add Sensor that polls until a day of the week

2018-12-22 Thread Kaxil Naik (JIRA)
Kaxil Naik created AIRFLOW-3560:
---

 Summary: Add Sensor that polls until a day of the week
 Key: AIRFLOW-3560
 URL: https://issues.apache.org/jira/browse/AIRFLOW-3560
 Project: Apache Airflow
  Issue Type: New Feature
Reporter: Kaxil Naik
Assignee: Kaxil Naik
 Fix For: 1.10.2


One of the use-case we have is we want to run certain tasks only on Weekends



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3150) Make execution_date a template field in TriggerDagRunOperator

2018-12-22 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3150:

Fix Version/s: 1.10.2

> Make execution_date a template field in TriggerDagRunOperator
> -
>
> Key: AIRFLOW-3150
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3150
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: operators
>Reporter: Kyle Hamlin
>Assignee: Kaxil Naik
>Priority: Minor
>  Labels: easy-fix
> Fix For: 1.10.2
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-850) Airflow should support a general purpose PythonSensor

2018-12-22 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-850.

   Resolution: Fixed
Fix Version/s: 1.10.2

Resolved by https://github.com/apache/incubator-airflow/pull/4349

> Airflow should support a general purpose PythonSensor
> -
>
> Key: AIRFLOW-850
> URL: https://issues.apache.org/jira/browse/AIRFLOW-850
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: operators
>Affects Versions: 1.8.0
>Reporter: Daniel Gies
>Assignee: Daniel Gies
>Priority: Major
> Fix For: 1.10.2
>
>
> Today I found myself trying to use a sensor to postpone execution until data 
> for the current execution date appeared in a file.  It occurred to me that 
> having a general purpose PythonSensor would allow developers to use the 
> sensor paradigm with arbitrary code.
> We should add a PythonSensor to the core sensors module which takes a 
> python_callable and optional args like the PythonOperator does.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3557) Various typos

2018-12-22 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3557?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3557.
-
   Resolution: Fixed
Fix Version/s: 2.0.0

Resolved by https://github.com/apache/incubator-airflow/pull/4357

> Various typos
> -
>
> Key: AIRFLOW-3557
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3557
> Project: Apache Airflow
>  Issue Type: Bug
>Reporter: Bas Harenslak
>Priority: Major
> Fix For: 2.0.0
>
>
> Fix various typos, checked with 
> [misspell|https://github.com/client9/misspell].



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3398) Google Cloud Spanner instance database query operator

2018-12-19 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3398?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3398.
-
   Resolution: Fixed
 Assignee: (was: Szymon Przedwojski)
Fix Version/s: 1.10.2

Resolved by https://github.com/apache/incubator-airflow/pull/4314

> Google Cloud Spanner instance database query operator
> -
>
> Key: AIRFLOW-3398
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3398
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: gcp
>Reporter: Szymon Przedwojski
>Priority: Minor
> Fix For: 1.10.2
>
>
> Creating an operator to enable executing arbitrary SQL in a Transaction in 
> Cloud Spanner.
> https://googleapis.github.io/google-cloud-python/latest/spanner/index.html#executing-arbitrary-sql-in-a-transaction



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3500) Make task duration display user friendly

2018-12-15 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3500:

Fix Version/s: (was: 1.10.2)
   2.0.0

> Make task duration display user friendly
> 
>
> Key: AIRFLOW-3500
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3500
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: DAG
>Affects Versions: 1.10.1
>Reporter: Ofer Zelig
>Assignee: Ofer Zelig
>Priority: Major
> Fix For: 2.0.0
>
>
> When hovering over a task (in Graph mode), the duration it took is displayed 
> as a plain number, which doesn't say what the number is (it's actually 
> seconds).
> When you see something like 2716 it's impractical to know how much time it 
> actually took, unless you can quickly do the math in your head.
> Change the display to read days/hours/minutes/seconds.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3447) Intended usage of ts_nodash macro broken with migration to new time system.

2018-12-15 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3447.
-
Resolution: Fixed

Resolved by https://github.com/apache/incubator-airflow/pull/4323

> Intended usage of ts_nodash macro broken with migration to new time system.
> ---
>
> Key: AIRFLOW-3447
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3447
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: core
>Reporter: Luka Draksler
>Assignee: Kaxil Naik
>Priority: Minor
>  Labels: easyfix
> Fix For: 1.10.2
>
>
> Migration to timezone aware times broke the intended usage of ts_nodash macro.
> ts_nodash is used in certain placeholders to create different names (table 
> names, cluster names...). As such it is alphanumeric only, it contains no 
> characters that could be deemed illegal by various naming restrictions. 
> Migration to new time system changed that.
> As an example, this would be returned currently: 
> {{20181205T125657.169324+}}
> {{before:}}
> {{20181204T03}}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3521) Airflow Jira Compare script is limited to 50 items

2018-12-15 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3521?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3521.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

> Airflow Jira Compare script is limited to 50 items
> --
>
> Key: AIRFLOW-3521
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3521
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Kaxil Naik
>Priority: Minor
> Fix For: 1.10.2
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (AIRFLOW-3521) Airflow Jira Compare script is limited to 50 items

2018-12-15 Thread Kaxil Naik (JIRA)
Kaxil Naik created AIRFLOW-3521:
---

 Summary: Airflow Jira Compare script is limited to 50 items
 Key: AIRFLOW-3521
 URL: https://issues.apache.org/jira/browse/AIRFLOW-3521
 Project: Apache Airflow
  Issue Type: Improvement
Reporter: Kaxil Naik






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-1919) Add option to query for DAG runs given a DAG ID

2018-12-15 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-1919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-1919.
-
   Resolution: Fixed
Fix Version/s: (was: 2.0.0)
   1.10.2

> Add option to query for DAG runs given a DAG ID
> ---
>
> Key: AIRFLOW-1919
> URL: https://issues.apache.org/jira/browse/AIRFLOW-1919
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: cli
>Affects Versions: 1.8.0
>Reporter: Steen Manniche
>Assignee: Tao Feng
>Priority: Trivial
> Fix For: 1.10.2
>
>
> Having a way to list all DAG runs for a given DAG identifier would be useful 
> when trying to get a programmatic overview of running DAGs. Something along 
> the lines of
> {code}
> airflow list_runs $DAG_ID
> {code}
> Which would return the running DAGs for {{$DAG_ID}}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Reopened] (AIRFLOW-1919) Add option to query for DAG runs given a DAG ID

2018-12-15 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-1919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik reopened AIRFLOW-1919:
-

> Add option to query for DAG runs given a DAG ID
> ---
>
> Key: AIRFLOW-1919
> URL: https://issues.apache.org/jira/browse/AIRFLOW-1919
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: cli
>Affects Versions: 1.8.0
>Reporter: Steen Manniche
>Assignee: Tao Feng
>Priority: Trivial
> Fix For: 1.10.2
>
>
> Having a way to list all DAG runs for a given DAG identifier would be useful 
> when trying to get a programmatic overview of running DAGs. Something along 
> the lines of
> {code}
> airflow list_runs $DAG_ID
> {code}
> Which would return the running DAGs for {{$DAG_ID}}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-1919) Add option to query for DAG runs given a DAG ID

2018-12-15 Thread Kaxil Naik (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-1919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16722114#comment-16722114
 ] 

Kaxil Naik commented on AIRFLOW-1919:
-

This feature will be available in 1.10.2. Currently it is only available in the 
master

> Add option to query for DAG runs given a DAG ID
> ---
>
> Key: AIRFLOW-1919
> URL: https://issues.apache.org/jira/browse/AIRFLOW-1919
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: cli
>Affects Versions: 1.8.0
>Reporter: Steen Manniche
>Assignee: Tao Feng
>Priority: Trivial
> Fix For: 1.10.2
>
>
> Having a way to list all DAG runs for a given DAG identifier would be useful 
> when trying to get a programmatic overview of running DAGs. Something along 
> the lines of
> {code}
> airflow list_runs $DAG_ID
> {code}
> Which would return the running DAGs for {{$DAG_ID}}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-1919) Add option to query for DAG runs given a DAG ID

2018-12-15 Thread Kaxil Naik (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-1919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16722113#comment-16722113
 ] 

Kaxil Naik commented on AIRFLOW-1919:
-

[~villasv]: This is already available in the documentation at 
https://airflow.readthedocs.io/en/latest/cli.html#list_dag_runs

> Add option to query for DAG runs given a DAG ID
> ---
>
> Key: AIRFLOW-1919
> URL: https://issues.apache.org/jira/browse/AIRFLOW-1919
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: cli
>Affects Versions: 1.8.0
>Reporter: Steen Manniche
>Assignee: Tao Feng
>Priority: Trivial
> Fix For: 2.0.0
>
>
> Having a way to list all DAG runs for a given DAG identifier would be useful 
> when trying to get a programmatic overview of running DAGs. Something along 
> the lines of
> {code}
> airflow list_runs $DAG_ID
> {code}
> Which would return the running DAGs for {{$DAG_ID}}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3444) Expand templated fields in gcs transfer service operator

2018-12-14 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3444?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3444.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

Resolved by https://github.com/apache/incubator-airflow/pull/4279

> Expand templated fields in gcs transfer service operator
> 
>
> Key: AIRFLOW-3444
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3444
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Josh Carp
>Assignee: Josh Carp
>Priority: Trivial
> Fix For: 1.10.2
>
>
> The `S3ToGoogleCloudStorageTransferOperator` should support an explicit 
> `description` parameter and allow that parameter to be templated. This will 
> make it easier to set job descriptions, and we'll be able to drop the 
> `job_kwargs` attribute, which doesn't have any other use.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3411) create openfaas hook

2018-12-12 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3411?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3411.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

> create openfaas hook
> 
>
> Key: AIRFLOW-3411
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3411
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: tal
>Assignee: tal
>Priority: Minor
> Fix For: 1.10.2
>
>
> add openfaas hook to trigger functions
> (https://github.com/openfaas/faas)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3310) Google Cloud Spanner deploy / delete operators

2018-12-12 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3310?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3310.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

> Google Cloud Spanner deploy / delete operators
> --
>
> Key: AIRFLOW-3310
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3310
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: gcp
>Reporter: Szymon Przedwojski
>Assignee: Szymon Przedwojski
>Priority: Minor
> Fix For: 1.10.2
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Reopened] (AIRFLOW-2770) kubernetes: add support for dag folder in the docker image

2018-12-12 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2770?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik reopened AIRFLOW-2770:
-

> kubernetes: add support for dag folder in the docker image
> --
>
> Key: AIRFLOW-2770
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2770
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Rurui Ye
>Assignee: Rurui Ye
>Priority: Critical
> Fix For: 1.10.2
>
>
> currently the kube executor need to provider dag_volume_chain or git repo in 
> the config file, but if the user has build dag into their docker image, they 
> doesn't need to provider these two options, and they can manager their dag 
> version by manager the docker image version. 
> So I suppose we can add the a new configuration as 
> kube.config.dag_folder_path along with dag_volume_chain and git repo. with 
> this config, we can run the worker just from the dags in docker image.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-2770) kubernetes: add support for dag folder in the docker image

2018-12-12 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2770?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-2770.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

> kubernetes: add support for dag folder in the docker image
> --
>
> Key: AIRFLOW-2770
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2770
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Rurui Ye
>Assignee: Rurui Ye
>Priority: Critical
> Fix For: 1.10.2
>
>
> currently the kube executor need to provider dag_volume_chain or git repo in 
> the config file, but if the user has build dag into their docker image, they 
> doesn't need to provider these two options, and they can manager their dag 
> version by manager the docker image version. 
> So I suppose we can add the a new configuration as 
> kube.config.dag_folder_path along with dag_volume_chain and git repo. with 
> this config, we can run the worker just from the dags in docker image.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3327) BiqQuery job checking doesn't include location, which api requires outside US/EU

2018-12-12 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3327:

Fix Version/s: 1.10.2

> BiqQuery job checking doesn't include location, which api requires outside 
> US/EU
> 
>
> Key: AIRFLOW-3327
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3327
> Project: Apache Airflow
>  Issue Type: Bug
>Reporter: Daniel Swiegers
>Assignee: Kaxil Naik
>Priority: Minor
>  Labels: google-cloud-bigquery
> Fix For: 1.10.2
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> We use this api but don't set / pass through the geographical location.
> Which is required in areas other than US and EU.
> Can be seen in contrib/hooks/big_query_hook.py poll_job_complete
> [https://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/get]
> |The geographic location of the job. Required except for US and EU. See 
> details at 
> https://cloud.google.com/bigquery/docs/locations#specifying_your_location.|



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-3327) BiqQuery job checking doesn't include location, which api requires outside US/EU

2018-12-12 Thread Kaxil Naik (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3327?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16718740#comment-16718740
 ] 

Kaxil Naik commented on AIRFLOW-3327:
-

I have assigned this to myself [~dsdinter] :-) Will work on it and should 
resolve it by this weekend

> BiqQuery job checking doesn't include location, which api requires outside 
> US/EU
> 
>
> Key: AIRFLOW-3327
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3327
> Project: Apache Airflow
>  Issue Type: Bug
>Reporter: Daniel Swiegers
>Assignee: Kaxil Naik
>Priority: Minor
>  Labels: google-cloud-bigquery
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> We use this api but don't set / pass through the geographical location.
> Which is required in areas other than US and EU.
> Can be seen in contrib/hooks/big_query_hook.py poll_job_complete
> [https://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/get]
> |The geographic location of the job. Required except for US and EU. See 
> details at 
> https://cloud.google.com/bigquery/docs/locations#specifying_your_location.|



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (AIRFLOW-3327) BiqQuery job checking doesn't include location, which api requires outside US/EU

2018-12-12 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik reassigned AIRFLOW-3327:
---

Assignee: Kaxil Naik

> BiqQuery job checking doesn't include location, which api requires outside 
> US/EU
> 
>
> Key: AIRFLOW-3327
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3327
> Project: Apache Airflow
>  Issue Type: Bug
>Reporter: Daniel Swiegers
>Assignee: Kaxil Naik
>Priority: Minor
>  Labels: google-cloud-bigquery
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> We use this api but don't set / pass through the geographical location.
> Which is required in areas other than US and EU.
> Can be seen in contrib/hooks/big_query_hook.py poll_job_complete
> [https://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/get]
> |The geographic location of the job. Required except for US and EU. See 
> details at 
> https://cloud.google.com/bigquery/docs/locations#specifying_your_location.|



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2785) Improve context management of mongo hook

2018-12-09 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2785?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2785:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Improve context management of mongo hook
> 
>
> Key: AIRFLOW-2785
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2785
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Andy Cooper
>Assignee: Andy Cooper
>Priority: Minor
> Fix For: 1.10.2
>
>
> Add __enter__ and __exit__ methods to improve context management of mongohook



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3484) The KubernetesExecutor is overly verbose and this can kill schedulers.

2018-12-09 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3484:

Fix Version/s: (was: 2.0.0)
   1.10.2

> The KubernetesExecutor is overly verbose and this can kill schedulers.
> --
>
> Key: AIRFLOW-3484
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3484
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: kubernetes
>Affects Versions: 1.10.1
>Reporter: Daniel Imberman
>Assignee: Daniel Imberman
>Priority: Major
> Fix For: 1.10.2
>
>
> There are two log lines in the k8sexecutor that can cause schedulers to crash 
> just by their sheer verbosity. This PR will switch these lines to debug as to 
> not mess with normal workflows.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3438) BigQueryOperator should default udf_config to None instead of false.

2018-12-09 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3438?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3438.
-
Resolution: Fixed

Resolved by https://github.com/apache/incubator-airflow/pull/4274

> BigQueryOperator should default udf_config to None instead of false.
> 
>
> Key: AIRFLOW-3438
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3438
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: operators
>Reporter: Luka Draksler
>Assignee: Kaxil Naik
>Priority: Major
>  Labels: easyfix
> Fix For: 1.10.2
>
>
> BigQueryOperator currently sets default value of udf_config to False. This no 
> longer works due to [https://github.com/apache/incubator-airflow/pull/3733] 
> validating the type of that parameter as either None or list. Default value 
> needs to be changed to None.
> The line in question added in the commit referenced above
> {code:java}
> (udf_config, 'userDefinedFunctionResources', None, list),
> {code}
>  
>  
> Note, other users of the hook may potentially encounter the same issue.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-3479) Records of "Log" table should be kept (by default) when users delete a DAG.

2018-12-07 Thread Kaxil Naik (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3479?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16713458#comment-16713458
 ] 

Kaxil Naik commented on AIRFLOW-3479:
-

Resolved by https://github.com/apache/incubator-airflow/pull/4287

> Records of "Log" table should be kept (by default) when users delete a DAG.
> ---
>
> Key: AIRFLOW-3479
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3479
> Project: Apache Airflow
>  Issue Type: Improvement
>Affects Versions: 1.10.1
>Reporter: Xiaodong DENG
>Assignee: Xiaodong DENG
>Priority: Minor
> Fix For: 1.10.2
>
>
> Currently when we delete a DAG (using API or from the UI), it will delete all 
> related records in all tables (all tables in which "dag_id" is available), 
> including "log" table.
> However, the records in "log" table should be kept (by default). This would 
> be ideal for multiple reasons, like auditing.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3479) Records of "Log" table should be kept (by default) when users delete a DAG.

2018-12-07 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3479?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3479.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

> Records of "Log" table should be kept (by default) when users delete a DAG.
> ---
>
> Key: AIRFLOW-3479
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3479
> Project: Apache Airflow
>  Issue Type: Improvement
>Affects Versions: 1.10.1
>Reporter: Xiaodong DENG
>Assignee: Xiaodong DENG
>Priority: Minor
> Fix For: 1.10.2
>
>
> Currently when we delete a DAG (using API or from the UI), it will delete all 
> related records in all tables (all tables in which "dag_id" is available), 
> including "log" table.
> However, the records in "log" table should be kept (by default). This would 
> be ideal for multiple reasons, like auditing.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3484) The KubernetesExecutor is overly verbose and this can kill schedulers.

2018-12-07 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3484.
-
   Resolution: Fixed
Fix Version/s: 2.0.0

Resolved by https://github.com/apache/incubator-airflow/pull/4296

> The KubernetesExecutor is overly verbose and this can kill schedulers.
> --
>
> Key: AIRFLOW-3484
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3484
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: kubernetes
>Affects Versions: 1.10.1
>Reporter: Daniel Imberman
>Assignee: Daniel Imberman
>Priority: Major
> Fix For: 2.0.0
>
>
> There are two log lines in the k8sexecutor that can cause schedulers to crash 
> just by their sheer verbosity. This PR will switch these lines to debug as to 
> not mess with normal workflows.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3408) Systemd setup instructions mention deprecated variable

2018-12-06 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3408?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3408.
-
Resolution: Fixed

> Systemd setup instructions mention deprecated variable
> --
>
> Key: AIRFLOW-3408
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3408
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Victor Villas Bôas Chaves
>Assignee: Kaxil Naik
>Priority: Minor
> Fix For: 1.10.2
>
>
> AIRFLOW-1698 was solved in code, but the documentation drifted.
> Places where {{SCHEDULER_RUNS}} is still mentioned as necessary adjustment:
> [https://github.com/apache/incubator-airflow/blob/53b89b98371c7bb993b242c341d3941e9ce09f9a/scripts/systemd/README]
> [https://github.com/apache/incubator-airflow/blob/b9fc03ea1ad5cea3c3aa668fcaca103f84167b9c/docs/howto/run-with-systemd.rst]
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3406) Implement an Azure CosmosDB operator

2018-12-06 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3406?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3406.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

Resolved by https://github.com/apache/incubator-airflow/pull/4265

> Implement an Azure CosmosDB operator 
> -
>
> Key: AIRFLOW-3406
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3406
> Project: Apache Airflow
>  Issue Type: New Feature
>Reporter: Tom Miller
>Assignee: Tom Miller
>Priority: Major
> Fix For: 1.10.2
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (AIRFLOW-3447) Intended usage of ts_nodash macro broken with migration to new time system.

2018-12-06 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik reassigned AIRFLOW-3447:
---

Assignee: Kaxil Naik

> Intended usage of ts_nodash macro broken with migration to new time system.
> ---
>
> Key: AIRFLOW-3447
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3447
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: core
>Reporter: Luka Draksler
>Assignee: Kaxil Naik
>Priority: Minor
>  Labels: easyfix
> Fix For: 1.10.2
>
>
> Migration to timezone aware times broke the intended usage of ts_nodash macro.
> ts_nodash is used in certain placeholders to create different names (table 
> names, cluster names...). As such it is alphanumeric only, it contains no 
> characters that could be deemed illegal by various naming restrictions. 
> Migration to new time system changed that.
> As an example, this would be returned currently: 
> {{20181205T125657.169324+}}
> {{before:}}
> {{20181204T03}}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3447) Intended usage of ts_nodash macro broken with migration to new time system.

2018-12-05 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3447:

Fix Version/s: 1.10.2

> Intended usage of ts_nodash macro broken with migration to new time system.
> ---
>
> Key: AIRFLOW-3447
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3447
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: core
>Reporter: Luka Draksler
>Priority: Minor
>  Labels: easyfix
> Fix For: 1.10.2
>
>
> Migration to timezone aware times broke the intended usage of ts_nodash macro.
> ts_nodash is used in certain placeholders to create different names (table 
> names, cluster names...). As such it is alphanumeric only, it contains no 
> characters that could be deemed illegal by various naming restrictions. 
> Migration to new time system changed that.
> As an example, this would be returned currently: 
> {{20181205T125657.169324+}}
> {{before:}}
> {{20181204T03}}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-3447) Intended usage of ts_nodash macro broken with migration to new time system.

2018-12-05 Thread Kaxil Naik (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3447?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16710606#comment-16710606
 ] 

Kaxil Naik commented on AIRFLOW-3447:
-

[~ashb] Timezone is giving few headaches isn't it. For 1.10.2 can we just 
resolve this with this solution given above? If there are many issues.

> Intended usage of ts_nodash macro broken with migration to new time system.
> ---
>
> Key: AIRFLOW-3447
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3447
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: core
>Reporter: Luka Draksler
>Priority: Minor
>  Labels: easyfix
> Fix For: 1.10.2
>
>
> Migration to timezone aware times broke the intended usage of ts_nodash macro.
> ts_nodash is used in certain placeholders to create different names (table 
> names, cluster names...). As such it is alphanumeric only, it contains no 
> characters that could be deemed illegal by various naming restrictions. 
> Migration to new time system changed that.
> As an example, this would be returned currently: 
> {{20181205T125657.169324+}}
> {{before:}}
> {{20181204T03}}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2440) Google Cloud SQL import/export operator

2018-12-05 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2440:

Priority: Trivial  (was: Major)

> Google Cloud SQL import/export operator
> ---
>
> Key: AIRFLOW-2440
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2440
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: contrib, gcp
>Reporter: Guillermo Rodríguez Cano
>Assignee: Szymon Przedwojski
>Priority: Trivial
> Fix For: 1.10.2
>
>
> Implement Google Cloud SQL Operator for import/export operations as a start.
> Such operator implies the corresponding hook, which can be inspired by the 
> Google Dataflow hook for the management of the import/export jobs (checking 
> the status of an issued job).
> API for [mysql|https://cloud.google.com/sql/docs/mysql/admin-api/v1beta4/] 
> and [postgres|https://cloud.google.com/sql/docs/postgres/admin-api/v1beta4/] 
> (they seem alike for these two operations at least)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2440) Google Cloud SQL import/export operator

2018-12-05 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2440:

Affects Version/s: (was: 2.0.0)

> Google Cloud SQL import/export operator
> ---
>
> Key: AIRFLOW-2440
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2440
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: contrib, gcp
>Reporter: Guillermo Rodríguez Cano
>Assignee: Szymon Przedwojski
>Priority: Trivial
> Fix For: 1.10.2
>
>
> Implement Google Cloud SQL Operator for import/export operations as a start.
> Such operator implies the corresponding hook, which can be inspired by the 
> Google Dataflow hook for the management of the import/export jobs (checking 
> the status of an issued job).
> API for [mysql|https://cloud.google.com/sql/docs/mysql/admin-api/v1beta4/] 
> and [postgres|https://cloud.google.com/sql/docs/postgres/admin-api/v1beta4/] 
> (they seem alike for these two operations at least)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-2440) Google Cloud SQL import/export operator

2018-12-05 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-2440.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

> Google Cloud SQL import/export operator
> ---
>
> Key: AIRFLOW-2440
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2440
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: contrib, gcp
>Reporter: Guillermo Rodríguez Cano
>Assignee: Szymon Przedwojski
>Priority: Major
> Fix For: 1.10.2
>
>
> Implement Google Cloud SQL Operator for import/export operations as a start.
> Such operator implies the corresponding hook, which can be inspired by the 
> Google Dataflow hook for the management of the import/export jobs (checking 
> the status of an issued job).
> API for [mysql|https://cloud.google.com/sql/docs/mysql/admin-api/v1beta4/] 
> and [postgres|https://cloud.google.com/sql/docs/postgres/admin-api/v1beta4/] 
> (they seem alike for these two operations at least)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-1552) Airflow Filter_by_owner not working with password_auth

2018-12-05 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-1552?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-1552:

Fix Version/s: 1.10.2

> Airflow Filter_by_owner not working with password_auth
> --
>
> Key: AIRFLOW-1552
> URL: https://issues.apache.org/jira/browse/AIRFLOW-1552
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: configuration
>Affects Versions: 1.8.0
> Environment: CentOS , python 2.7
>Reporter: raghu ram reddy
>Priority: Major
> Fix For: 1.10.2
>
>
> Airflow Filter_by_owner parameter is not working with password_auth.
> I created sample user using the below code from airflow documentation and 
> enabled password_auth.
> I'm able to login as the user created but by default this user is superuser 
> and there is noway to modify it, default all users created by PasswordUser 
> are superusers.
> import airflow
> from airflow import models, settings
> from airflow.contrib.auth.backends.password_auth import PasswordUser
> user = PasswordUser(models.User())
> user.username = 'test1'
> user.password = 'test1'
> user.is_superuser()
> session = settings.Session()
> session.add(user)
> session.commit()
> session.close()
> exit()



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2670) SSHOperator's timeout parameter doesn't affect SSHook timeoot

2018-12-04 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2670?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2670:

Fix Version/s: (was: 2.0.0)
   1.10.2

> SSHOperator's timeout parameter doesn't affect SSHook timeoot
> -
>
> Key: AIRFLOW-2670
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2670
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: contrib
>Affects Versions: 2.0.0
>Reporter: jin zhang
>Priority: Major
> Fix For: 1.10.2
>
>
> when I use SSHOperator, SSHOperator's timeout parameter can't set in SSHHook 
> and it's just effect exce_command. 
> old version:
> self.ssh_hook = SSHHook(ssh_conn_id=self.ssh_conn_id)
> I change it to :
> self.ssh_hook = SSHHook(ssh_conn_id=self.ssh_conn_id, timeout=self.timeout)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3309) Missing Mongo DB connection type

2018-12-04 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3309?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3309:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Missing Mongo DB connection type
> 
>
> Key: AIRFLOW-3309
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3309
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: database
>Affects Versions: 1.10.0
>Reporter: John Cheng
>Assignee: John Cheng
>Priority: Minor
> Fix For: 1.10.2
>
>
> Unable to choose Mongo DB on the admin console connection page.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Work started] (AIRFLOW-3438) BigQueryOperator should default udf_config to None instead of false.

2018-12-03 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3438?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on AIRFLOW-3438 started by Kaxil Naik.
---
> BigQueryOperator should default udf_config to None instead of false.
> 
>
> Key: AIRFLOW-3438
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3438
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: operators
>Reporter: Luka Draksler
>Assignee: Kaxil Naik
>Priority: Major
>  Labels: easyfix
> Fix For: 1.10.2
>
>
> BigQueryOperator currently sets default value of udf_config to False. This no 
> longer works due to [https://github.com/apache/incubator-airflow/pull/3733] 
> validating the type of that parameter as either None or list. Default value 
> needs to be changed to None.
> The line in question added in the commit referenced above
> {code:java}
> (udf_config, 'userDefinedFunctionResources', None, list),
> {code}
>  
>  
> Note, other users of the hook may potentially encounter the same issue.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3438) BigQueryOperator should default udf_config to None instead of false.

2018-12-03 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3438?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3438:

Fix Version/s: 1.10.2

> BigQueryOperator should default udf_config to None instead of false.
> 
>
> Key: AIRFLOW-3438
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3438
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: operators
>Reporter: Luka Draksler
>Assignee: Kaxil Naik
>Priority: Major
>  Labels: easyfix
> Fix For: 1.10.2
>
>
> BigQueryOperator currently sets default value of udf_config to False. This no 
> longer works due to [https://github.com/apache/incubator-airflow/pull/3733] 
> validating the type of that parameter as either None or list. Default value 
> needs to be changed to None.
> The line in question added in the commit referenced above
> {code:java}
> (udf_config, 'userDefinedFunctionResources', None, list),
> {code}
>  
>  
> Note, other users of the hook may potentially encounter the same issue.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3434) SFTPOperator does not create intermediate directories

2018-12-03 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3434?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3434.
-
Resolution: Fixed

Resolved by https://github.com/apache/incubator-airflow/pull/4270

> SFTPOperator does not create intermediate directories
> -
>
> Key: AIRFLOW-3434
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3434
> Project: Apache Airflow
>  Issue Type: New Feature
>Reporter: Kaxil Naik
>Assignee: Kaxil Naik
>Priority: Minor
> Fix For: 1.10.2
>
>
> When using SFTPOperator with either 'get' or 'put', it doesn't create the 
> intermediate directories when copying the file and fails with directory does 
> not exist.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (AIRFLOW-3434) SFTPOperator does not create intermediate directories

2018-12-02 Thread Kaxil Naik (JIRA)
Kaxil Naik created AIRFLOW-3434:
---

 Summary: SFTPOperator does not create intermediate directories
 Key: AIRFLOW-3434
 URL: https://issues.apache.org/jira/browse/AIRFLOW-3434
 Project: Apache Airflow
  Issue Type: New Feature
Reporter: Kaxil Naik
Assignee: Kaxil Naik
 Fix For: 1.10.2


When using SFTPOperator with either 'get' or 'put', it doesn't create the 
intermediate directories when copying the file and fails with directory does 
not exist.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3414) reload_module not working with custom logging class

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3414?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3414:

Affects Version/s: (was: 1.10.2)
   1.10.1

> reload_module not working with custom logging class
> ---
>
> Key: AIRFLOW-3414
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3414
> Project: Apache Airflow
>  Issue Type: Bug
>Affects Versions: 1.10.1
>Reporter: Kevin Yang
>Assignee: Kevin Yang
>Priority: Major
> Fix For: 1.10.2
>
>
> If using custom logging class, the reload_module in dag_processing.py will 
> fail because it will try to reload default logging class, which is not loaded 
> at the first place.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3380) Metrics documentation

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3380?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3380.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

Resolved by https://github.com/apache/incubator-airflow/pull/4219

> Metrics documentation
> -
>
> Key: AIRFLOW-3380
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3380
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Bartosz Ługowski
>Assignee: Bartosz Ługowski
>Priority: Trivial
> Fix For: 1.10.2
>
>
> At the moment there is no documentation about Airflow metrics:
>  * how to enable it
>  * which metrics are generated
>  * description of each metric



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3414) reload_module not working with custom logging class

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3414?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3414:

Fix Version/s: 1.10.2

> reload_module not working with custom logging class
> ---
>
> Key: AIRFLOW-3414
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3414
> Project: Apache Airflow
>  Issue Type: Bug
>Affects Versions: 1.10.1
>Reporter: Kevin Yang
>Assignee: Kevin Yang
>Priority: Major
> Fix For: 1.10.2
>
>
> If using custom logging class, the reload_module in dag_processing.py will 
> fail because it will try to reload default logging class, which is not loaded 
> at the first place.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3408) Systemd setup instructions mention deprecated variable

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3408?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3408:

Fix Version/s: 1.10.2

> Systemd setup instructions mention deprecated variable
> --
>
> Key: AIRFLOW-3408
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3408
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Victor Villas Bôas Chaves
>Assignee: Kaxil Naik
>Priority: Minor
> Fix For: 1.10.2
>
>
> AIRFLOW-1698 was solved in code, but the documentation drifted.
> Places where {{SCHEDULER_RUNS}} is still mentioned as necessary adjustment:
> [https://github.com/apache/incubator-airflow/blob/53b89b98371c7bb993b242c341d3941e9ce09f9a/scripts/systemd/README]
> [https://github.com/apache/incubator-airflow/blob/b9fc03ea1ad5cea3c3aa668fcaca103f84167b9c/docs/howto/run-with-systemd.rst]
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Work started] (AIRFLOW-3408) Systemd setup instructions mention deprecated variable

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3408?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on AIRFLOW-3408 started by Kaxil Naik.
---
> Systemd setup instructions mention deprecated variable
> --
>
> Key: AIRFLOW-3408
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3408
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Victor Villas Bôas Chaves
>Assignee: Kaxil Naik
>Priority: Minor
>
> AIRFLOW-1698 was solved in code, but the documentation drifted.
> Places where {{SCHEDULER_RUNS}} is still mentioned as necessary adjustment:
> [https://github.com/apache/incubator-airflow/blob/53b89b98371c7bb993b242c341d3941e9ce09f9a/scripts/systemd/README]
> [https://github.com/apache/incubator-airflow/blob/b9fc03ea1ad5cea3c3aa668fcaca103f84167b9c/docs/howto/run-with-systemd.rst]
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (AIRFLOW-3408) Systemd setup instructions mention deprecated variable

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3408?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik reassigned AIRFLOW-3408:
---

Assignee: Kaxil Naik  (was: Anoop Kunjuraman)

> Systemd setup instructions mention deprecated variable
> --
>
> Key: AIRFLOW-3408
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3408
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Victor Villas Bôas Chaves
>Assignee: Kaxil Naik
>Priority: Minor
>
> AIRFLOW-1698 was solved in code, but the documentation drifted.
> Places where {{SCHEDULER_RUNS}} is still mentioned as necessary adjustment:
> [https://github.com/apache/incubator-airflow/blob/53b89b98371c7bb993b242c341d3941e9ce09f9a/scripts/systemd/README]
> [https://github.com/apache/incubator-airflow/blob/b9fc03ea1ad5cea3c3aa668fcaca103f84167b9c/docs/howto/run-with-systemd.rst]
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3432) Test is missing for feature "Delete DAG in UI"

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3432?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3432.
-
Resolution: Fixed

Resolved by https://github.com/apache/incubator-airflow/pull/4266

> Test is missing for feature "Delete DAG in UI"
> --
>
> Key: AIRFLOW-3432
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3432
> Project: Apache Airflow
>  Issue Type: Test
>  Components: tests
>Affects Versions: 1.10.1
>Reporter: Xiaodong DENG
>Assignee: Xiaodong DENG
>Priority: Minor
> Fix For: 1.10.2
>
>
> In [https://github.com/apache/incubator-airflow/pull/4069,] "delete DAG in 
> UI" feature is fixed, but no test was added for it.
>  
> We should have test for it to release it with more confidence.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3266) AWS Athena Operator in Airflow

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3266?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3266:

Fix Version/s: (was: 2.0.0)
   1.10.2

> AWS Athena Operator in Airflow
> --
>
> Key: AIRFLOW-3266
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3266
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: aws
>Affects Versions: 1.10.0
>Reporter: Sai Phanindhra
>Assignee: Sai Phanindhra
>Priority: Minor
> Fix For: 1.10.2
>
>
> There is no official athena operator as of now airflow. Either one has do it 
> using boto3 in python operator or using aws cli in bash operator. Either of 
> these does not take care of total life cycle of query. Create a Athena 
> operator and hook to submit presto query and update task based of state of 
> submitted query.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3022) Add volume mount to KubernetesExecutorConfig

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3022?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3022.
-
Resolution: Fixed

Resolved by https://github.com/apache/incubator-airflow/pull/3855

> Add volume mount to KubernetesExecutorConfig
> 
>
> Key: AIRFLOW-3022
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3022
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: executor
>Affects Versions: 1.10.0
>Reporter: John Cheng
>Assignee: John Cheng
>Priority: Minor
> Fix For: 1.10.2
>
>
> Allows volume mount to the worker pod with executor_config.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3361) Add the task_id to the Deprecation Warning when passing unsupported keywords to BaseOperator

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3361?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3361:

Fix Version/s: (was: 2.0.0)

> Add the task_id to the Deprecation Warning when passing unsupported keywords 
> to BaseOperator
> 
>
> Key: AIRFLOW-3361
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3361
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: logging
>Affects Versions: 1.9.0
>Reporter: Martin Black
>Assignee: Martin Black
>Priority: Trivial
> Fix For: 1.10.2
>
>
> In 2.0 passing invalid keywords to {{BaseOperator}} will be deprecated. Prior 
> to that, there is a {{PendingDeprecationWarning}} raised, however it can be 
> hard to track down which specific task is raising this warning.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3425) GoogleCloudBaseHook doesn't set default scope when google_cloud_default connection is empty

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3425?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3425:

Fix Version/s: (was: 2.0.0)
   1.10.2

> GoogleCloudBaseHook doesn't set default scope when google_cloud_default 
> connection is empty
> ---
>
> Key: AIRFLOW-3425
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3425
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: gcp, hooks
>Affects Versions: 1.10.1
>Reporter: Alexander Sokolov
>Priority: Minor
> Fix For: 1.10.2
>
>
> When using Application Default Credentials to authenticating to GCP.
> _If the environment variable {{GOOGLE_APPLICATION_CREDENTIALS}} is set to the 
> path of a valid service account JSON private key file, then it is loaded and 
> returned. The project ID returned is the project ID defined in the service 
> account file if available (some older files do not contain project ID 
> information)._
> {code:java}
> [2018-11-30 11:18:35,038] {models.py:1569} INFO - Executing 
>  on 
> 2018-11-29T17:54:03.110575+00:00
> [2018-11-30 11:18:35,039] {base_task_runner.py:124} INFO - Running: ['bash', 
> '-c', 'airflow run transform_raw_data upload_transformed_data_to_gcs 
> 2018-11-29T17:54:03.110575+00:00 --job_id 98 --raw -sd 
> DAGS_FOLDER/transform_raw_data.py --cfg_path 
> /var/folders/5j/n3y23gt93l19lmt1tf4mgpr8gn/T/tmprbpmva2a']
> [2018-11-30 11:18:35,606] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:35,606] {settings.py:174} 
> INFO - setting.configure_orm(): Using pool settings. pool_size=5, 
> pool_recycle=1800
> [2018-11-30 11:18:35,981] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:35,979] {_init_.py:51} INFO 
> - Using executor LocalExecutor
> [2018-11-30 11:18:36,202] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:36,201] {models.py:258} INFO 
> - Filling up the DagBag from /Users/alex/airflow/dags/transform_raw_data.py
> [2018-11-30 11:18:38,391] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:38,390] {base_hook.py:83} 
> INFO - Using connection to: localhost
> [2018-11-30 11:18:43,427] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:43,426] {cli.py:492} INFO - 
> Running  2018-11-29T17:54:03.110575+00:00 [running]> on host laptop.local
> [2018-11-30 11:18:48,574] {logging_mixin.py:95} INFO - [2018-11-30 
> 11:18:48,574] {gcp_api_base_hook.py:86} INFO - Getting connection using 
> `google.auth.default()` since no key file is defined for hook.
> [2018-11-30 11:18:48,583] {logging_mixin.py:95} INFO - [2018-11-30 
> 11:18:48,583] {discovery.py:267} INFO - URL being requested: GET 
> https://www.googleapis.com/discovery/v1/apis/storage/v1/rest
> [2018-11-30 11:18:48,852] {models.py:1736} ERROR - ('invalid_scope: Empty or 
> missing scope not allowed.', '{\n  "error": "invalid_scope",\n  
> "error_description": "Empty or missing scope not allowed."\n}')
> Traceback (most recent call last):
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/models.py",
>  line 1633, in _run_raw_task
>     result = task_copy.execute(context=context)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/contrib/operators/file_to_gcs.py",
>  line 75, in execute
>     filename=self.src)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/contrib/hooks/gcs_hook.py",
>  line 187, in upload
>     service = self.get_conn()
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/contrib/hooks/gcs_hook.py",
>  line 48, in get_conn
>     'storage', 'v1', http=http_authorized, cache_discovery=False)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/googleapiclient/_helpers.py",
>  line 130, in positional_wrapper
>     return wrapped(*args, **kwargs)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/googleapiclient/discovery.py",
>  line 222, in build
>     requested_url, discovery_http, cache_discovery, cache)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/googleapiclient/discovery.py",
>  line 269, in _retrieve_discovery_doc
>     resp, content = http.request(actual_url)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/google_auth_httplib2.py",
>  line 187, in request
>     self._request, method, uri, 

[jira] [Updated] (AIRFLOW-3345) Add Google Cloud Storage (GCS) operators for ACL

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3345?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3345:

Fix Version/s: (was: 2.0.0)

> Add Google Cloud Storage (GCS) operators for ACL
> 
>
> Key: AIRFLOW-3345
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3345
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: gcp
>Reporter: Szymon Przedwojski
>Assignee: Szymon Przedwojski
>Priority: Minor
> Fix For: 1.10.2
>
>
> Add operators to be able to add ACL entries to GCS buckets and objects:
>  * 
> [https://cloud.google.com/storage/docs/json_api/v1/objectAccessControls/insert]
>  * 
> [https://cloud.google.com/storage/docs/json_api/v1/bucketAccessControls/insert]
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3213) Create ADLS to GCS operator

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3213?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3213:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Create ADLS to GCS operator 
> 
>
> Key: AIRFLOW-3213
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3213
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: gcp, operators
>Reporter: Brandon Kvarda
>Assignee: Brandon Kvarda
>Priority: Minor
> Fix For: 1.10.2
>
>
> Create ADLS to GCS operator that supports copying of files from ADLS to GCS



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3384) Allow higher versions of sqlalchemy and jinja

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3384:

Fix Version/s: (was: 2.0.0)

> Allow higher versions of sqlalchemy and jinja
> -
>
> Key: AIRFLOW-3384
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3384
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: dependencies
>Reporter: Jose Luis Ricon
>Assignee: Jose Luis Ricon
>Priority: Major
> Fix For: 1.10.2
>
>
> At the moment airflow doesn't allow the installation of sqlalchemy version 
> 1.2.11 and jinja2==2.10 . Airflow works with both, and there is no reason to 
> allow higher versions. Projects downstream who are currently forcing the 
> installation of said versions, overriding airflow's dependencies, will 
> benefit for this as it will allow for version-compatible installations 
> without loss in functionality.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-2983) Add prev_ds_nodash and next_ds_nodash macro

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2983?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-2983.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

> Add prev_ds_nodash and next_ds_nodash macro 
> 
>
> Key: AIRFLOW-2983
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2983
> Project: Apache Airflow
>  Issue Type: Bug
>Reporter: Tao Feng
>Assignee: Tao Feng
>Priority: Minor
> Fix For: 1.10.2
>
>
> Similar to [#3418|https://github.com/apache/incubator-airflow/pull/3418], but 
> introduce two nodash macros:
> {\{ prev_ds_nodash }}: the previous execution date as \{{ MMDD }}
> {\{ next_ds_nodash }}: the next execution date as \{{ MMDD }}
> We introduced prev_ds and next_ds in 
> [#3418|https://github.com/apache/incubator-airflow/pull/3418] which is 
> beneficial for backfill. By using these two macros, we could override the 
> schedule_internal from daily to certain days(e.g 7 days) in backfill.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Reopened] (AIRFLOW-2983) Add prev_ds_nodash and next_ds_nodash macro

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2983?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik reopened AIRFLOW-2983:
-

> Add prev_ds_nodash and next_ds_nodash macro 
> 
>
> Key: AIRFLOW-2983
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2983
> Project: Apache Airflow
>  Issue Type: Bug
>Reporter: Tao Feng
>Assignee: Tao Feng
>Priority: Minor
> Fix For: 1.10.2
>
>
> Similar to [#3418|https://github.com/apache/incubator-airflow/pull/3418], but 
> introduce two nodash macros:
> {\{ prev_ds_nodash }}: the previous execution date as \{{ MMDD }}
> {\{ next_ds_nodash }}: the next execution date as \{{ MMDD }}
> We introduced prev_ds and next_ds in 
> [#3418|https://github.com/apache/incubator-airflow/pull/3418] which is 
> beneficial for backfill. By using these two macros, we could override the 
> schedule_internal from daily to certain days(e.g 7 days) in backfill.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3416) CloudSqlQueryOperator with sql proxy does not work with Python 3.x

2018-12-02 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3416?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3416.
-
Resolution: Fixed

Resolved by https://github.com/apache/incubator-airflow/pull/4254

> CloudSqlQueryOperator with sql proxy does not work with Python 3.x
> --
>
> Key: AIRFLOW-3416
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3416
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.1
>Reporter: Jarek Potiuk
>Assignee: Jarek Potiuk
>Priority: Major
> Fix For: 1.10.2
>
>
> There are compatibility issues with Python 3.x for CloudSQLoperator. Output 
> of cloud_sql_proxy binary is parsed and the output in Python3 is bytes rather 
> than string so several "in"s raise an exception in Python 3. It needs 
> explicit decode('utf-8')



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3396) Redundant sql param check in BigQueryOperator

2018-12-01 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3396?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3396.
-
Resolution: Fixed

Resolved by https://github.com/apache/incubator-airflow/pull/4224

> Redundant sql param check in BigQueryOperator
> -
>
> Key: AIRFLOW-3396
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3396
> Project: Apache Airflow
>  Issue Type: Task
>  Components: gcp
>Reporter: Kaxil Naik
>Assignee: Kaxil Naik
>Priority: Trivial
> Fix For: 2.0.0
>
>
> The 'sql' parameter in BigQueryOperator accepts None and then there is a 
> check which checks if it is None and gives error.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3403) Add AWS Athena Sensor

2018-12-01 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3403?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3403.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

Resolved by https://github.com/apache/incubator-airflow/pull/4244

> Add AWS Athena Sensor
> -
>
> Key: AIRFLOW-3403
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3403
> Project: Apache Airflow
>  Issue Type: New Feature
>Reporter: tal
>Assignee: tal
>Priority: Minor
> Fix For: 1.10.2
>
>
> It will be nice to have an Athena sensor to monitor the progress of the query



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3403) Add AWS Athena Sensor

2018-12-01 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3403?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3403:

Summary: Add AWS Athena Sensor  (was: create athena sensor)

> Add AWS Athena Sensor
> -
>
> Key: AIRFLOW-3403
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3403
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: tal
>Assignee: tal
>Priority: Minor
>
> It will be nice to have an Athena sensor to monitor the progress of the query



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3403) Add AWS Athena Sensor

2018-12-01 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3403?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3403:

Issue Type: New Feature  (was: Improvement)

> Add AWS Athena Sensor
> -
>
> Key: AIRFLOW-3403
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3403
> Project: Apache Airflow
>  Issue Type: New Feature
>Reporter: tal
>Assignee: tal
>Priority: Minor
>
> It will be nice to have an Athena sensor to monitor the progress of the query



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3323) Support Basic Authentication for Flower

2018-12-01 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3323:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Support Basic Authentication for Flower
> ---
>
> Key: AIRFLOW-3323
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3323
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: celery
>Affects Versions: 1.10.0, 1.10.1
>Reporter: Xiaodong DENG
>Assignee: Xiaodong DENG
>Priority: Critical
> Fix For: 1.10.2
>
>
> The current `airflow flower` doesn't come with any authentication. This may 
> make essential information exposed to in an untrusted environment.
> Currently Flower itself supports
>  * HTTP Basic Authentication
>  * Google OAuth 2.0
>  * GitHub OAuth
> Given Flower is not really the most essential component of Airflow, we don't 
> have to support all its authentication methods. But may be good to at least 
> support Basic Authentication.
>  
> This ticket adds support to Basic Authentication for `Airflow Flower`.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-3425) GoogleCloudBaseHook doesn't set default scope when google_cloud_default connection is empty

2018-12-01 Thread Kaxil Naik (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3425?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16705951#comment-16705951
 ] 

Kaxil Naik commented on AIRFLOW-3425:
-

Btw [~alsokoloff] it is not a bug, it already checked for `is not None`, and 
not `is None` as you mentioned.

However, I am merging this because it is a better convention.

> GoogleCloudBaseHook doesn't set default scope when google_cloud_default 
> connection is empty
> ---
>
> Key: AIRFLOW-3425
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3425
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: gcp, hooks
>Affects Versions: 1.10.1
>Reporter: Alexander Sokolov
>Priority: Minor
> Fix For: 2.0.0
>
>
> When using Application Default Credentials to authenticating to GCP.
> _If the environment variable {{GOOGLE_APPLICATION_CREDENTIALS}} is set to the 
> path of a valid service account JSON private key file, then it is loaded and 
> returned. The project ID returned is the project ID defined in the service 
> account file if available (some older files do not contain project ID 
> information)._
> {code:java}
> [2018-11-30 11:18:35,038] {models.py:1569} INFO - Executing 
>  on 
> 2018-11-29T17:54:03.110575+00:00
> [2018-11-30 11:18:35,039] {base_task_runner.py:124} INFO - Running: ['bash', 
> '-c', 'airflow run transform_raw_data upload_transformed_data_to_gcs 
> 2018-11-29T17:54:03.110575+00:00 --job_id 98 --raw -sd 
> DAGS_FOLDER/transform_raw_data.py --cfg_path 
> /var/folders/5j/n3y23gt93l19lmt1tf4mgpr8gn/T/tmprbpmva2a']
> [2018-11-30 11:18:35,606] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:35,606] {settings.py:174} 
> INFO - setting.configure_orm(): Using pool settings. pool_size=5, 
> pool_recycle=1800
> [2018-11-30 11:18:35,981] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:35,979] {_init_.py:51} INFO 
> - Using executor LocalExecutor
> [2018-11-30 11:18:36,202] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:36,201] {models.py:258} INFO 
> - Filling up the DagBag from /Users/alex/airflow/dags/transform_raw_data.py
> [2018-11-30 11:18:38,391] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:38,390] {base_hook.py:83} 
> INFO - Using connection to: localhost
> [2018-11-30 11:18:43,427] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:43,426] {cli.py:492} INFO - 
> Running  2018-11-29T17:54:03.110575+00:00 [running]> on host laptop.local
> [2018-11-30 11:18:48,574] {logging_mixin.py:95} INFO - [2018-11-30 
> 11:18:48,574] {gcp_api_base_hook.py:86} INFO - Getting connection using 
> `google.auth.default()` since no key file is defined for hook.
> [2018-11-30 11:18:48,583] {logging_mixin.py:95} INFO - [2018-11-30 
> 11:18:48,583] {discovery.py:267} INFO - URL being requested: GET 
> https://www.googleapis.com/discovery/v1/apis/storage/v1/rest
> [2018-11-30 11:18:48,852] {models.py:1736} ERROR - ('invalid_scope: Empty or 
> missing scope not allowed.', '{\n  "error": "invalid_scope",\n  
> "error_description": "Empty or missing scope not allowed."\n}')
> Traceback (most recent call last):
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/models.py",
>  line 1633, in _run_raw_task
>     result = task_copy.execute(context=context)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/contrib/operators/file_to_gcs.py",
>  line 75, in execute
>     filename=self.src)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/contrib/hooks/gcs_hook.py",
>  line 187, in upload
>     service = self.get_conn()
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/contrib/hooks/gcs_hook.py",
>  line 48, in get_conn
>     'storage', 'v1', http=http_authorized, cache_discovery=False)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/googleapiclient/_helpers.py",
>  line 130, in positional_wrapper
>     return wrapped(*args, **kwargs)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/googleapiclient/discovery.py",
>  line 222, in build
>     requested_url, discovery_http, cache_discovery, cache)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/googleapiclient/discovery.py",
>  line 269, in _retrieve_discovery_doc
>     resp, content = http.request(actual_url)
>   File 
> 

[jira] [Resolved] (AIRFLOW-3425) GoogleCloudBaseHook doesn't set default scope when google_cloud_default connection is empty

2018-12-01 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3425?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3425.
-
   Resolution: Fixed
Fix Version/s: 2.0.0

Resolved by https://github.com/apache/incubator-airflow/pull/4261

> GoogleCloudBaseHook doesn't set default scope when google_cloud_default 
> connection is empty
> ---
>
> Key: AIRFLOW-3425
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3425
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: gcp, hooks
>Affects Versions: 1.10.1
>Reporter: Alexander Sokolov
>Priority: Minor
> Fix For: 2.0.0
>
>
> When using Application Default Credentials to authenticating to GCP.
> _If the environment variable {{GOOGLE_APPLICATION_CREDENTIALS}} is set to the 
> path of a valid service account JSON private key file, then it is loaded and 
> returned. The project ID returned is the project ID defined in the service 
> account file if available (some older files do not contain project ID 
> information)._
> {code:java}
> [2018-11-30 11:18:35,038] {models.py:1569} INFO - Executing 
>  on 
> 2018-11-29T17:54:03.110575+00:00
> [2018-11-30 11:18:35,039] {base_task_runner.py:124} INFO - Running: ['bash', 
> '-c', 'airflow run transform_raw_data upload_transformed_data_to_gcs 
> 2018-11-29T17:54:03.110575+00:00 --job_id 98 --raw -sd 
> DAGS_FOLDER/transform_raw_data.py --cfg_path 
> /var/folders/5j/n3y23gt93l19lmt1tf4mgpr8gn/T/tmprbpmva2a']
> [2018-11-30 11:18:35,606] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:35,606] {settings.py:174} 
> INFO - setting.configure_orm(): Using pool settings. pool_size=5, 
> pool_recycle=1800
> [2018-11-30 11:18:35,981] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:35,979] {_init_.py:51} INFO 
> - Using executor LocalExecutor
> [2018-11-30 11:18:36,202] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:36,201] {models.py:258} INFO 
> - Filling up the DagBag from /Users/alex/airflow/dags/transform_raw_data.py
> [2018-11-30 11:18:38,391] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:38,390] {base_hook.py:83} 
> INFO - Using connection to: localhost
> [2018-11-30 11:18:43,427] {base_task_runner.py:107} INFO - Job 98: Subtask 
> upload_transformed_data_to_gcs [2018-11-30 11:18:43,426] {cli.py:492} INFO - 
> Running  2018-11-29T17:54:03.110575+00:00 [running]> on host laptop.local
> [2018-11-30 11:18:48,574] {logging_mixin.py:95} INFO - [2018-11-30 
> 11:18:48,574] {gcp_api_base_hook.py:86} INFO - Getting connection using 
> `google.auth.default()` since no key file is defined for hook.
> [2018-11-30 11:18:48,583] {logging_mixin.py:95} INFO - [2018-11-30 
> 11:18:48,583] {discovery.py:267} INFO - URL being requested: GET 
> https://www.googleapis.com/discovery/v1/apis/storage/v1/rest
> [2018-11-30 11:18:48,852] {models.py:1736} ERROR - ('invalid_scope: Empty or 
> missing scope not allowed.', '{\n  "error": "invalid_scope",\n  
> "error_description": "Empty or missing scope not allowed."\n}')
> Traceback (most recent call last):
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/models.py",
>  line 1633, in _run_raw_task
>     result = task_copy.execute(context=context)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/contrib/operators/file_to_gcs.py",
>  line 75, in execute
>     filename=self.src)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/contrib/hooks/gcs_hook.py",
>  line 187, in upload
>     service = self.get_conn()
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/airflow/contrib/hooks/gcs_hook.py",
>  line 48, in get_conn
>     'storage', 'v1', http=http_authorized, cache_discovery=False)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/googleapiclient/_helpers.py",
>  line 130, in positional_wrapper
>     return wrapped(*args, **kwargs)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/googleapiclient/discovery.py",
>  line 222, in build
>     requested_url, discovery_http, cache_discovery, cache)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/googleapiclient/discovery.py",
>  line 269, in _retrieve_discovery_doc
>     resp, content = http.request(actual_url)
>   File 
> "/Users/alex/.local/share/virtualenvs/project-X9xQGmcz/lib/python3.6/site-packages/google_auth_httplib2.py",
>  line 

[jira] [Updated] (AIRFLOW-3392) Add index on dag_id in sla_miss table

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3392?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3392:

Fix Version/s: (was: 1.10.2)
   2.0.0

> Add index on dag_id in sla_miss table
> -
>
> Key: AIRFLOW-3392
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3392
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Kevin Yang
>Assignee: Kevin Yang
>Priority: Major
> Fix For: 2.0.0
>
>
> The select queries on sla_miss table produce a great % of DB traffic and thus 
> made the DB CPU usage unnecessarily high. It would be a low hanging fruit to 
> add an index and reduce the load.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3062) Add Qubole in integration docs

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3062?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3062:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Add Qubole in integration docs
> --
>
> Key: AIRFLOW-3062
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3062
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Sumit Maheshwari
>Assignee: Sumit Maheshwari
>Priority: Major
> Fix For: 1.10.2
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2596) Oracle to Azure Datalake Transfer Operator

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2596?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2596:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Oracle to Azure Datalake Transfer Operator 
> ---
>
> Key: AIRFLOW-2596
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2596
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: operators
>Reporter: Marcus Rehm
>Assignee: Marcus Rehm
>Priority: Trivial
> Fix For: 1.10.2
>
>
> Like MySQLToHveTransfer  operator would be nice to have a 
> OracleToAzureDataLakeTransfer operator so we can simplify data extraction 
> from Oracle database and put it right into Azure Datalake.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2795) Oracle to Oracle Transfer Operator

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2795?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2795:

Fix Version/s: (was: 2.0.0)

> Oracle to Oracle Transfer Operator 
> ---
>
> Key: AIRFLOW-2795
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2795
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: operators
>Reporter: Marcus Rehm
>Assignee: Marcus Rehm
>Priority: Trivial
> Fix For: 1.10.2
>
>
> This operator should help in transfer data from one Oracle instance to 
> another or between tables in the same instance. t's suitable in use cases 
> where you don't want to or it's not allowed use dblink.
> The operator needs a sql query and a destination table in order to work.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2753) dataproc_operator job operators provide a way to get the actual jobId submitted to Dataproc

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2753?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2753:

Fix Version/s: 1.10.2

> dataproc_operator job operators provide a way to get the actual jobId 
> submitted to Dataproc
> ---
>
> Key: AIRFLOW-2753
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2753
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: contrib, operators
>Reporter: Jeffrey Payne
>Assignee: Jeffrey Payne
>Priority: Minor
>  Labels: dataproc, google, google-compute-engine, operator
> Fix For: 1.10.2, 2.0.0
>
>
> I'd like to be able to get the {{jobId}} that's submitted by the 
> {{DataProcHook}} in order to do things like build the URL to the Dataproc job 
> UI, etc.  The {{job_name}} that the caller passes into 
> {{DataProcPySparkOperator}}, etc gets an 8 character random string appended 
> to it before the job is submitted.  I don't currently see a way to access 
> that value, which is needed for the above mentioned use case.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2795) Oracle to Oracle Transfer Operator

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2795?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2795:

Fix Version/s: 1.10.2

> Oracle to Oracle Transfer Operator 
> ---
>
> Key: AIRFLOW-2795
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2795
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: operators
>Reporter: Marcus Rehm
>Assignee: Marcus Rehm
>Priority: Trivial
> Fix For: 1.10.2
>
>
> This operator should help in transfer data from one Oracle instance to 
> another or between tables in the same instance. t's suitable in use cases 
> where you don't want to or it's not allowed use dblink.
> The operator needs a sql query and a destination table in order to work.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2705) Move class-level moto decorator to method-level

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2705?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2705:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Move class-level moto decorator to method-level
> ---
>
> Key: AIRFLOW-2705
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2705
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: redshift, tests
>Reporter: Kengo Seki
>Assignee: Kengo Seki
>Priority: Major
> Fix For: 1.10.2
>
>
> In tests/contrib/hooks/test_redshift_hook.py and 
> tests/contrib/sensors/test_aws_redshift_cluster_sensor.py, moto decorators 
> are at class-level, whereas they are at method-level in other tests using 
> moto.
> {code:title=tests/contrib/hooks/test_redshift_hook.py}
> @mock_redshift
> class TestRedshiftHook(unittest.TestCase):
> {code}
> I found it can affect other tests using HTTP request. Let's say the following 
> test is added to tests/hooks/test_http_hook.py:
> {code:title=tests/hooks/test_http_hook.py}
> def test_sample(self):
> HttpHook("GET").run("")
> {code}
> This test sends a GET request to https://www.google.com/ and succeeds of 
> course:
> {code}
> $ ./run_unit_tests.sh tests.hooks.test_http_hook:TestHttpHook.test_sample
> (snip)
> /home/sekikn/a3/lib/python3.5/site-packages/urllib3/connectionpool.py:858: 
> InsecureRequestWarning: Unverified HTTPS request is being made. Adding 
> certificate verification is strongly advised. See: 
> https://urllib3.readthedocs.io/en/latest/advanced-usage.html#ssl-warnings
>   InsecureRequestWarning)
> .
> --
> Ran 1 test in 0.189s
> OK
> {code}
> But strangely, it fails with ConnectionError when it's executed after 
> tests/contrib/hooks/test_redshift_hook.py or 
> tests/contrib/sensors/test_aws_redshift_cluster_sensor.py:
> {code}
> $ ./run_unit_tests.sh tests.contrib.hooks.test_redshift_hook 
> tests.hooks.test_http_hook:TestHttpHook.test_sample
> (snip)
> E
> ==
> ERROR: test_sample (tests.hooks.test_http_hook.TestHttpHook)
> --
> Traceback (most recent call last):
>   File "/home/sekikn/dev/incubator-airflow/tests/hooks/test_http_hook.py", 
> line 250, in test_sample
> HttpHook("GET").run("")
>   File "/home/sekikn/dev/incubator-airflow/airflow/hooks/http_hook.py", line 
> 119, in run
> return self.run_and_check(session, prepped_request, extra_options)
>   File "/home/sekikn/dev/incubator-airflow/airflow/hooks/http_hook.py", line 
> 167, in run_and_check
> raise ex
>   File "/home/sekikn/dev/incubator-airflow/airflow/hooks/http_hook.py", line 
> 159, in run_and_check
> allow_redirects=extra_options.get("allow_redirects", True))
>   File "/home/sekikn/a3/lib/python3.5/site-packages/requests/sessions.py", 
> line 618, in send
> r = adapter.send(request, **kwargs)
>   File 
> "/home/sekikn/a3/lib/python3.5/site-packages/moto/packages/responses/responses.py",
>  line 308, in unbound_on_send
> return self._on_request(adapter, request, *a, **kwargs)
>   File 
> "/home/sekikn/a3/lib/python3.5/site-packages/moto/packages/responses/responses.py",
>  line 250, in _on_request
> raise response
> requests.exceptions.ConnectionError: Connection refused: GET 
> https://www.google.com/
>  >> begin captured stdout << -
> [2018-06-30 09:41:21,029] {base_hook.py:83} INFO - Using connection to: 
> https://www.google.com/
> [2018-06-30 09:41:21,030] {http_hook.py:118} INFO - Sending 'GET' to url: 
> https://www.google.com/
> [2018-06-30 09:41:21,030] {http_hook.py:166} WARNING - Connection refused: 
> GET https://www.google.com/ Tenacity will retry to execute the operation
> - >> end captured stdout << --
>  >> begin captured logging << 
> airflow.utils.log.logging_mixin.LoggingMixin: INFO: Using connection to: 
> https://www.google.com/
> airflow.hooks.http_hook.HttpHook: INFO: Sending 'GET' to url: 
> https://www.google.com/
> airflow.hooks.http_hook.HttpHook: WARNING: Connection refused: GET 
> https://www.google.com/ Tenacity will retry to execute the operation
> - >> end captured logging << -
> --
> Ran 7 tests in 1.838s
> FAILED (errors=1)
> {code}
> This problem is solved by using method-level moto decorator instead of 
> class-level one. I suspect class-level decorator patches method that sends 
> HTTP request and doesn't restore the original state.



--
This message was sent by Atlassian JIRA

[jira] [Updated] (AIRFLOW-3365) Missing visibility_timeout will overwrite celery broker transport options

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3365?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3365:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Missing visibility_timeout will overwrite celery broker transport options
> -
>
> Key: AIRFLOW-3365
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3365
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: configuration
>Affects Versions: 1.10.0
>Reporter: Victor Villas Bôas Chaves
>Assignee: Victor Villas Bôas Chaves
>Priority: Major
> Fix For: 1.10.2
>
>
> If you specify the region in your airflow.cfg like this:
> {{[celery_broker_transport_options] }}
> {{region = us-east-2 }}
> it gets ignored.
> This is ignored because Airflow checks to see if the {{visibility_timeout}} 
> parameter is set in the transport options and, if it's not set, replaces the 
> transport options completely with {{{'visibility_timeout': 21600}}} (at 
> [https://github.com/apache/incubator-airflow/blob/master/airflow/config_templates/default_celery.py#L40]).
> If you add {{visibility_timeout}}, the region is correctly set:
> {{[celery_broker_transport_options] }}
> {{region = us-east-2 }}
> {{visibility_timeout = 21600 }}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3294) Add QuboleCheckOperator in integration docs

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3294?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3294:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Add QuboleCheckOperator in integration docs
> ---
>
> Key: AIRFLOW-3294
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3294
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Sumit Maheshwari
>Assignee: Sumit Maheshwari
>Priority: Major
> Fix For: 1.10.2
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2867) Airflow Python Code not compatible to coding guidelines and standards

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2867:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Airflow Python Code not compatible to coding guidelines and standards 
> --
>
> Key: AIRFLOW-2867
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2867
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Kaxil Naik
>Assignee: Kaxil Naik
>Priority: Minor
> Fix For: 1.10.2
>
>
> Some of the Airflow code doesn't conform to python coding guidelines and 
> standards.
> The improvement I have analyzed are below:
> - Dictionary creation should be written by dictionary literal
> - Mutable default argument. Python’s default arguments are evaluated once 
> when the function is defined, not each time the function is called (like it 
> is in say, Ruby). This means that if you use a mutable default argument and 
> mutate it, you will and have mutated that object for all future calls to the 
> function as well.
> - Functions calling sets can be replaced by set literal 
> - Replace list literals
> - Some of the static methods haven't been set static
> - Redundant parentheses



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3343) DockerOperator always reports containers as failed

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3343?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3343:

Fix Version/s: (was: 1.10.1)
   1.10.2

> DockerOperator always reports containers as failed
> --
>
> Key: AIRFLOW-3343
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3343
> Project: Apache Airflow
>  Issue Type: Bug
>Affects Versions: 1.10.1
>Reporter: Ash Berlin-Taylor
>Priority: Major
> Fix For: 1.10.2
>
>
> The docker-py 3.0.0 has some API breaking changes that didn't get picked up 
> because our mocking was still returning the old API. From 
> https://docker-py.readthedocs.io/en/stable/change-log.html#breaking-changes:
> > APIClient.wait and Container.wait now return a dict representing the API’s 
> > response instead of returning the status code directly.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3392) Add index on dag_id in sla_miss table

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3392?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3392.
-
   Resolution: Fixed
Fix Version/s: 1.10.2

> Add index on dag_id in sla_miss table
> -
>
> Key: AIRFLOW-3392
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3392
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Kevin Yang
>Assignee: Kevin Yang
>Priority: Major
> Fix For: 1.10.2
>
>
> The select queries on sla_miss table produce a great % of DB traffic and thus 
> made the DB CPU usage unnecessarily high. It would be a low hanging fruit to 
> add an index and reduce the load.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3301) Update CI test for [AIRFLOW-3132] (PR #3977)

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3301?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3301:

Fix Version/s: 1.10.2

> Update CI test for [AIRFLOW-3132] (PR #3977)
> 
>
> Key: AIRFLOW-3301
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3301
> Project: Apache Airflow
>  Issue Type: Test
>  Components: tests
>Reporter: Xiaodong DENG
>Assignee: Xiaodong DENG
>Priority: Critical
> Fix For: 1.10.2
>
>
> In PR [https://github.com/apache/incubator-airflow/pull/3977,] test is not 
> updated accordingly, and it results in CI failure.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2691) Make Airflow's JS code (and dependencies) manageable via npm and webpack

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2691:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Make Airflow's JS code (and dependencies) manageable via npm and webpack
> 
>
> Key: AIRFLOW-2691
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2691
> Project: Apache Airflow
>  Issue Type: New Feature
>Reporter: Verdan Mahmood
>Assignee: Verdan Mahmood
>Priority: Major
> Fix For: 1.10.2
>
>
> Airflow's JS code is hard to maintain and upgrade. The dependencies are 
> locally existing files making it hard to upgrade versions. 
> Make sure Airflow uses *npm* and *webpack* for the dependencies management. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3236) Create AzureDataLakeStorageListOperator

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3236?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3236:

Fix Version/s: (was: 2.0.0)
   1.10.2

> Create AzureDataLakeStorageListOperator
> ---
>
> Key: AIRFLOW-3236
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3236
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: operators
>Reporter: Brandon Kvarda
>Assignee: Brandon Kvarda
>Priority: Minor
> Fix For: 1.10.2
>
>
> Creates an Operator that is similar to GoogleCloudStorageListOperator and 
> S3ListOperator that returns a list of files at some specified path. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3395) add to documentation all existed REST API endpoints and example how to pass dag_runs params

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3395?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3395:

Fix Version/s: (was: 2.0.0)
   1.10.2

> add to documentation all existed REST API endpoints and example how to pass 
> dag_runs params
> ---
>
> Key: AIRFLOW-3395
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3395
> Project: Apache Airflow
>  Issue Type: Task
>Affects Versions: 1.10.2
>Reporter: Iuliia Volkova
>Assignee: Iuliia Volkova
>Priority: Minor
> Fix For: 1.10.2
>
>
> In doc exist only 2 endpoints https://airflow.apache.org/api.html#endpoints 
> In source code 
> (https://github.com/apache/incubator-airflow/blob/v1-10-stable/airflow/www_rbac/api/experimental/endpoints.py)
>  we have more, caused issues when users think that there are not more methods 
> - I got on work project several questions about it and also saw in 
> stackoverflow relative questions: 
> https://stackoverflow.com/questions/50121593/pass-parameters-to-airflow-experimental-rest-api-when-creating-dag-run
>  
> I want to add more information about REST API



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2642) [kubernetes executor worker] the value of git-sync init container ENV GIT_SYNC_ROOT is wrong

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2642?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2642:

Fix Version/s: (was: 2.0.0)
   1.10.2

> [kubernetes executor worker] the value of git-sync init container ENV 
> GIT_SYNC_ROOT is wrong
> 
>
> Key: AIRFLOW-2642
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2642
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.0
>Reporter: pengchen
>Assignee: pengchen
>Priority: Major
> Fix For: 1.10.2
>
>
> There are two way of syncing dags, pvc and git-sync. When we use git-sync 
> this way, the generated worker pod yaml file fragment is as follows
>  
> {code:java}
> worker container:
> ---
> containers:
> - args:
> - airflow run tutorial1 print_date 2018-06-19T07:57:15.011693+00:00 --local 
> -sd
> /root/airflow/dags/dags/example_dags/tutorial1.py
> command:
> - bash
> - -cx
> - --
> env:
> - name: AIRFLOW__CORE__AIRFLOW_HOME
> value: /root/airflow
> - name: AIRFLOW__CORE__EXECUTOR
> value: LocalExecutor
> - name: AIRFLOW__CORE__DAGS_FOLDER
> value: /tmp/dags
> - name: SQL_ALCHEMY_CONN
> valueFrom:
> secretKeyRef:
> key: sql_alchemy_conn
> name: airflow-secrets
> init container:
> ---
> initContainers:
> - env:
> - name: GIT_SYNC_REPO
> value: https://code.devops.xiaohongshu.com/pengchen/Airflow-DAGs.git
> - name: GIT_SYNC_BRANCH
> value: master
> - name: GIT_SYNC_ROOT
> value: /tmp
> - name: GIT_SYNC_DEST
> value: dags
> - name: GIT_SYNC_ONE_TIME
> value: "true"
> - name: GIT_SYNC_USERNAME
> value: XXX
> - name: GIT_SYNC_PASSWORD
> value: XXX
> image: library/git-sync-amd64:v2.0.5
> imagePullPolicy: IfNotPresent
> name: git-sync-clone
> resources: {}
> securityContext:
> runAsUser: 0
> terminationMessagePath: /dev/termination-log
> terminationMessagePolicy: File
> volumeMounts:
> - mountPath: /root/airflow/dags/
> name: airflow-dags
> - mountPath: /root/airflow/logs
> name: airflow-logs
> - mountPath: /root/airflow/airflow.cfg
> name: airflow-config
> readOnly: true
> subPath: airflow.cfg
> - mountPath: /var/run/secrets/kubernetes.io/serviceaccount
> name: default-token-xz87t
> readOnly: true
> {code}
> According to the configuration, git-sync will synchronize dags to /tmp/dags 
> directory. However the worker container command args(airflow run tutorial1 
> print_date 2018-06-19T07:57:15.011693+00:00 --local -sd
>  /root/airflow/dags/dags/example_dags/tutorial1.py) are generated by the 
> scheduler. Therefore, the task error is as follows
> {code:java}
> + airflow run tutorial1 print_date 2018-06-19T07:57:15.011693+00:00 --local 
> -sd /root/airflow/dags/dags/example_dags/tutorial1.py
> [2018-06-19 07:57:29,075] {settings.py:174} INFO - setting.configure_orm(): 
> Using pool settings. pool_size=5, pool_recycle=1800
> [2018-06-19 07:57:29,232] {__init__.py:51} INFO - Using executor LocalExecutor
> [2018-06-19 07:57:29,373] {models.py:219} INFO - Filling up the DagBag from 
> /root/airflow/dags/dags/example_dags/tutorial1.py
> [2018-06-19 07:57:29,648] {models.py:310} INFO - File 
> /usr/local/lib/python2.7/dist-packages/airflow/example_dags/__init__.py 
> assumed to contain no DAGs. Skipping.
> Traceback (most recent call last):
> File "/usr/local/bin/airflow", line 32, in 
> args.func(args)
> File "/usr/local/lib/python2.7/dist-packages/airflow/utils/cli.py", line 74, 
> in wrapper
> return f(*args, **kwargs)
> File "/usr/local/lib/python2.7/dist-packages/airflow/bin/cli.py", line 475, 
> in run
> dag = get_dag(args)
> File "/usr/local/lib/python2.7/dist-packages/airflow/bin/cli.py", line 146, 
> in get_dag
> 'parse.'.format(args.dag_id))
> airflow.exceptions.AirflowException: dag_id could not be found: tutorial1. 
> Either the dag did not exist or it failed to parse.
> {code}
>  
> The log shows that the worker cannot find the corresponding dag, so I think 
> the environment variable GIT_SYNC_ROOT should be consistent with 
> dag_volume_mount_path.   



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3366) AirflowConfigParser getsection() doesn't pick environment overrides

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3366:

Fix Version/s: (was: 2.0.0)
   1.10.2

> AirflowConfigParser getsection() doesn't pick environment overrides
> ---
>
> Key: AIRFLOW-3366
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3366
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: configuration
>Affects Versions: 1.10.0
>Reporter: Victor Villas Bôas Chaves
>Assignee: Victor Villas Bôas Chaves
>Priority: Major
> Fix For: 1.10.2
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3353) redis-py 3.0.0 dependency breaks celery executor

2018-11-30 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3353?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3353:

Fix Version/s: (was: 2.0.0)
   1.10.2

> redis-py 3.0.0 dependency breaks celery executor
> 
>
> Key: AIRFLOW-3353
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3353
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: celery
>Affects Versions: 1.10.0
>Reporter: Stefan Seelmann
>Assignee: Ash Berlin-Taylor
>Priority: Major
> Fix For: 1.10.2
>
>
> redis-py 3.0.0 was just released. Airflow 1.10.0 defines redis>=2.10.5 so 
> installs redis-py 3.0.0 now.
> Error in worker below.
> Workaround: Pin redis==2.10.6 (e.g. in constraints.txt)
> {code}
> [2018-11-15 12:06:18,441: CRITICAL/MainProcess] Unrecoverable error: 
> AttributeError("'float' object has no attribute 'items'",)
> Traceback (most recent call last):
>   File "/usr/local/lib/python3.6/site-packages/celery/worker/worker.py", line 
> 205, in start
> self.blueprint.start(self)
>   File "/usr/local/lib/python3.6/site-packages/celery/bootsteps.py", line 
> 119, in start
> step.start(parent)
>   File "/usr/local/lib/python3.6/site-packages/celery/bootsteps.py", line 
> 369, in start
> return self.obj.start()
>   File 
> "/usr/local/lib/python3.6/site-packages/celery/worker/consumer/consumer.py", 
> line 317, in start
> blueprint.start(self)
>   File "/usr/local/lib/python3.6/site-packages/celery/bootsteps.py", line 
> 119, in start
> step.start(parent)
>   File 
> "/usr/local/lib/python3.6/site-packages/celery/worker/consumer/consumer.py", 
> line 593, in start
> c.loop(*c.loop_args())
>   File "/usr/local/lib/python3.6/site-packages/celery/worker/loops.py", line 
> 91, in asynloop
> next(loop)
>   File "/usr/local/lib/python3.6/site-packages/kombu/asynchronous/hub.py", 
> line 354, in create_loop
> cb(*cbargs)
>   File "/usr/local/lib/python3.6/site-packages/kombu/transport/redis.py", 
> line 1040, in on_readable
> self.cycle.on_readable(fileno)
>   File "/usr/local/lib/python3.6/site-packages/kombu/transport/redis.py", 
> line 337, in on_readable
> chan.handlers[type]()
>   File "/usr/local/lib/python3.6/site-packages/kombu/transport/redis.py", 
> line 724, in _brpop_read
> self.connection._deliver(loads(bytes_to_str(item)), dest)
>   File 
> "/usr/local/lib/python3.6/site-packages/kombu/transport/virtual/base.py", 
> line 983, in _deliver
> callback(message)
>   File 
> "/usr/local/lib/python3.6/site-packages/kombu/transport/virtual/base.py", 
> line 632, in _callback
> self.qos.append(message, message.delivery_tag)
>   File "/usr/local/lib/python3.6/site-packages/kombu/transport/redis.py", 
> line 149, in append
> pipe.zadd(self.unacked_index_key, time(), delivery_tag) \
>   File "/usr/local/lib/python3.6/site-packages/redis/client.py", line 2263, 
> in zadd
> for pair in iteritems(mapping):
>   File "/usr/local/lib/python3.6/site-packages/redis/_compat.py", line 123, 
> in iteritems
> return iter(x.items())
> AttributeError: 'float' object has no attribute 'items'
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3271) Airflow RBAC Permissions modification via UI do not persist

2018-11-28 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3271:

Fix Version/s: (was: 1.10.2)
   2.0.0

> Airflow RBAC Permissions modification via UI do not persist
> ---
>
> Key: AIRFLOW-3271
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3271
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: ui
>Affects Versions: 1.10.0
>Reporter: Smitha Koduri
>Assignee: Smitha Koduri
>Priority: Major
> Fix For: 2.0.0
>
>
> After upgrading Airflow to 1.10, we have noticed that when attempting to add 
> a new permission-role mapping (via UI), initially it gets successfully added 
> to db. But later, the entry doesn't persist in the db. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2761) Parallelize Celery Executor enqueuing

2018-11-28 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2761?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2761:

Component/s: celery

> Parallelize Celery Executor enqueuing
> -
>
> Key: AIRFLOW-2761
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2761
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: celery
>Reporter: Kevin Yang
>Priority: Major
> Fix For: 1.10.2, 2.0.0
>
>
> Currently celery executor enqueues in an async fashion but still doing that 
> in a single process loop. This can slows down scheduler loop and creates 
> scheduling delay if we have large # of task to schedule in a short time, e.g. 
> UTC midnight we need to schedule large # of sensors in a short period.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-2399) TriggerDagRunOperator Should Permit Implementor to Choose Dag Id at Runtime

2018-11-28 Thread Kaxil Naik (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-2399?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16702446#comment-16702446
 ] 

Kaxil Naik commented on AIRFLOW-2399:
-

Resolved by https://github.com/apache/incubator-airflow/pull/4228

> TriggerDagRunOperator Should Permit Implementor to Choose Dag Id at Runtime
> ---
>
> Key: AIRFLOW-2399
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2399
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: operators
>Reporter: andy dreyfuss
>Assignee: Kaxil Naik
>Priority: Minor
> Fix For: 2.0.0
>
>
> Currently TriggerDagRunOperator takes a string argument to specify the one 
> dag id that is triggered on success. But it would be just as easy and more 
> flexible to allow any. Incoming PR to demo what I mean ...



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2399) TriggerDagRunOperator Should Permit Implementor to Choose Dag Id at Runtime

2018-11-28 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2399?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2399:

External issue URL:   (was: 
https://github.com/apache/incubator-airflow/pull/4228)

> TriggerDagRunOperator Should Permit Implementor to Choose Dag Id at Runtime
> ---
>
> Key: AIRFLOW-2399
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2399
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: operators
>Reporter: andy dreyfuss
>Assignee: Kaxil Naik
>Priority: Minor
> Fix For: 2.0.0
>
>
> Currently TriggerDagRunOperator takes a string argument to specify the one 
> dag id that is triggered on success. But it would be just as easy and more 
> flexible to allow any. Incoming PR to demo what I mean ...



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2399) TriggerDagRunOperator Should Permit Implementor to Choose Dag Id at Runtime

2018-11-28 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2399?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2399:

External issue URL: https://github.com/apache/incubator-airflow/pull/4228

> TriggerDagRunOperator Should Permit Implementor to Choose Dag Id at Runtime
> ---
>
> Key: AIRFLOW-2399
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2399
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: operators
>Reporter: andy dreyfuss
>Assignee: Kaxil Naik
>Priority: Minor
> Fix For: 2.0.0
>
>
> Currently TriggerDagRunOperator takes a string argument to specify the one 
> dag id that is triggered on success. But it would be just as easy and more 
> flexible to allow any. Incoming PR to demo what I mean ...



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (AIRFLOW-3410) Remote Host Identification Has Changed Error while using SSHOperator

2018-11-28 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3410?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik resolved AIRFLOW-3410.
-
Resolution: Fixed

Resolved by https://github.com/apache/incubator-airflow/pull/4249

> Remote Host Identification Has Changed Error while using SSHOperator
> 
>
> Key: AIRFLOW-3410
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3410
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: contrib
>Reporter: Kaxil Naik
>Assignee: Kaxil Naik
>Priority: Minor
> Fix For: 1.10.2
>
>
> Currently, there is no provision in Airflow's implementation of SSHOperator 
> to disable Host Checking and use a Null Known Host file to allow connecting 
> to host that has a new IP.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3346) Add gcp transfer service hook and operators

2018-11-28 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3346:

Issue Type: New Feature  (was: Bug)

> Add gcp transfer service hook and operators
> ---
>
> Key: AIRFLOW-3346
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3346
> Project: Apache Airflow
>  Issue Type: New Feature
>Reporter: Josh Carp
>Priority: Minor
> Fix For: 1.10.2
>
>
> Add a hook and operator(s) to connect to gcp storage transfer service and 
> transfer files from s3 to gcp (and gcp to gcp) without copying to local disk.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3306) Disable unused flask-sqlalchemy modification tracking

2018-11-28 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3306?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3306:

Issue Type: Improvement  (was: Bug)

> Disable unused flask-sqlalchemy modification tracking
> -
>
> Key: AIRFLOW-3306
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3306
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Josh Carp
>Assignee: Josh Carp
>Priority: Trivial
> Fix For: 1.10.2, 2.0.0
>
>
> By default, flask-sqlalchemy tracks model changes for its event system, which 
> adds some overhead. Since I don't think we're using the flask-sqlalchemy 
> event system, we should be able to turn off modification tracking and improve 
> performance.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-2760) DAG parsing loop coupled with scheduler loop

2018-11-28 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2760?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-2760:

Fix Version/s: (was: 2.0.0)
   1.10.2

> DAG parsing loop coupled with scheduler loop
> 
>
> Key: AIRFLOW-2760
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2760
> Project: Apache Airflow
>  Issue Type: Improvement
>Reporter: Kevin Yang
>Assignee: Kevin Yang
>Priority: Major
> Fix For: 1.10.2
>
>
> Currently DAG parsing loop is coupled with scheduler loop, meaning that if 
> scheduler loop became slow, we will parse DAG slower.
> As a simple producer and consumer pattern, we shall have them decoupled and 
> completely remove the scheduling bottleneck placed by DAG parsing--which is 
> identified in Airbnb as the current biggest bottleneck.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (AIRFLOW-3410) Remote Host Identification Has Changed Error while using SSHOperator

2018-11-27 Thread Kaxil Naik (JIRA)
Kaxil Naik created AIRFLOW-3410:
---

 Summary: Remote Host Identification Has Changed Error while using 
SSHOperator
 Key: AIRFLOW-3410
 URL: https://issues.apache.org/jira/browse/AIRFLOW-3410
 Project: Apache Airflow
  Issue Type: New Feature
  Components: contrib
Reporter: Kaxil Naik
Assignee: Kaxil Naik
 Fix For: 1.10.2


Currently, there is no provision in Airflow's implementation of SSHOperator to 
disable Host Checking and use a Null Known Host file to allow connecting to 
host that has a new IP.





--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (AIRFLOW-2104) Recommendation for Airflow design/architecture

2018-11-27 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-2104?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik closed AIRFLOW-2104.
---
Resolution: Not A Problem

> Recommendation for Airflow design/architecture
> --
>
> Key: AIRFLOW-2104
> URL: https://issues.apache.org/jira/browse/AIRFLOW-2104
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: celery, configuration, DagRun
>Affects Versions: 1.9.0
>Reporter: Anil Kumar
>Priority: Major
> Fix For: 1.9.0
>
>
> Hello,
> We are in the implementation phase for Apache Airflow and looking for 
> Recommendation on Airflow configuration/structure.
> Which one will be best option
>  # Airflow standalone installation with multi worker nodes or
>  # Airflow running as docker image on Kubernetes
> Requirement :-
> We are suppose to schedule a long running ETL process that includes complex 
> unzip/decrypt task, execution of data wrangling UDF's.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3396) Redundant sql param check in BigQueryOperator

2018-11-25 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3396?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3396:

Description: The 'sql' parameter in BigQueryOperator accepts None and then 
there is a check which checks if it is None and gives error.  (was: The 'sql' p)

> Redundant sql param check in BigQueryOperator
> -
>
> Key: AIRFLOW-3396
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3396
> Project: Apache Airflow
>  Issue Type: Task
>  Components: gcp
>Reporter: Kaxil Naik
>Assignee: Kaxil Naik
>Priority: Trivial
> Fix For: 2.0.0
>
>
> The 'sql' parameter in BigQueryOperator accepts None and then there is a 
> check which checks if it is None and gives error.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (AIRFLOW-3396) Redundant sql param check in BigQueryOperator

2018-11-25 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-3396?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik updated AIRFLOW-3396:

Description: The 'sql' p

> Redundant sql param check in BigQueryOperator
> -
>
> Key: AIRFLOW-3396
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3396
> Project: Apache Airflow
>  Issue Type: Task
>  Components: gcp
>Reporter: Kaxil Naik
>Assignee: Kaxil Naik
>Priority: Trivial
> Fix For: 2.0.0
>
>
> The 'sql' p



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (AIRFLOW-3396) Redundant sql param check in BigQueryOperator

2018-11-25 Thread Kaxil Naik (JIRA)
Kaxil Naik created AIRFLOW-3396:
---

 Summary: Redundant sql param check in BigQueryOperator
 Key: AIRFLOW-3396
 URL: https://issues.apache.org/jira/browse/AIRFLOW-3396
 Project: Apache Airflow
  Issue Type: Task
  Components: gcp
Reporter: Kaxil Naik
Assignee: Kaxil Naik
 Fix For: 2.0.0






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (AIRFLOW-156) Add date option to trigger_dag

2018-11-23 Thread Kaxil Naik (JIRA)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-156?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kaxil Naik closed AIRFLOW-156.
--
Resolution: Duplicate

> Add date option to trigger_dag
> --
>
> Key: AIRFLOW-156
> URL: https://issues.apache.org/jira/browse/AIRFLOW-156
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: cli
>Reporter: Jeffrey Picard
>Priority: Minor
>
> Currently the trigger_dag command always sets the execution date to
> datetime.now(). This seems like a rather arbitrary restriction and there
> are use cases when running dags ad-hoc where one may wish to set a
> different execution date.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


  1   2   3   4   5   6   >