[GitHub] [airflow] RosterIn commented on issue #7398: [AIRFLOW-6505] Let emoji encoded properly for json.dumps()

2020-02-10 Thread GitBox
RosterIn commented on issue #7398: [AIRFLOW-6505]  Let emoji encoded properly 
for json.dumps()
URL: https://github.com/apache/airflow/pull/7398#issuecomment-584514650
 
 
   @damon09273 
   `airflow/contrib/operators/sql_to_gcs.py` doesn't exist in master (which is 
why you get conflict)
   You need to edit:
   
https://github.com/apache/airflow/blob/master/airflow/providers/google/cloud/operators/sql_to_gcs.py#L181


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] damon09273 closed pull request #7398: [AIRFLOW-6505] Let emoji encoded properly for json.dumps()

2020-02-10 Thread GitBox
damon09273 closed pull request #7398: [AIRFLOW-6505]  Let emoji encoded 
properly for json.dumps()
URL: https://github.com/apache/airflow/pull/7398
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (AIRFLOW-6505) Let emoji encoded properly for json.dumps() -- BaseSQLToGoogleCloudStorageOperator

2020-02-10 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-6505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17034207#comment-17034207
 ] 

ASF GitHub Bot commented on AIRFLOW-6505:
-

damon09273 commented on pull request #7398: [AIRFLOW-6505]  Let emoji encoded 
properly for json.dumps()
URL: https://github.com/apache/airflow/pull/7398
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Let emoji encoded properly for json.dumps() -- 
> BaseSQLToGoogleCloudStorageOperator
> --
>
> Key: AIRFLOW-6505
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6505
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.8
> Environment: 1.10.7
>Reporter: Damon Liao
>Assignee: Damon Liao
>Priority: Blocker
> Fix For: 1.10.7
>
>
> Emoji can't be encoded properly when ` json.dumps()` and 'UTF-8', problem 
> fixed by adding parameter `ensure_ascii=False`. 
>  I've created a [PR|[https://github.com/apache/airflow/pull/7398]] to fix 
> this problem.
> In [line|#L175]] BaseSQLToGoogleCloudStorageOperator.
>  
> For example
> the emoji  encoded differentially when in use or not use  
> `ensure_ascii=False`.
> In Use (correct UTF-8 encode): "\xf0\x9f\x8d\xbb"
> Not Use (only 2 slash):"\\\ud83c\\\udf7b"
>  
> Ref: 
> [https://stackoverflow.com/questions/51183947/python-json-dumps-doesnt-encode-emojis-properly]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (AIRFLOW-6505) Let emoji encoded properly for json.dumps() -- BaseSQLToGoogleCloudStorageOperator

2020-02-10 Thread Damon Liao (Jira)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-6505?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Damon Liao updated AIRFLOW-6505:

Description: 
Emoji can't be encoded properly when ` json.dumps()` and 'UTF-8', problem fixed 
by adding parameter `ensure_ascii=False`. 

 I've created a [PR|[https://github.com/apache/airflow/pull/7398]] to fix this 
problem.

In [line|#L175]] BaseSQLToGoogleCloudStorageOperator.

 

For example

the emoji  encoded differentially when in use or not use  `ensure_ascii=False`.

In Use (correct UTF-8 encode): "\xf0\x9f\x8d\xbb"

Not Use (only 2 slash):"\\\ud83c\\\udf7b"

 

Ref: 
[https://stackoverflow.com/questions/51183947/python-json-dumps-doesnt-encode-emojis-properly]

  was:
Emoji can't be encoded properly when ` json.dumps()` and 'UTF-8', problem fixed 
by adding parameter `ensure_ascii=False`. 

 I've created a [PR|[https://github.com/damon09273/airflow/pull/1]] to fix this 
problem.

In [line|#L175]] BaseSQLToGoogleCloudStorageOperato.

 

For example

the emoji  encoded differentially when in use or not use  `ensure_ascii=False`.

In Use (correct UTF-8 encode): "\xf0\x9f\x8d\xbb"

Not Use (only 2 slash):"\\\ud83c\\\udf7b"

 

Ref: 
[https://stackoverflow.com/questions/51183947/python-json-dumps-doesnt-encode-emojis-properly]


> Let emoji encoded properly for json.dumps() -- 
> BaseSQLToGoogleCloudStorageOperator
> --
>
> Key: AIRFLOW-6505
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6505
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.8
> Environment: 1.10.7
>Reporter: Damon Liao
>Assignee: Damon Liao
>Priority: Blocker
> Fix For: 1.10.7
>
>
> Emoji can't be encoded properly when ` json.dumps()` and 'UTF-8', problem 
> fixed by adding parameter `ensure_ascii=False`. 
>  I've created a [PR|[https://github.com/apache/airflow/pull/7398]] to fix 
> this problem.
> In [line|#L175]] BaseSQLToGoogleCloudStorageOperator.
>  
> For example
> the emoji  encoded differentially when in use or not use  
> `ensure_ascii=False`.
> In Use (correct UTF-8 encode): "\xf0\x9f\x8d\xbb"
> Not Use (only 2 slash):"\\\ud83c\\\udf7b"
>  
> Ref: 
> [https://stackoverflow.com/questions/51183947/python-json-dumps-doesnt-encode-emojis-properly]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (AIRFLOW-6505) Let emoji encoded properly for json.dumps() -- BaseSQLToGoogleCloudStorageOperator

2020-02-10 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-6505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17034196#comment-17034196
 ] 

ASF GitHub Bot commented on AIRFLOW-6505:
-

damon09273 commented on pull request #7398: [AIRFLOW-6505]  Let emoji encoded 
properly for json.dumps()
URL: https://github.com/apache/airflow/pull/7398
 
 
   Make sure you have checked _all_ steps below.
   
   ### Jira
   
   - [x] My PR addresses the following [Airflow 
Jira](https://issues.apache.org/jira/browse/AIRFLOW-6505) issues and references 
them in the PR title. For example, "\[AIRFLOW-XXX\] My Airflow PR"
 - https://issues.apache.org/jira/browse/AIRFLOW-6505
   
   ### Description
   Emoji can't be encoded properly when ` json.dumps()` and 'UTF-8', problem 
fixed by adding parameter `ensure_ascii=False`. 
   In 
[line](https://github.com/apache/airflow/blob/1.10.7/airflow/contrib/operators/sql_to_gcs.py#L175)
 BaseSQLToGoogleCloudStorageOperator.
   
   For example
   
   the emoji  encoded differentially when in use or not use  
`ensure_ascii=False`.
   
   In Use (correct UTF-8 encode): "\xf0\x9f\x8d\xbb"
   
   Not Use:"ud83cudf7b"
   

   
   Ref: 
https://stackoverflow.com/questions/51183947/python-json-dumps-doesnt-encode-emojis-properly
   
   
   
   ### Tests
   
   - [x] My PR adds the following unit tests __OR__ does not need testing for 
this extremely good reason:
   
   ### Commits
   
   - [x] My commits all reference Jira issues in their subject lines, and I 
have squashed multiple commits if they address the same issue. In addition, my 
commits follow the guidelines from "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)":
 1. Subject is separated from body by a blank line
 1. Subject is limited to 50 characters (not including Jira issue reference)
 1. Subject does not end with a period
 1. Subject uses the imperative mood ("add", not "adding")
 1. Body wraps at 72 characters
 1. Body explains "what" and "why", not "how"
   
   ### Documentation
   
   - [x] In case of new functionality, my PR adds documentation that describes 
how to use it.
 - All the public functions and the classes in the PR contain docstrings 
that explain what it does
 - If you implement backwards incompatible changes, please leave a note in 
the [Updating.md](https://github.com/apache/airflow/blob/master/UPDATING.md) so 
we can assign it to a appropriate release
   
   ### Code Quality
   
   - [x] Passes `flake8`
   
   Make sure you have checked _all_ steps below.
   
   ### Jira
   
   - [ ] My PR addresses the following [Airflow 
Jira](https://issues.apache.org/jira/browse/AIRFLOW/) issues and references 
them in the PR title. For example, "\[AIRFLOW-XXX\] My Airflow PR"
 - https://issues.apache.org/jira/browse/AIRFLOW-XXX
 - In case you are fixing a typo in the documentation you can prepend your 
commit with \[AIRFLOW-XXX\], code changes always need a Jira issue.
 - In case you are proposing a fundamental code change, you need to create 
an Airflow Improvement Proposal 
([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals)).
 - In case you are adding a dependency, check if the license complies with 
the [ASF 3rd Party License 
Policy](https://www.apache.org/legal/resolved.html#category-x).
   
   ### Description
   
   - [ ] Here are some details about my PR, including screenshots of any UI 
changes:
   
   ### Tests
   
   - [ ] My PR adds the following unit tests __OR__ does not need testing for 
this extremely good reason:
   
   ### Commits
   
   - [ ] My commits all reference Jira issues in their subject lines, and I 
have squashed multiple commits if they address the same issue. In addition, my 
commits follow the guidelines from "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)":
 1. Subject is separated from body by a blank line
 1. Subject is limited to 50 characters (not including Jira issue reference)
 1. Subject does not end with a period
 1. Subject uses the imperative mood ("add", not "adding")
 1. Body wraps at 72 characters
 1. Body explains "what" and "why", not "how"
   
   ### Documentation
   
   - [ ] In case of new functionality, my PR adds documentation that describes 
how to use it.
 - All the public functions and the classes in the PR contain docstrings 
that explain what it does
 - If you implement backwards incompatible changes, please leave a note in 
the [Updating.md](https://github.com/apache/airflow/blob/master/UPDATING.md) so 
we can assign it to a appropriate release
   
   ### Code Quality
   
   - [ ] Passes `flake8`
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go 

[GitHub] [airflow-site] XD-DENG commented on issue #238: Typo in community page

2020-02-10 Thread GitBox
XD-DENG commented on issue #238: Typo in community page
URL: https://github.com/apache/airflow-site/issues/238#issuecomment-584510155
 
 
   Thanks @louisguitton 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] damon09273 opened a new pull request #7398: [AIRFLOW-6505] Let emoji encoded properly for json.dumps()

2020-02-10 Thread GitBox
damon09273 opened a new pull request #7398: [AIRFLOW-6505]  Let emoji encoded 
properly for json.dumps()
URL: https://github.com/apache/airflow/pull/7398
 
 
   Make sure you have checked _all_ steps below.
   
   ### Jira
   
   - [x] My PR addresses the following [Airflow 
Jira](https://issues.apache.org/jira/browse/AIRFLOW-6505) issues and references 
them in the PR title. For example, "\[AIRFLOW-XXX\] My Airflow PR"
 - https://issues.apache.org/jira/browse/AIRFLOW-6505
   
   ### Description
   Emoji can't be encoded properly when ` json.dumps()` and 'UTF-8', problem 
fixed by adding parameter `ensure_ascii=False`. 
   In 
[line](https://github.com/apache/airflow/blob/1.10.7/airflow/contrib/operators/sql_to_gcs.py#L175)
 BaseSQLToGoogleCloudStorageOperator.
   
   For example
   
   the emoji  encoded differentially when in use or not use  
`ensure_ascii=False`.
   
   In Use (correct UTF-8 encode): "\xf0\x9f\x8d\xbb"
   
   Not Use:"ud83cudf7b"
   

   
   Ref: 
https://stackoverflow.com/questions/51183947/python-json-dumps-doesnt-encode-emojis-properly
   
   
   
   ### Tests
   
   - [x] My PR adds the following unit tests __OR__ does not need testing for 
this extremely good reason:
   
   ### Commits
   
   - [x] My commits all reference Jira issues in their subject lines, and I 
have squashed multiple commits if they address the same issue. In addition, my 
commits follow the guidelines from "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)":
 1. Subject is separated from body by a blank line
 1. Subject is limited to 50 characters (not including Jira issue reference)
 1. Subject does not end with a period
 1. Subject uses the imperative mood ("add", not "adding")
 1. Body wraps at 72 characters
 1. Body explains "what" and "why", not "how"
   
   ### Documentation
   
   - [x] In case of new functionality, my PR adds documentation that describes 
how to use it.
 - All the public functions and the classes in the PR contain docstrings 
that explain what it does
 - If you implement backwards incompatible changes, please leave a note in 
the [Updating.md](https://github.com/apache/airflow/blob/master/UPDATING.md) so 
we can assign it to a appropriate release
   
   ### Code Quality
   
   - [x] Passes `flake8`
   
   Make sure you have checked _all_ steps below.
   
   ### Jira
   
   - [ ] My PR addresses the following [Airflow 
Jira](https://issues.apache.org/jira/browse/AIRFLOW/) issues and references 
them in the PR title. For example, "\[AIRFLOW-XXX\] My Airflow PR"
 - https://issues.apache.org/jira/browse/AIRFLOW-XXX
 - In case you are fixing a typo in the documentation you can prepend your 
commit with \[AIRFLOW-XXX\], code changes always need a Jira issue.
 - In case you are proposing a fundamental code change, you need to create 
an Airflow Improvement Proposal 
([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals)).
 - In case you are adding a dependency, check if the license complies with 
the [ASF 3rd Party License 
Policy](https://www.apache.org/legal/resolved.html#category-x).
   
   ### Description
   
   - [ ] Here are some details about my PR, including screenshots of any UI 
changes:
   
   ### Tests
   
   - [ ] My PR adds the following unit tests __OR__ does not need testing for 
this extremely good reason:
   
   ### Commits
   
   - [ ] My commits all reference Jira issues in their subject lines, and I 
have squashed multiple commits if they address the same issue. In addition, my 
commits follow the guidelines from "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)":
 1. Subject is separated from body by a blank line
 1. Subject is limited to 50 characters (not including Jira issue reference)
 1. Subject does not end with a period
 1. Subject uses the imperative mood ("add", not "adding")
 1. Body wraps at 72 characters
 1. Body explains "what" and "why", not "how"
   
   ### Documentation
   
   - [ ] In case of new functionality, my PR adds documentation that describes 
how to use it.
 - All the public functions and the classes in the PR contain docstrings 
that explain what it does
 - If you implement backwards incompatible changes, please leave a note in 
the [Updating.md](https://github.com/apache/airflow/blob/master/UPDATING.md) so 
we can assign it to a appropriate release
   
   ### Code Quality
   
   - [ ] Passes `flake8`
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[airflow-site] branch master updated: Add mail list unsubscribe link (#250)

2020-02-10 Thread xddeng
This is an automated email from the ASF dual-hosted git repository.

xddeng pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/airflow-site.git


The following commit(s) were added to refs/heads/master by this push:
 new c598425  Add mail list unsubscribe link (#250)
c598425 is described below

commit c598425ce4f515814ff943fe0bf442c950fecf64
Author: Fokko Driesprong 
AuthorDate: Tue Feb 11 08:33:41 2020 +0100

Add mail list unsubscribe link (#250)
---
 landing-pages/site/content/en/community/_index.html | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/landing-pages/site/content/en/community/_index.html 
b/landing-pages/site/content/en/community/_index.html
index 8628577..c77199f 100644
--- a/landing-pages/site/content/en/community/_index.html
+++ b/landing-pages/site/content/en/community/_index.html
@@ -16,10 +16,10 @@ menu:
 
 
 
-Users list: mailto:users-subscr...@airflow.apache.org;>users-subscr...@airflow.apache.org
+Users list: mailto:users-subscr...@airflow.apache.org;>users-subscr...@airflow.apache.org
 and mailto:users-unsubscr...@airflow.apache.org;>users-unsubscr...@airflow.apache.org
 to unsubscribe
 
 
-Dev list: mailto:dev-subscr...@airflow.apache.org;>dev-subscr...@airflow.apache.org
+Dev list: mailto:dev-subscr...@airflow.apache.org;>dev-subscr...@airflow.apache.org
 and mailto:dev-unsubscr...@airflow.apache.org;>dev-unsubscr...@airflow.apache.org
 to unsubscribe
 
 
 



[GitHub] [airflow-site] XD-DENG merged pull request #250: Add unsubscribe link

2020-02-10 Thread GitBox
XD-DENG merged pull request #250: Add unsubscribe link
URL: https://github.com/apache/airflow-site/pull/250
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] damon09273 closed pull request #7397: Let emoji encoded properly for json.dumps() -- BaseSQLToGoogleCloudStorageOperator

2020-02-10 Thread GitBox
damon09273 closed pull request #7397: Let emoji encoded properly for 
json.dumps() -- BaseSQLToGoogleCloudStorageOperator
URL: https://github.com/apache/airflow/pull/7397
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] damon09273 opened a new pull request #7397: Let emoji encoded properly for json.dumps() -- BaseSQLToGoogleCloudStorageOperator

2020-02-10 Thread GitBox
damon09273 opened a new pull request #7397: Let emoji encoded properly for 
json.dumps() -- BaseSQLToGoogleCloudStorageOperator
URL: https://github.com/apache/airflow/pull/7397
 
 
   Make sure you have checked _all_ steps below.
   
   ### Jira
   
   - [x] My PR addresses the following [Airflow 
Jira](https://issues.apache.org/jira/browse/AIRFLOW-6505) issues and references 
them in the PR title. For example, "\[AIRFLOW-XXX\] My Airflow PR"
 - https://issues.apache.org/jira/browse/AIRFLOW-6505
   
   ### Description
   Emoji can't be encoded properly when ` json.dumps()` and 'UTF-8', problem 
fixed by adding parameter `ensure_ascii=False`. 
   In 
[line](https://github.com/apache/airflow/blob/1.10.7/airflow/contrib/operators/sql_to_gcs.py#L175)
 BaseSQLToGoogleCloudStorageOperato.
   
   For example
   
   the emoji  encoded differentially when in use or not use  
`ensure_ascii=False`.
   
   In Use (correct UTF-8 encode): "\xf0\x9f\x8d\xbb"
   
   Not Use:"ud83cudf7b"
   

   
   Ref: 
https://stackoverflow.com/questions/51183947/python-json-dumps-doesnt-encode-emojis-properly
   
   
   
   ### Tests
   
   - [x] My PR adds the following unit tests __OR__ does not need testing for 
this extremely good reason:
   
   ### Commits
   
   - [x] My commits all reference Jira issues in their subject lines, and I 
have squashed multiple commits if they address the same issue. In addition, my 
commits follow the guidelines from "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)":
 1. Subject is separated from body by a blank line
 1. Subject is limited to 50 characters (not including Jira issue reference)
 1. Subject does not end with a period
 1. Subject uses the imperative mood ("add", not "adding")
 1. Body wraps at 72 characters
 1. Body explains "what" and "why", not "how"
   
   ### Documentation
   
   - [x] In case of new functionality, my PR adds documentation that describes 
how to use it.
 - All the public functions and the classes in the PR contain docstrings 
that explain what it does
 - If you implement backwards incompatible changes, please leave a note in 
the [Updating.md](https://github.com/apache/airflow/blob/master/UPDATING.md) so 
we can assign it to a appropriate release
   
   ### Code Quality
   
   - [x] Passes `flake8`
   
   Make sure you have checked _all_ steps below.
   
   ### Jira
   
   - [ ] My PR addresses the following [Airflow 
Jira](https://issues.apache.org/jira/browse/AIRFLOW/) issues and references 
them in the PR title. For example, "\[AIRFLOW-XXX\] My Airflow PR"
 - https://issues.apache.org/jira/browse/AIRFLOW-XXX
 - In case you are fixing a typo in the documentation you can prepend your 
commit with \[AIRFLOW-XXX\], code changes always need a Jira issue.
 - In case you are proposing a fundamental code change, you need to create 
an Airflow Improvement Proposal 
([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals)).
 - In case you are adding a dependency, check if the license complies with 
the [ASF 3rd Party License 
Policy](https://www.apache.org/legal/resolved.html#category-x).
   
   ### Description
   
   - [ ] Here are some details about my PR, including screenshots of any UI 
changes:
   
   ### Tests
   
   - [ ] My PR adds the following unit tests __OR__ does not need testing for 
this extremely good reason:
   
   ### Commits
   
   - [ ] My commits all reference Jira issues in their subject lines, and I 
have squashed multiple commits if they address the same issue. In addition, my 
commits follow the guidelines from "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)":
 1. Subject is separated from body by a blank line
 1. Subject is limited to 50 characters (not including Jira issue reference)
 1. Subject does not end with a period
 1. Subject uses the imperative mood ("add", not "adding")
 1. Body wraps at 72 characters
 1. Body explains "what" and "why", not "how"
   
   ### Documentation
   
   - [ ] In case of new functionality, my PR adds documentation that describes 
how to use it.
 - All the public functions and the classes in the PR contain docstrings 
that explain what it does
 - If you implement backwards incompatible changes, please leave a note in 
the [Updating.md](https://github.com/apache/airflow/blob/master/UPDATING.md) so 
we can assign it to a appropriate release
   
   ### Code Quality
   
   - [ ] Passes `flake8`
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] feluelle commented on issue #6576: [AIRFLOW-5922] Add option to specify the mysql client library used in MySqlHook

2020-02-10 Thread GitBox
feluelle commented on issue #6576: [AIRFLOW-5922] Add option to specify the 
mysql client library used in MySqlHook
URL: https://github.com/apache/airflow/pull/6576#issuecomment-584507024
 
 
   @potiuk @ashb @nuclearpinguin do you have any idea what I can do to fix the 
segmentation error?
   
   I tried to reorder the imports but that didn't help :/


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow-site] Fokko opened a new pull request #250: Add unsubscribe link

2020-02-10 Thread GitBox
Fokko opened a new pull request #250: Add unsubscribe link
URL: https://github.com/apache/airflow-site/pull/250
 
 
   It isn't clear how to unsubscribe from the mail list.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[airflow-site] branch master updated: Fix typo - Commiter => Committer (#247)

2020-02-10 Thread fokko
This is an automated email from the ASF dual-hosted git repository.

fokko pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/airflow-site.git


The following commit(s) were added to refs/heads/master by this push:
 new 72de3af  Fix typo - Commiter => Committer (#247)
72de3af is described below

commit 72de3af50a099b4537175fff0e714eaa5f47ca3f
Author: Louis Guitton 
AuthorDate: Tue Feb 11 08:20:10 2020 +0100

Fix typo - Commiter => Committer (#247)

Follow-up #243, Fix #238
---
 docs-archive/1.10.4/changelog.html |  2 +-
 docs-archive/1.10.5/changelog.html |  2 +-
 docs-archive/1.10.6/_static/_gen/css/main-custom.min.css   |  2 +-
 docs-archive/1.10.6/changelog.html |  4 ++--
 docs-archive/1.10.7/_static/_gen/css/main-custom.min.css   |  2 +-
 docs-archive/1.10.7/changelog.html |  4 ++--
 docs-archive/1.10.8/_static/_gen/css/main-custom.min.css   |  2 +-
 docs-archive/1.10.8/changelog.html |  4 ++--
 landing-pages/site/assets/scss/_community-page.scss|  4 ++--
 landing-pages/site/assets/scss/_list-boxes.scss|  2 +-
 ...pe-2019-thoughts-and-insights-by-airflow-committers.md} |  4 ++--
 .../site/data/{commiters.json => committers.json}  |  0
 landing-pages/site/layouts/community/list.html | 14 +++---
 .../partials/boxes/{commiter.html => committer.html}   | 14 +++---
 landing-pages/src/index.js |  6 +++---
 .../src/js/{commitersList.js => committersList.js} |  0
 16 files changed, 33 insertions(+), 33 deletions(-)

diff --git a/docs-archive/1.10.4/changelog.html 
b/docs-archive/1.10.4/changelog.html
index bb268a4..458bb7e 100644
--- a/docs-archive/1.10.4/changelog.html
+++ b/docs-archive/1.10.4/changelog.html
@@ -620,7 +620,7 @@
 [AIRFLOW-XXX] Reduce log spam in tests (#5174)
 [AIRFLOW-XXX] Speed up tests for PythonSensor (#5158)
 [AIRFLOW-XXX] Add Bas Harenslak to committer list (#5157)
-[AIRFLOW-XXX] Add Jarek Potiuk to commiter list (#5132)
+[AIRFLOW-XXX] Add Jarek Potiuk to committer list (#5132)
 [AIRFLOW-XXX] Update docstring for SchedulerJob (#5105)
 [AIRFLOW-XXX] Fix docstrings for CassandraToGoogleCloudStorageOperator 
(#5103)
 [AIRFLOW-XXX] update SlackWebhookHook and SlackWebhookOperator 
docstring (#5074)
diff --git a/docs-archive/1.10.5/changelog.html 
b/docs-archive/1.10.5/changelog.html
index 566365c..b67121e 100644
--- a/docs-archive/1.10.5/changelog.html
+++ b/docs-archive/1.10.5/changelog.html
@@ -738,7 +738,7 @@
 [AIRFLOW-XXX] Reduce log spam in tests (#5174)
 [AIRFLOW-XXX] Speed up tests for PythonSensor (#5158)
 [AIRFLOW-XXX] Add Bas Harenslak to committer list (#5157)
-[AIRFLOW-XXX] Add Jarek Potiuk to commiter list (#5132)
+[AIRFLOW-XXX] Add Jarek Potiuk to committer list (#5132)
 [AIRFLOW-XXX] Update docstring for SchedulerJob (#5105)
 [AIRFLOW-XXX] Fix docstrings for CassandraToGoogleCloudStorageOperator 
(#5103)
 [AIRFLOW-XXX] update SlackWebhookHook and SlackWebhookOperator 
docstring (#5074)
diff --git a/docs-archive/1.10.6/_static/_gen/css/main-custom.min.css 
b/docs-archive/1.10.6/_static/_gen/css/main-custom.min.css
index 2d686bf..c8ebf36 100644
--- a/docs-archive/1.10.6/_static/_gen/css/main-custom.min.css
+++ b/docs-archive/1.10.6/_static/_gen/css/main-custom.min.css
@@ -1 +1 @@
-@charset "UTF-8";@import 
"https://fonts.googleapis.com/css?family=Rubik:500=swap";@import 
"https://fonts.googleapis.com/css?family=Roboto:400,400i,500,700=swap";@import
 
"https://fonts.googleapis.com/css?family=Roboto+Mono:400,700=swap";.header__large--cerulean-blue{font-family:rubik,sans-serif;font-weight:500;font-size:72px;line-height:1.17;color:#017cee}.header__large--shamrock{font-family:rubik,sans-serif;font-weight:500;font-size:72px;line-height:1.17;color:#00
 [...]
\ No newline at end of file
+@charset "UTF-8";@import 
"https://fonts.googleapis.com/css?family=Rubik:500=swap";@import 
"https://fonts.googleapis.com/css?family=Roboto:400,400i,500,700=swap";@import
 
"https://fonts.googleapis.com/css?family=Roboto+Mono:400,700=swap";.header__large--cerulean-blue{font-family:rubik,sans-serif;font-weight:500;font-size:72px;line-height:1.17;color:#017cee}.header__large--shamrock{font-family:rubik,sans-serif;font-weight:500;font-size:72px;line-height:1.17;color:#00
 [...]
diff --git a/docs-archive/1.10.6/changelog.html 
b/docs-archive/1.10.6/changelog.html
index c5f048d..7ac9e42 100644
--- a/docs-archive/1.10.6/changelog.html
+++ b/docs-archive/1.10.6/changelog.html
@@ -1285,7 +1285,7 @@ https://www.sphinx-doc.org/en/master/templating.html
 [AIRFLOW-XXX] Reduce log spam in tests (#5174)
 [AIRFLOW-XXX] Speed up tests for PythonSensor (#5158)
 [AIRFLOW-XXX] Add Bas Harenslak to committer list (#5157)
-[AIRFLOW-XXX] Add Jarek Potiuk to commiter list (#5132)
+[AIRFLOW-XXX] Add Jarek Potiuk to committer list (#5132)
 [AIRFLOW-XXX] Update 

[GitHub] [airflow-site] Fokko commented on a change in pull request #246: Expand the guide for contributors

2020-02-10 Thread GitBox
Fokko commented on a change in pull request #246: Expand the guide for 
contributors
URL: https://github.com/apache/airflow-site/pull/246#discussion_r377472874
 
 

 ##
 File path: CONTRIBUTE.md
 ##
 @@ -17,9 +17,10 @@
  under the License.
 -->
 
-## General directory structure
+Contrubitor Guide
 
 Review comment:
   ```suggestion
   Contributor Guide
   ```


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow-site] Fokko merged pull request #247: Fix typo - Commiter => Committer

2020-02-10 Thread GitBox
Fokko merged pull request #247: Fix typo - Commiter => Committer
URL: https://github.com/apache/airflow-site/pull/247
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow-site] Fokko opened a new pull request #249: Remove AIP-11 in favor of master

2020-02-10 Thread GitBox
Fokko opened a new pull request #249: Remove AIP-11 in favor of master
URL: https://github.com/apache/airflow-site/pull/249
 
 
   I believe this is a relic from the past. I tried to click the _Suggest a 
change on this page_ on here: https://airflow.apache.org/community/
   
   But I got a 404.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (AIRFLOW-6771) Error in formatting in _collect_errors

2020-02-10 Thread Gregory Golberg (Jira)
Gregory Golberg created AIRFLOW-6771:


 Summary: Error in formatting in _collect_errors
 Key: AIRFLOW-6771
 URL: https://issues.apache.org/jira/browse/AIRFLOW-6771
 Project: Apache Airflow
  Issue Type: Bug
  Components: DagRun
Affects Versions: 1.9.0
Reporter: Gregory Golberg


In airflow/jobs.py, in _collect_errors() method (on [line 
2323|https://github.com/apache/airflow/blob/v1-9-stable/airflow/jobs.py#L2323] 
the code is
{code}
@provide_session
def _collect_errors(self, ti_status, session=None):
err = ''
if ti_status.failed:
err += (
"---\n"
"Some task instances failed:\n%s".format(ti_status.failed))
{code}

But that just results in %s -- last line should be instead

{code}
 "Some task instances failed:\n{}".format(ti_status.failed))
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] jithin97 commented on a change in pull request #7396: [AIRFLOW-6770] Run particular test using breeze CLI bug fix

2020-02-10 Thread GitBox
jithin97 commented on a change in pull request #7396: [AIRFLOW-6770] Run 
particular test using breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396#discussion_r377459173
 
 

 ##
 File path: .travis.yml
 ##
 @@ -101,7 +101,7 @@ jobs:
 ENABLED_INTEGRATIONS="cassandra kerberos mongo openldap rabbitmq redis"
 RUN_INTEGRATION_TESTS=all
   stage: test
-  script: ./scripts/ci/ci_run_airflow_testing.sh tests/tests_core.py
+  script: ./scripts/ci/ci_run_airflow_testing.sh tests/test_core.py
 
 Review comment:
   fixed!!! 
https://github.com/apache/airflow/pull/7396/commits/f6be34fd8725ae44b156f0e51076d92bf69a36a0


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] jithin97 commented on a change in pull request #7396: [AIRFLOW-6770] Run particular test using breeze CLI bug fix

2020-02-10 Thread GitBox
jithin97 commented on a change in pull request #7396: [AIRFLOW-6770] Run 
particular test using breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396#discussion_r377459124
 
 

 ##
 File path: scripts/ci/in_container/entrypoint_ci.sh
 ##
 @@ -211,6 +211,10 @@ fi
 
 TEST_DIR="tests/"
 
+if [[ ${#@} -gt 0 && -n "$1" ]]; then
+TEST_DIR="$1"
+fi
+
 if [[ -n ${RUNTIME} ]]; then
 CI_ARGS+=("--runtime" "${RUNTIME}" "-rpfExX")
 TEST_DIR="tests/runtime"
 
 Review comment:
   fixed!!! 
https://github.com/apache/airflow/pull/7396/commits/f6be34fd8725ae44b156f0e51076d92bf69a36a0


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] jithin97 commented on a change in pull request #7396: [AIRFLOW-6770] Run particular test using breeze CLI bug fix

2020-02-10 Thread GitBox
jithin97 commented on a change in pull request #7396: [AIRFLOW-6770] Run 
particular test using breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396#discussion_r377459152
 
 

 ##
 File path: scripts/ci/in_container/entrypoint_ci.sh
 ##
 @@ -211,6 +211,10 @@ fi
 
 TEST_DIR="tests/"
 
 Review comment:
   fixed!!! 
https://github.com/apache/airflow/pull/7396/commits/f6be34fd8725ae44b156f0e51076d92bf69a36a0


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on issue #7396: [AIRFLOW-6770] Run particular test using breeze CLI bug fix

2020-02-10 Thread GitBox
potiuk commented on issue #7396: [AIRFLOW-6770] Run particular test using 
breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396#issuecomment-584491660
 
 
   Thanks @jithin97 !


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] codecov-io commented on issue #7396: [AIRFLOW-6770] Run particular test using breeze CLI bug fix

2020-02-10 Thread GitBox
codecov-io commented on issue #7396: [AIRFLOW-6770] Run particular test using 
breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396#issuecomment-584488896
 
 
   # [Codecov](https://codecov.io/gh/apache/airflow/pull/7396?src=pr=h1) 
Report
   > Merging 
[#7396](https://codecov.io/gh/apache/airflow/pull/7396?src=pr=desc) into 
[master](https://codecov.io/gh/apache/airflow/commit/1e00243014382d4cb7152ca7c5011b97cbd733b0?src=pr=desc)
 will **decrease** coverage by `0.14%`.
   > The diff coverage is `n/a`.
   
   [![Impacted file tree 
graph](https://codecov.io/gh/apache/airflow/pull/7396/graphs/tree.svg?width=650=WdLKlKHOAU=150=pr)](https://codecov.io/gh/apache/airflow/pull/7396?src=pr=tree)
   
   ```diff
   @@Coverage Diff @@
   ##   master#7396  +/-   ##
   ==
   - Coverage86.6%   86.46%   -0.15% 
   ==
 Files 873  873  
 Lines   4072540725  
   ==
   - Hits3527135213  -58 
   - Misses   5454 5512  +58
   ```
   
   
   | [Impacted 
Files](https://codecov.io/gh/apache/airflow/pull/7396?src=pr=tree) | 
Coverage Δ | |
   |---|---|---|
   | 
[airflow/executors/celery\_executor.py](https://codecov.io/gh/apache/airflow/pull/7396/diff?src=pr=tree#diff-YWlyZmxvdy9leGVjdXRvcnMvY2VsZXJ5X2V4ZWN1dG9yLnB5)
 | `50.67% <0%> (-37.84%)` | :arrow_down: |
   | 
[airflow/executors/base\_executor.py](https://codecov.io/gh/apache/airflow/pull/7396/diff?src=pr=tree#diff-YWlyZmxvdy9leGVjdXRvcnMvYmFzZV9leGVjdXRvci5weQ==)
 | `93.58% <0%> (-2.57%)` | :arrow_down: |
   
   --
   
   [Continue to review full report at 
Codecov](https://codecov.io/gh/apache/airflow/pull/7396?src=pr=continue).
   > **Legend** - [Click here to learn 
more](https://docs.codecov.io/docs/codecov-delta)
   > `Δ = absolute  (impact)`, `ø = not affected`, `? = missing data`
   > Powered by 
[Codecov](https://codecov.io/gh/apache/airflow/pull/7396?src=pr=footer). 
Last update 
[1e00243...8421b26](https://codecov.io/gh/apache/airflow/pull/7396?src=pr=lastupdated).
 Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments).
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on a change in pull request #7396: [AIRFLOW-6770] Run particular test using breeze CLI bug fix

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7396: [AIRFLOW-6770] Run 
particular test using breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396#discussion_r377454766
 
 

 ##
 File path: scripts/ci/in_container/entrypoint_ci.sh
 ##
 @@ -211,6 +211,10 @@ fi
 
 TEST_DIR="tests/"
 
+if [[ ${#@} -gt 0 && -n "$1" ]]; then
+TEST_DIR="$1"
+fi
+
 if [[ -n ${RUNTIME} ]]; then
 CI_ARGS+=("--runtime" "${RUNTIME}" "-rpfExX")
 TEST_DIR="tests/runtime"
 
 Review comment:
   ```suggestion
   TESTS_TO_RUN="tests/runtime"
   ```


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on a change in pull request #7396: [AIRFLOW-6770] Run particular test using breeze CLI bug fix

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7396: [AIRFLOW-6770] Run 
particular test using breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396#discussion_r377454608
 
 

 ##
 File path: scripts/ci/in_container/entrypoint_ci.sh
 ##
 @@ -211,6 +211,10 @@ fi
 
 TEST_DIR="tests/"
 
 Review comment:
   ```suggestion
   TESTS_TO_RUN="tests/"
   ```


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on a change in pull request #7396: [AIRFLOW-6770] Run particular test using breeze CLI bug fix

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7396: [AIRFLOW-6770] Run 
particular test using breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396#discussion_r377454957
 
 

 ##
 File path: scripts/ci/in_container/entrypoint_ci.sh
 ##
 @@ -211,6 +211,10 @@ fi
 
 TEST_DIR="tests/"
 
+if [[ ${#@} -gt 0 && -n "$1" ]]; then
+TEST_DIR="$1"
+fi
+
 if [[ -n ${RUNTIME} ]]; then
 CI_ARGS+=("--runtime" "${RUNTIME}" "-rpfExX")
 TEST_DIR="tests/runtime"
 
 Review comment:
   And also we should change TEST_DIR below to TESTS_TO_RUN :)


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on a change in pull request #7396: [AIRFLOW-6770] Run particular test using breeze CLI bug fix

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7396: [AIRFLOW-6770] Run 
particular test using breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396#discussion_r377454669
 
 

 ##
 File path: scripts/ci/in_container/entrypoint_ci.sh
 ##
 @@ -211,6 +211,10 @@ fi
 
 TEST_DIR="tests/"
 
 Review comment:
   I think this will be better name :)


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on a change in pull request #7396: [AIRFLOW-6770] Run particular test using breeze CLI bug fix

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7396: [AIRFLOW-6770] Run 
particular test using breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396#discussion_r377454371
 
 

 ##
 File path: .travis.yml
 ##
 @@ -101,7 +101,7 @@ jobs:
 ENABLED_INTEGRATIONS="cassandra kerberos mongo openldap rabbitmq redis"
 RUN_INTEGRATION_TESTS=all
   stage: test
-  script: ./scripts/ci/ci_run_airflow_testing.sh tests/tests_core.py
+  script: ./scripts/ci/ci_run_airflow_testing.sh tests/test_core.py
 
 Review comment:
   ```suggestion
   ```
   That's an interesting one. Actually it's a mistake here. We should not 
constraint ourselves in this test only to test_core. It is likely a remnant 
from some testing done when I was implementing it. Good that you noticed it.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (AIRFLOW-6770) Unable to run specific test using breeze CLI

2020-02-10 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-6770?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17034126#comment-17034126
 ] 

ASF GitHub Bot commented on AIRFLOW-6770:
-

jithin97 commented on pull request #7396: [AIRFLOW-6770] Run particular test 
using breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396
 
 
   ### Description
   Running a particular test from breeze CLI runs all the tests within _tests/_ 
directory instead of running only the tests specified by **test-target** flag
   
   **Before:**
   
   ```
   ⇒  ./breeze --test-target tests/test_core.py
   .
   .
   .
   Starting the tests with those pytest arguments: tests/
   
   

 test session starts 

   platform linux -- Python 3.6.10, pytest-5.3.5, py-1.8.1, pluggy-0.13.1 -- 
/usr/local/bin/python
   cachedir: .pytest_cache
   rootdir: /opt/airflow, inifile: pytest.ini
   plugins: requests-mock-1.7.0, cov-2.8.1, flaky-3.6.1, instafail-0.4.1.post0, 
celery-4.4.0
   collected 5055 items / 1 skipped / 5054 selected
   ```
   
   **After:**
   
   ```
   ⇒  ./breeze --test-target tests/test_core.py
   .
   .
   .
   Starting the tests with those pytest arguments: tests/test_core.py
   
   

 test session starts 

   platform linux -- Python 3.6.10, pytest-5.3.5, py-1.8.1, pluggy-0.13.1 -- 
/usr/local/bin/python
   cachedir: .pytest_cache
   rootdir: /opt/airflow, inifile: pytest.ini
   plugins: requests-mock-1.7.0, cov-2.8.1, flaky-3.6.1, instafail-0.4.1.post0, 
celery-4.4.0
   collected 24 items 
   ```
   ---
   Issue link: WILL BE INSERTED BY 
[boring-cyborg](https://github.com/kaxil/boring-cyborg)
   
   Make sure to mark the boxes below before creating PR: [x]
   
   - [x] Description above provides context of the change
   - [x] Commit message/PR title starts with `[AIRFLOW-]`. AIRFLOW- = 
JIRA ID*
   - [x] Unit tests coverage for changes (not needed for documentation changes)
   - [x] Commits follow "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)"
   - [x] Relevant documentation is updated including usage instructions.
   - [x] I will engage committers as explained in [Contribution Workflow 
Example](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#contribution-workflow-example).
   
   * For document-only changes commit message can start with 
`[AIRFLOW-]`.
   
   ---
   In case of fundamental code change, Airflow Improvement Proposal 
([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals))
 is needed.
   In case of a new dependency, check compliance with the [ASF 3rd Party 
License Policy](https://www.apache.org/legal/resolved.html#category-x).
   In case of backwards incompatible changes please leave a note in 
[UPDATING.md](https://github.com/apache/airflow/blob/master/UPDATING.md).
   Read the [Pull Request 
Guidelines](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#pull-request-guidelines)
 for more information.
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Unable to run specific test using breeze CLI
> 
>
> Key: AIRFLOW-6770
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6770
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: breeze
>Affects Versions: 1.10.9
>Reporter: Jithin Sukumar
>Assignee: Jithin Sukumar
>Priority: Minor
>
> {code:bash}
> ./breeze -t tests/test_core.py
> {code}
> is supposed to run only tests/test_core.py but instead runs all tests within 
> tests/ directory.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] jithin97 opened a new pull request #7396: [AIRFLOW-6770] Run particular test using breeze CLI bug fix

2020-02-10 Thread GitBox
jithin97 opened a new pull request #7396: [AIRFLOW-6770] Run particular test 
using breeze CLI bug fix
URL: https://github.com/apache/airflow/pull/7396
 
 
   ### Description
   Running a particular test from breeze CLI runs all the tests within _tests/_ 
directory instead of running only the tests specified by **test-target** flag
   
   **Before:**
   
   ```
   ⇒  ./breeze --test-target tests/test_core.py
   .
   .
   .
   Starting the tests with those pytest arguments: tests/
   
   

 test session starts 

   platform linux -- Python 3.6.10, pytest-5.3.5, py-1.8.1, pluggy-0.13.1 -- 
/usr/local/bin/python
   cachedir: .pytest_cache
   rootdir: /opt/airflow, inifile: pytest.ini
   plugins: requests-mock-1.7.0, cov-2.8.1, flaky-3.6.1, instafail-0.4.1.post0, 
celery-4.4.0
   collected 5055 items / 1 skipped / 5054 selected
   ```
   
   **After:**
   
   ```
   ⇒  ./breeze --test-target tests/test_core.py
   .
   .
   .
   Starting the tests with those pytest arguments: tests/test_core.py
   
   

 test session starts 

   platform linux -- Python 3.6.10, pytest-5.3.5, py-1.8.1, pluggy-0.13.1 -- 
/usr/local/bin/python
   cachedir: .pytest_cache
   rootdir: /opt/airflow, inifile: pytest.ini
   plugins: requests-mock-1.7.0, cov-2.8.1, flaky-3.6.1, instafail-0.4.1.post0, 
celery-4.4.0
   collected 24 items 
   ```
   ---
   Issue link: WILL BE INSERTED BY 
[boring-cyborg](https://github.com/kaxil/boring-cyborg)
   
   Make sure to mark the boxes below before creating PR: [x]
   
   - [x] Description above provides context of the change
   - [x] Commit message/PR title starts with `[AIRFLOW-]`. AIRFLOW- = 
JIRA ID*
   - [x] Unit tests coverage for changes (not needed for documentation changes)
   - [x] Commits follow "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)"
   - [x] Relevant documentation is updated including usage instructions.
   - [x] I will engage committers as explained in [Contribution Workflow 
Example](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#contribution-workflow-example).
   
   * For document-only changes commit message can start with 
`[AIRFLOW-]`.
   
   ---
   In case of fundamental code change, Airflow Improvement Proposal 
([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals))
 is needed.
   In case of a new dependency, check compliance with the [ASF 3rd Party 
License Policy](https://www.apache.org/legal/resolved.html#category-x).
   In case of backwards incompatible changes please leave a note in 
[UPDATING.md](https://github.com/apache/airflow/blob/master/UPDATING.md).
   Read the [Pull Request 
Guidelines](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#pull-request-guidelines)
 for more information.
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (AIRFLOW-6770) Unable to run specific test using breeze CLI

2020-02-10 Thread Jithin Sukumar (Jira)
Jithin Sukumar created AIRFLOW-6770:
---

 Summary: Unable to run specific test using breeze CLI
 Key: AIRFLOW-6770
 URL: https://issues.apache.org/jira/browse/AIRFLOW-6770
 Project: Apache Airflow
  Issue Type: Bug
  Components: breeze
Affects Versions: 1.10.9
Reporter: Jithin Sukumar
Assignee: Jithin Sukumar


{code:bash}
./breeze -t tests/test_core.py
{code}
is supposed to run only tests/test_core.py but instead runs all tests within 
tests/ directory.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] yuqian90 removed a comment on issue #7276: [AIRFLOW-5391] Do not run skipped tasks when they are cleared

2020-02-10 Thread GitBox
yuqian90 removed a comment on issue #7276: [AIRFLOW-5391] Do not run skipped 
tasks when they are cleared
URL: https://github.com/apache/airflow/pull/7276#issuecomment-579271946
 
 
   This is work-in-progress and untested. I want to get some opinions from 
committers first to see if it's worth pursuing. I'll then add tests and make 
the PR complete.
   
   One immediate shortcoming that I can see:
   This PR only makes skipped tasks respect the decision of upstream 
`BranchPythonOperator` when cleared. Other variants of branching exist, namely 
`BaseBranchOperator` and `ShortCircuitOperator`. This PR does not make tasks 
respect the decision of those operators when cleared.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (AIRFLOW-5629) Support specify priority and priorityClassName spec in KubernetesPodOperator

2020-02-10 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-5629?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17034018#comment-17034018
 ] 

ASF GitHub Bot commented on AIRFLOW-5629:
-

simis2626 commented on pull request #7395: [AIRFLOW-5629] Implement K8s 
priorityClassName in KubernetesPo…
URL: https://github.com/apache/airflow/pull/7395
 
 
   …dOperator
   
   Changes allow user to specify priority_class_name within 
KubernetesPodOperator
   which sets the Pod.spec value to influence scheduling and preemption 
behaviour
   of workloads generated by Airflow.
   
   pre-commit hook issue with BREEZE.rst adding whitespace - Can't push without 
committing whitespace change.
   
   ---
   Issue link: WILL BE INSERTED BY 
[boring-cyborg](https://github.com/kaxil/boring-cyborg)
   
   Make sure to mark the boxes below before creating PR: [x]
   
   - [x] Description above provides context of the change
   - [x] Commit message/PR title starts with `[AIRFLOW-]`. AIRFLOW- = 
JIRA ID*
   - [ ] Unit tests coverage for changes (not needed for documentation changes)
   - [x] Commits follow "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)"
   - [x] Relevant documentation is updated including usage instructions.
   - [x] I will engage committers as explained in [Contribution Workflow 
Example](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#contribution-workflow-example).
   
   * For document-only changes commit message can start with 
`[AIRFLOW-]`.
   
   ---
   In case of fundamental code change, Airflow Improvement Proposal 
([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals))
 is needed.
   In case of a new dependency, check compliance with the [ASF 3rd Party 
License Policy](https://www.apache.org/legal/resolved.html#category-x).
   In case of backwards incompatible changes please leave a note in 
[UPDATING.md](https://github.com/apache/airflow/blob/master/UPDATING.md).
   Read the [Pull Request 
Guidelines](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#pull-request-guidelines)
 for more information.
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Support specify priority and priorityClassName spec in KubernetesPodOperator
> 
>
> Key: AIRFLOW-5629
> URL: https://issues.apache.org/jira/browse/AIRFLOW-5629
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: operators
>Affects Versions: 2.0.0
>Reporter: Yaliang Wang
>Assignee: Simon Levett
>Priority: Major
>  Labels: features
>
> Since Kubernetes 1.14, Pod can have priority and priority class to indicate 
> the importance of the 
> pod([https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/]).
>  We should support to config the priority and priority class in the 
> KubernetesPodOperator 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] simis2626 opened a new pull request #7395: [AIRFLOW-5629] Implement K8s priorityClassName in KubernetesPo…

2020-02-10 Thread GitBox
simis2626 opened a new pull request #7395: [AIRFLOW-5629] Implement K8s 
priorityClassName in KubernetesPo…
URL: https://github.com/apache/airflow/pull/7395
 
 
   …dOperator
   
   Changes allow user to specify priority_class_name within 
KubernetesPodOperator
   which sets the Pod.spec value to influence scheduling and preemption 
behaviour
   of workloads generated by Airflow.
   
   pre-commit hook issue with BREEZE.rst adding whitespace - Can't push without 
committing whitespace change.
   
   ---
   Issue link: WILL BE INSERTED BY 
[boring-cyborg](https://github.com/kaxil/boring-cyborg)
   
   Make sure to mark the boxes below before creating PR: [x]
   
   - [x] Description above provides context of the change
   - [x] Commit message/PR title starts with `[AIRFLOW-]`. AIRFLOW- = 
JIRA ID*
   - [ ] Unit tests coverage for changes (not needed for documentation changes)
   - [x] Commits follow "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)"
   - [x] Relevant documentation is updated including usage instructions.
   - [x] I will engage committers as explained in [Contribution Workflow 
Example](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#contribution-workflow-example).
   
   * For document-only changes commit message can start with 
`[AIRFLOW-]`.
   
   ---
   In case of fundamental code change, Airflow Improvement Proposal 
([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals))
 is needed.
   In case of a new dependency, check compliance with the [ASF 3rd Party 
License Policy](https://www.apache.org/legal/resolved.html#category-x).
   In case of backwards incompatible changes please leave a note in 
[UPDATING.md](https://github.com/apache/airflow/blob/master/UPDATING.md).
   Read the [Pull Request 
Guidelines](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#pull-request-guidelines)
 for more information.
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] boring-cyborg[bot] commented on issue #7395: [AIRFLOW-5629] Implement K8s priorityClassName in KubernetesPo…

2020-02-10 Thread GitBox
boring-cyborg[bot] commented on issue #7395: [AIRFLOW-5629] Implement K8s 
priorityClassName in KubernetesPo…
URL: https://github.com/apache/airflow/pull/7395#issuecomment-584409286
 
 
   Congratulations on your first Pull Request and welcome to the Apache Airflow 
community! If you have any issues or are unsure about any anything please check 
our Contribution Guide 
(https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst)
   Here are some useful points:
   - Pay attention to the quality of your code (flake8, pylint and type 
annotations). Our [pre-commits]( 
https://github.com/apache/airflow/blob/master/STATIC_CODE_CHECKS.rst#prerequisites-for-pre-commit-hooks)
 will help you with that.
   - In case of a new feature add useful documentation (in docstrings or in 
`docs/` directory). Adding a new operator? Check this short 
[guide](https://github.com/apache/airflow/blob/master/docs/howto/custom-operator.rst)
 Consider adding an example DAG that shows how users should use it.
   - Consider using [Breeze 
environment](https://github.com/apache/airflow/blob/master/BREEZE.rst) for 
testing locally, it’s a heavy docker but it ships with a working Airflow and a 
lot of integrations.
   - Be patient and persistent. It might take some time to get a review or get 
the final approval from Committers.
   Apache Airflow is a community-driven project and together we are making it 
better .
   In case of doubts contact the developers at:
   Mailing List: d...@airflow.apache.org
   Slack: https://apache-airflow-slack.herokuapp.com/
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Assigned] (AIRFLOW-5629) Support specify priority and priorityClassName spec in KubernetesPodOperator

2020-02-10 Thread Simon Levett (Jira)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-5629?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Levett reassigned AIRFLOW-5629:
-

Assignee: Simon Levett  (was: Yaliang Wang)

Working on a PR for this.

> Support specify priority and priorityClassName spec in KubernetesPodOperator
> 
>
> Key: AIRFLOW-5629
> URL: https://issues.apache.org/jira/browse/AIRFLOW-5629
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: operators
>Affects Versions: 2.0.0
>Reporter: Yaliang Wang
>Assignee: Simon Levett
>Priority: Major
>  Labels: features
>
> Since Kubernetes 1.14, Pod can have priority and priority class to indicate 
> the importance of the 
> pod([https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/]).
>  We should support to config the priority and priority class in the 
> KubernetesPodOperator 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] stale[bot] closed pull request #2708: [AIRFLOW-1746] Add a Nomad operator to trigger job from Airflow

2020-02-10 Thread GitBox
stale[bot] closed pull request #2708: [AIRFLOW-1746] Add a Nomad operator to 
trigger job from Airflow
URL: https://github.com/apache/airflow/pull/2708
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (AIRFLOW-1746) Add a Nomad operator to trigger job from Airflow

2020-02-10 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-1746?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033986#comment-17033986
 ] 

ASF GitHub Bot commented on AIRFLOW-1746:
-

stale[bot] commented on pull request #2708: [AIRFLOW-1746] Add a Nomad operator 
to trigger job from Airflow
URL: https://github.com/apache/airflow/pull/2708
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Add a Nomad operator to trigger job from Airflow
> 
>
> Key: AIRFLOW-1746
> URL: https://issues.apache.org/jira/browse/AIRFLOW-1746
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: contrib
>Reporter: Eyal Trabelsi
>Assignee: Eyal Trabelsi
>Priority: Major
>
> We recently face the need to trigger nomad jobs from Airflow and no operator 
> are available for that. 
> The way the operator works is to register a nomad job and dispatch the job , 
> than check the status of the job using similar method like boto-core 
> (https://github.com/boto/botocore/blob/5a07b477114b11e6dc5f676f5db810972565b113/botocore/docs/waiter.py)
> The operator uses https://github.com/jrxFive/python-nomad which is a wrap 
> over nomad rest api of nomad written in python.
> Link to the PR : https://github.com/apache/incubator-airflow/pull/2708



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] spektom commented on issue #5785: [AIRFLOW-5176] Add Azure Data Explorer (Kusto) operator

2020-02-10 Thread GitBox
spektom commented on issue #5785: [AIRFLOW-5176] Add Azure Data Explorer 
(Kusto) operator
URL: https://github.com/apache/airflow/pull/5785#issuecomment-584335178
 
 
   > Great work @spektom, sorry it took 6 months to land!
   
   No problem, thanks for reviewing and helping shape it!


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Resolved] (AIRFLOW-5176) Add integration with Azure Data Explorer

2020-02-10 Thread Ash Berlin-Taylor (Jira)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-5176?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ash Berlin-Taylor resolved AIRFLOW-5176.

Fix Version/s: 2.0.0
   Resolution: Fixed

> Add integration with Azure Data Explorer
> 
>
> Key: AIRFLOW-5176
> URL: https://issues.apache.org/jira/browse/AIRFLOW-5176
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: hooks, operators
>Affects Versions: 1.10.4, 2.0.0
>Reporter: Michael Spector
>Assignee: Michael Spector
>Priority: Major
> Fix For: 2.0.0
>
>
> Add a hook and an operator that allow sending queries to Azure Data Explorer 
> (Kusto) cluster.
> ADX (Azure Data Explorer) is relatively new but very promising analytics data 
> store / data processing offering in Azure:
> [https://azure.microsoft.com/en-us/services/data-explorer/]
>  
> PR: [https://github.com/apache/airflow/pull/5785]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (AIRFLOW-5176) Add integration with Azure Data Explorer

2020-02-10 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-5176?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033921#comment-17033921
 ] 

ASF GitHub Bot commented on AIRFLOW-5176:
-

ashb commented on pull request #5785: [AIRFLOW-5176] Add Azure Data Explorer 
(Kusto) operator
URL: https://github.com/apache/airflow/pull/5785
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Add integration with Azure Data Explorer
> 
>
> Key: AIRFLOW-5176
> URL: https://issues.apache.org/jira/browse/AIRFLOW-5176
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: hooks, operators
>Affects Versions: 1.10.4, 2.0.0
>Reporter: Michael Spector
>Assignee: Michael Spector
>Priority: Major
>
> Add a hook and an operator that allow sending queries to Azure Data Explorer 
> (Kusto) cluster.
> ADX (Azure Data Explorer) is relatively new but very promising analytics data 
> store / data processing offering in Azure:
> [https://azure.microsoft.com/en-us/services/data-explorer/]
>  
> PR: [https://github.com/apache/airflow/pull/5785]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (AIRFLOW-5176) Add integration with Azure Data Explorer

2020-02-10 Thread ASF subversion and git services (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-5176?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033922#comment-17033922
 ] 

ASF subversion and git services commented on AIRFLOW-5176:
--

Commit 1e00243014382d4cb7152ca7c5011b97cbd733b0 in airflow's branch 
refs/heads/master from Michael Spector
[ https://gitbox.apache.org/repos/asf?p=airflow.git;h=1e00243 ]

[AIRFLOW-5176] Add Azure Data Explorer (Kusto) operator (#5785)



> Add integration with Azure Data Explorer
> 
>
> Key: AIRFLOW-5176
> URL: https://issues.apache.org/jira/browse/AIRFLOW-5176
> Project: Apache Airflow
>  Issue Type: New Feature
>  Components: hooks, operators
>Affects Versions: 1.10.4, 2.0.0
>Reporter: Michael Spector
>Assignee: Michael Spector
>Priority: Major
> Fix For: 2.0.0
>
>
> Add a hook and an operator that allow sending queries to Azure Data Explorer 
> (Kusto) cluster.
> ADX (Azure Data Explorer) is relatively new but very promising analytics data 
> store / data processing offering in Azure:
> [https://azure.microsoft.com/en-us/services/data-explorer/]
>  
> PR: [https://github.com/apache/airflow/pull/5785]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] boring-cyborg[bot] commented on issue #5785: [AIRFLOW-5176] Add Azure Data Explorer (Kusto) operator

2020-02-10 Thread GitBox
boring-cyborg[bot] commented on issue #5785: [AIRFLOW-5176] Add Azure Data 
Explorer (Kusto) operator
URL: https://github.com/apache/airflow/pull/5785#issuecomment-584334338
 
 
   Awesome work, congrats on your first merged pull request!
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] ashb commented on issue #5785: [AIRFLOW-5176] Add Azure Data Explorer (Kusto) operator

2020-02-10 Thread GitBox
ashb commented on issue #5785: [AIRFLOW-5176] Add Azure Data Explorer (Kusto) 
operator
URL: https://github.com/apache/airflow/pull/5785#issuecomment-584334509
 
 
   Great work @spektom, sorry it took 6 months to land!


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] ashb merged pull request #5785: [AIRFLOW-5176] Add Azure Data Explorer (Kusto) operator

2020-02-10 Thread GitBox
ashb merged pull request #5785: [AIRFLOW-5176] Add Azure Data Explorer (Kusto) 
operator
URL: https://github.com/apache/airflow/pull/5785
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (AIRFLOW-6769) Multiple DAGs cause infinite loop in logging

2020-02-10 Thread Chris Boden (Jira)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-6769?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Boden updated AIRFLOW-6769:
-
Description: 
After adding a second DAG to Airflow the scheduler spirals into an infinite 
loop and is unable to schedule any tasks. The recursion appears to have to do 
with logging. I've attached the output from scheduler and my requirements.txt

If I reset Airflow (`airflow resetdb`) and remove all but 1 DAG from the folder 
there is no issue. I don't believe it's a specific DAG as I have tried with 
various different test ones. I don't even have to schedule/run them, just 
having Airflow recognize them causes the loop. 

I'm using MySQL and Celery as my store and worker. 

I have also attached part of the log file for one of my DAGs outputted by the 
scheduler ($AIRFLOW_HOME/logs/scheduler/\{dag_name}/latest/\{dag_name}.log). It 
seems `collect_dags` is being called every second by the scheduler. 

  was:
After adding a second DAG to Airflow the scheduler spirals into an infinite 
loop and is unable to schedule any tasks. The recursion appears to have to do 
with logging. I've attached the output from scheduler and my requirements.txt

If I reset Airflow (`airflow resetdb`) and remove all but 1 DAG from the folder 
there is no issue. I don't believe it's a specific DAG as I have tried with 
various different test ones. I don't even have to schedule/run them, just 
having Airflow recognize them causes the loop. 

I'm using MySQL and Celery as my store and worker. 


> Multiple DAGs cause infinite loop in logging
> 
>
> Key: AIRFLOW-6769
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6769
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: scheduler
>Affects Versions: 1.10.9
> Environment: macOS 10.15.2
> Python 3.7.5
>Reporter: Chris Boden
>Priority: Major
> Attachments: requirements.txt, scheduler-dag.log, scheduler.log
>
>
> After adding a second DAG to Airflow the scheduler spirals into an infinite 
> loop and is unable to schedule any tasks. The recursion appears to have to do 
> with logging. I've attached the output from scheduler and my requirements.txt
> If I reset Airflow (`airflow resetdb`) and remove all but 1 DAG from the 
> folder there is no issue. I don't believe it's a specific DAG as I have tried 
> with various different test ones. I don't even have to schedule/run them, 
> just having Airflow recognize them causes the loop. 
> I'm using MySQL and Celery as my store and worker. 
> I have also attached part of the log file for one of my DAGs outputted by the 
> scheduler ($AIRFLOW_HOME/logs/scheduler/\{dag_name}/latest/\{dag_name}.log). 
> It seems `collect_dags` is being called every second by the scheduler. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (AIRFLOW-6769) Multiple DAGs cause infinite loop in logging

2020-02-10 Thread Chris Boden (Jira)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-6769?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Boden updated AIRFLOW-6769:
-
Attachment: scheduler-dag.log

> Multiple DAGs cause infinite loop in logging
> 
>
> Key: AIRFLOW-6769
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6769
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: scheduler
>Affects Versions: 1.10.9
> Environment: macOS 10.15.2
> Python 3.7.5
>Reporter: Chris Boden
>Priority: Major
> Attachments: requirements.txt, scheduler-dag.log, scheduler.log
>
>
> After adding a second DAG to Airflow the scheduler spirals into an infinite 
> loop and is unable to schedule any tasks. The recursion appears to have to do 
> with logging. I've attached the output from scheduler and my requirements.txt
> If I reset Airflow (`airflow resetdb`) and remove all but 1 DAG from the 
> folder there is no issue. I don't believe it's a specific DAG as I have tried 
> with various different test ones. I don't even have to schedule/run them, 
> just having Airflow recognize them causes the loop. 
> I'm using MySQL and Celery as my store and worker. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (AIRFLOW-6769) Multiple DAGs cause infinite loop in logging

2020-02-10 Thread Chris Boden (Jira)
Chris Boden created AIRFLOW-6769:


 Summary: Multiple DAGs cause infinite loop in logging
 Key: AIRFLOW-6769
 URL: https://issues.apache.org/jira/browse/AIRFLOW-6769
 Project: Apache Airflow
  Issue Type: Bug
  Components: scheduler
Affects Versions: 1.10.9
 Environment: macOS 10.15.2
Python 3.7.5
Reporter: Chris Boden
 Attachments: requirements.txt, scheduler.log

After adding a second DAG to Airflow the scheduler spirals into an infinite 
loop and is unable to schedule any tasks. The recursion appears to have to do 
with logging. I've attached the output from scheduler and my requirements.txt

If I reset Airflow (`airflow resetdb`) and remove all but 1 DAG from the folder 
there is no issue. I don't believe it's a specific DAG as I have tried with 
various different test ones. I don't even have to schedule/run them, just 
having Airflow recognize them causes the loop. 

I'm using MySQL and Celery as my store and worker. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] stale[bot] commented on issue #6256: [AIRFLOW-5590] Add run_id to trigger DAG run API response

2020-02-10 Thread GitBox
stale[bot] commented on issue #6256: [AIRFLOW-5590] Add run_id to trigger DAG 
run API response
URL: https://github.com/apache/airflow/pull/6256#issuecomment-584258610
 
 
   This issue has been automatically marked as stale because it has not had 
recent activity. It will be closed if no further activity occurs. Thank you for 
your contributions.
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] blcksrx commented on issue #7358: [AIRFLOW-1202] Create Elasticsearch Hook

2020-02-10 Thread GitBox
blcksrx commented on issue #7358: [AIRFLOW-1202] Create Elasticsearch Hook
URL: https://github.com/apache/airflow/pull/7358#issuecomment-584257515
 
 
   Done


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] codecov-io edited a comment on issue #7358: [AIRFLOW-1202] Create Elasticsearch Hook

2020-02-10 Thread GitBox
codecov-io edited a comment on issue #7358: [AIRFLOW-1202] Create Elasticsearch 
Hook
URL: https://github.com/apache/airflow/pull/7358#issuecomment-581934830
 
 
   # [Codecov](https://codecov.io/gh/apache/airflow/pull/7358?src=pr=h1) 
Report
   > Merging 
[#7358](https://codecov.io/gh/apache/airflow/pull/7358?src=pr=desc) into 
[master](https://codecov.io/gh/apache/airflow/commit/319714f07c470bca999b2f6becbb340c81e79e29?src=pr=desc)
 will **decrease** coverage by `0.46%`.
   > The diff coverage is `47.5%`.
   
   [![Impacted file tree 
graph](https://codecov.io/gh/apache/airflow/pull/7358/graphs/tree.svg?width=650=WdLKlKHOAU=150=pr)](https://codecov.io/gh/apache/airflow/pull/7358?src=pr=tree)
   
   ```diff
   @@Coverage Diff @@
   ##   master#7358  +/-   ##
   ==
   - Coverage   86.58%   86.12%   -0.47% 
   ==
 Files 871  872   +1 
 Lines   4066040700  +40 
   ==
   - Hits3520735053 -154 
   - Misses   5453 5647 +194
   ```
   
   
   | [Impacted 
Files](https://codecov.io/gh/apache/airflow/pull/7358?src=pr=tree) | 
Coverage Δ | |
   |---|---|---|
   | 
[airflow/models/connection.py](https://codecov.io/gh/apache/airflow/pull/7358/diff?src=pr=tree#diff-YWlyZmxvdy9tb2RlbHMvY29ubmVjdGlvbi5weQ==)
 | `95.07% <ø> (ø)` | :arrow_up: |
   | 
[airflow/utils/db.py](https://codecov.io/gh/apache/airflow/pull/7358/diff?src=pr=tree#diff-YWlyZmxvdy91dGlscy9kYi5weQ==)
 | `98.26% <100%> (+0.01%)` | :arrow_up: |
   | 
[...low/providers/elasticsearch/hooks/elasticsearch.py](https://codecov.io/gh/apache/airflow/pull/7358/diff?src=pr=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvZWxhc3RpY3NlYXJjaC9ob29rcy9lbGFzdGljc2VhcmNoLnB5)
 | `46.15% <46.15%> (ø)` | |
   | 
[...w/providers/apache/hive/operators/mysql\_to\_hive.py](https://codecov.io/gh/apache/airflow/pull/7358/diff?src=pr=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvYXBhY2hlL2hpdmUvb3BlcmF0b3JzL215c3FsX3RvX2hpdmUucHk=)
 | `100% <0%> (ø)` | :arrow_up: |
   | 
[airflow/operators/generic\_transfer.py](https://codecov.io/gh/apache/airflow/pull/7358/diff?src=pr=tree#diff-YWlyZmxvdy9vcGVyYXRvcnMvZ2VuZXJpY190cmFuc2Zlci5weQ==)
 | `100% <0%> (ø)` | :arrow_up: |
   | 
[airflow/kubernetes/volume\_mount.py](https://codecov.io/gh/apache/airflow/pull/7358/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3ZvbHVtZV9tb3VudC5weQ==)
 | `44.44% <0%> (-55.56%)` | :arrow_down: |
   | 
[airflow/providers/postgres/operators/postgres.py](https://codecov.io/gh/apache/airflow/pull/7358/diff?src=pr=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvcG9zdGdyZXMvb3BlcmF0b3JzL3Bvc3RncmVzLnB5)
 | `100% <0%> (ø)` | :arrow_up: |
   | 
[airflow/kubernetes/volume.py](https://codecov.io/gh/apache/airflow/pull/7358/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3ZvbHVtZS5weQ==)
 | `52.94% <0%> (-47.06%)` | :arrow_down: |
   | 
[airflow/security/kerberos.py](https://codecov.io/gh/apache/airflow/pull/7358/diff?src=pr=tree#diff-YWlyZmxvdy9zZWN1cml0eS9rZXJiZXJvcy5weQ==)
 | `76.08% <0%> (ø)` | :arrow_up: |
   | 
[airflow/kubernetes/pod\_launcher.py](https://codecov.io/gh/apache/airflow/pull/7358/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3BvZF9sYXVuY2hlci5weQ==)
 | `47.18% <0%> (-45.08%)` | :arrow_down: |
   | ... and [14 
more](https://codecov.io/gh/apache/airflow/pull/7358/diff?src=pr=tree-more) 
| |
   
   --
   
   [Continue to review full report at 
Codecov](https://codecov.io/gh/apache/airflow/pull/7358?src=pr=continue).
   > **Legend** - [Click here to learn 
more](https://docs.codecov.io/docs/codecov-delta)
   > `Δ = absolute  (impact)`, `ø = not affected`, `? = missing data`
   > Powered by 
[Codecov](https://codecov.io/gh/apache/airflow/pull/7358?src=pr=footer). 
Last update 
[319714f...b168505](https://codecov.io/gh/apache/airflow/pull/7358?src=pr=lastupdated).
 Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments).
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Assigned] (AIRFLOW-6369) clear cli command needs a 'conf' option

2020-02-10 Thread Yajna Pandith (Jira)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-6369?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yajna Pandith reassigned AIRFLOW-6369:
--

Assignee: Yajna Pandith

> clear cli command needs a 'conf' option
> ---
>
> Key: AIRFLOW-6369
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6369
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: cli, core, DagRun
>Affects Versions: 1.10.6
>Reporter: t oo
>Assignee: Yajna Pandith
>Priority: Major
>  Labels: gsoc, gsoc2020, mentor
>
> key-value pairs of conf can be passed into trigger_dag command
> ie
> --conf '{"ric":"amzn"}'
> clear command needs this feature too
> ie in case exec_date is important and there was a failure halfway in the 1st 
> dagrun due to bad conf being sent on trigger_dag command and want to run the 
> same execdate but with new conf on 2nd dagrun
> alternative solution would be a new delete_dag_run cli command so never need 
> to 'clear' but can do a 2nd DagRun for same exec date



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (AIRFLOW-6767) Default workgroup name for Athena is incorrect

2020-02-10 Thread ASF subversion and git services (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-6767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033743#comment-17033743
 ] 

ASF subversion and git services commented on AIRFLOW-6767:
--

Commit b7aa778b38df2f116a1c20031e72fea8b97315bf in airflow's branch 
refs/heads/master from Bhavika Tekwani
[ https://gitbox.apache.org/repos/asf?p=airflow.git;h=b7aa778 ]

[AIRFLOW-6767] Correct name for default Athena workgroup (#7394)

* Change default wg name

* Change workgroup in tests


> Default workgroup name for Athena is incorrect
> --
>
> Key: AIRFLOW-6767
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6767
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: aws
>Affects Versions: 1.10.9
>Reporter: Bhavika
>Assignee: Bhavika
>Priority: Minor
>  Labels: easy-fix, pull-request-available
> Fix For: 1.10.10
>
>
> The default Athena workgroup name is set to "default" when it should be 
> "primary". See this comment - 
> [https://github.com/apache/airflow/pull/6871#issuecomment-584066252]
> I'll submit a fix for the incorrect name. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (AIRFLOW-6767) Default workgroup name for Athena is incorrect

2020-02-10 Thread Kamil Bregula (Jira)


 [ 
https://issues.apache.org/jira/browse/AIRFLOW-6767?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kamil Bregula resolved AIRFLOW-6767.

Resolution: Fixed

> Default workgroup name for Athena is incorrect
> --
>
> Key: AIRFLOW-6767
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6767
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: aws
>Affects Versions: 1.10.9
>Reporter: Bhavika
>Assignee: Bhavika
>Priority: Minor
>  Labels: easy-fix, pull-request-available
> Fix For: 1.10.10
>
>
> The default Athena workgroup name is set to "default" when it should be 
> "primary". See this comment - 
> [https://github.com/apache/airflow/pull/6871#issuecomment-584066252]
> I'll submit a fix for the incorrect name. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (AIRFLOW-6767) Default workgroup name for Athena is incorrect

2020-02-10 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-6767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033741#comment-17033741
 ] 

ASF GitHub Bot commented on AIRFLOW-6767:
-

mik-laj commented on pull request #7394: [AIRFLOW-6767] Correct name for 
default Athena workgroup
URL: https://github.com/apache/airflow/pull/7394
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Default workgroup name for Athena is incorrect
> --
>
> Key: AIRFLOW-6767
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6767
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: aws
>Affects Versions: 1.10.9
>Reporter: Bhavika
>Assignee: Bhavika
>Priority: Minor
>  Labels: easy-fix, pull-request-available
> Fix For: 1.10.10
>
>
> The default Athena workgroup name is set to "default" when it should be 
> "primary". See this comment - 
> [https://github.com/apache/airflow/pull/6871#issuecomment-584066252]
> I'll submit a fix for the incorrect name. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] mik-laj merged pull request #7394: [AIRFLOW-6767] Correct name for default Athena workgroup

2020-02-10 Thread GitBox
mik-laj merged pull request #7394: [AIRFLOW-6767] Correct name for default 
Athena workgroup
URL: https://github.com/apache/airflow/pull/7394
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] leonardoam commented on issue #6337: [AIRFLOW-5659] - Add support for ephemeral storage on KubernetesPodOp…

2020-02-10 Thread GitBox
leonardoam commented on issue #6337: [AIRFLOW-5659] - Add support for ephemeral 
storage on KubernetesPodOp…
URL: https://github.com/apache/airflow/pull/6337#issuecomment-584208987
 
 
   @potiuk branch rebased. Thx for your help!


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (AIRFLOW-5615) BaseJob subclasses shouldn't implement own heartbeat logic

2020-02-10 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-5615?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033725#comment-17033725
 ] 

ASF GitHub Bot commented on AIRFLOW-5615:
-

ashb commented on pull request #6311: [AIRFLOW-5615] Reduce duplicated logic 
around job heartbeating
URL: https://github.com/apache/airflow/pull/6311
 
 
   ### Jira
   
   - [x] https://issues.apache.org/jira/browse/AIRFLOW-5615
   
   ### Description
   
   - [x] Both SchedulerJob and LocalTaskJob have their own timers and decide 
when
   to call heartbeat based upon that. THis makes those jobs harder to
   follow, so I've moved the logic to BaseJob
   
   ### Tests
   
   - [x] Tests added
   
   ### Commits
   
   - [x] My commits all reference Jira issues in their subject lines, and I 
have squashed multiple commits if they address the same issue. In addition, my 
commits follow the guidelines from "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)":
 1. Subject is separated from body by a blank line
 1. Subject is limited to 50 characters (not including Jira issue reference)
 1. Subject does not end with a period
 1. Subject uses the imperative mood ("add", not "adding")
 1. Body wraps at 72 characters
 1. Body explains "what" and "why", not "how"
   
   ### Documentation
   
   - [x] In case of new functionality, my PR adds documentation that describes 
how to use it.
 - All the public functions and the classes in the PR contain docstrings 
that explain what it does
 - If you implement backwards incompatible changes, please leave a note in 
the [Updating.md](https://github.com/apache/airflow/blob/master/UPDATING.md) so 
we can assign it to a appropriate release
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> BaseJob subclasses shouldn't implement own heartbeat logic
> --
>
> Key: AIRFLOW-5615
> URL: https://issues.apache.org/jira/browse/AIRFLOW-5615
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: core
>Affects Versions: 1.10.5
>Reporter: Ash Berlin-Taylor
>Assignee: Ash Berlin-Taylor
>Priority: Trivial
> Fix For: 2.0.0
>
>
> Both SchedulerJob and LocalTaskJob have their own timers and decide when to 
> call heartbeat based upon that.
> That logic should be removed and live in BaseJob to simplify the code.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (AIRFLOW-5615) BaseJob subclasses shouldn't implement own heartbeat logic

2020-02-10 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-5615?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033724#comment-17033724
 ] 

ASF GitHub Bot commented on AIRFLOW-5615:
-

stale[bot] commented on pull request #6311: [AIRFLOW-5615] Reduce duplicated 
logic around job heartbeating
URL: https://github.com/apache/airflow/pull/6311
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> BaseJob subclasses shouldn't implement own heartbeat logic
> --
>
> Key: AIRFLOW-5615
> URL: https://issues.apache.org/jira/browse/AIRFLOW-5615
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: core
>Affects Versions: 1.10.5
>Reporter: Ash Berlin-Taylor
>Assignee: Ash Berlin-Taylor
>Priority: Trivial
> Fix For: 2.0.0
>
>
> Both SchedulerJob and LocalTaskJob have their own timers and decide when to 
> call heartbeat based upon that.
> That logic should be removed and live in BaseJob to simplify the code.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] ashb opened a new pull request #6311: [AIRFLOW-5615] Reduce duplicated logic around job heartbeating

2020-02-10 Thread GitBox
ashb opened a new pull request #6311: [AIRFLOW-5615] Reduce duplicated logic 
around job heartbeating
URL: https://github.com/apache/airflow/pull/6311
 
 
   ### Jira
   
   - [x] https://issues.apache.org/jira/browse/AIRFLOW-5615
   
   ### Description
   
   - [x] Both SchedulerJob and LocalTaskJob have their own timers and decide 
when
   to call heartbeat based upon that. THis makes those jobs harder to
   follow, so I've moved the logic to BaseJob
   
   ### Tests
   
   - [x] Tests added
   
   ### Commits
   
   - [x] My commits all reference Jira issues in their subject lines, and I 
have squashed multiple commits if they address the same issue. In addition, my 
commits follow the guidelines from "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)":
 1. Subject is separated from body by a blank line
 1. Subject is limited to 50 characters (not including Jira issue reference)
 1. Subject does not end with a period
 1. Subject uses the imperative mood ("add", not "adding")
 1. Body wraps at 72 characters
 1. Body explains "what" and "why", not "how"
   
   ### Documentation
   
   - [x] In case of new functionality, my PR adds documentation that describes 
how to use it.
 - All the public functions and the classes in the PR contain docstrings 
that explain what it does
 - If you implement backwards incompatible changes, please leave a note in 
the [Updating.md](https://github.com/apache/airflow/blob/master/UPDATING.md) so 
we can assign it to a appropriate release


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] stale[bot] closed pull request #6311: [AIRFLOW-5615] Reduce duplicated logic around job heartbeating

2020-02-10 Thread GitBox
stale[bot] closed pull request #6311: [AIRFLOW-5615] Reduce duplicated logic 
around job heartbeating
URL: https://github.com/apache/airflow/pull/6311
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] codecov-io commented on issue #7391: [AIRFLOW-6663] Prepare backporting packages

2020-02-10 Thread GitBox
codecov-io commented on issue #7391: [AIRFLOW-6663] Prepare backporting packages
URL: https://github.com/apache/airflow/pull/7391#issuecomment-584201606
 
 
   # [Codecov](https://codecov.io/gh/apache/airflow/pull/7391?src=pr=h1) 
Report
   > Merging 
[#7391](https://codecov.io/gh/apache/airflow/pull/7391?src=pr=desc) into 
[master](https://codecov.io/gh/apache/airflow/commit/53ba6d607c9a3548a996c0c716d1ffb8da0b291b?src=pr=desc)
 will **increase** coverage by `0.28%`.
   > The diff coverage is `n/a`.
   
   [![Impacted file tree 
graph](https://codecov.io/gh/apache/airflow/pull/7391/graphs/tree.svg?width=650=WdLKlKHOAU=150=pr)](https://codecov.io/gh/apache/airflow/pull/7391?src=pr=tree)
   
   ```diff
   @@Coverage Diff @@
   ##   master#7391  +/-   ##
   ==
   + Coverage86.3%   86.58%   +0.28% 
   ==
 Files 871  871  
 Lines   4066040660  
   ==
   + Hits3509135207 +116 
   + Misses   5569 5453 -116
   ```
   
   
   | [Impacted 
Files](https://codecov.io/gh/apache/airflow/pull/7391?src=pr=tree) | 
Coverage Δ | |
   |---|---|---|
   | 
[...viders/cncf/kubernetes/operators/kubernetes\_pod.py](https://codecov.io/gh/apache/airflow/pull/7391/diff?src=pr=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvY25jZi9rdWJlcm5ldGVzL29wZXJhdG9ycy9rdWJlcm5ldGVzX3BvZC5weQ==)
 | `93.61% <0%> (+23.4%)` | :arrow_up: |
   | 
[airflow/kubernetes/refresh\_config.py](https://codecov.io/gh/apache/airflow/pull/7391/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3JlZnJlc2hfY29uZmlnLnB5)
 | `74.5% <0%> (+23.52%)` | :arrow_up: |
   | 
[airflow/kubernetes/pod\_launcher.py](https://codecov.io/gh/apache/airflow/pull/7391/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3BvZF9sYXVuY2hlci5weQ==)
 | `92.25% <0%> (+45.07%)` | :arrow_up: |
   | 
[airflow/kubernetes/volume.py](https://codecov.io/gh/apache/airflow/pull/7391/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3ZvbHVtZS5weQ==)
 | `100% <0%> (+47.05%)` | :arrow_up: |
   | 
[airflow/kubernetes/volume\_mount.py](https://codecov.io/gh/apache/airflow/pull/7391/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3ZvbHVtZV9tb3VudC5weQ==)
 | `100% <0%> (+55.55%)` | :arrow_up: |
   
   --
   
   [Continue to review full report at 
Codecov](https://codecov.io/gh/apache/airflow/pull/7391?src=pr=continue).
   > **Legend** - [Click here to learn 
more](https://docs.codecov.io/docs/codecov-delta)
   > `Δ = absolute  (impact)`, `ø = not affected`, `? = missing data`
   > Powered by 
[Codecov](https://codecov.io/gh/apache/airflow/pull/7391?src=pr=footer). 
Last update 
[53ba6d6...c450e45](https://codecov.io/gh/apache/airflow/pull/7391?src=pr=lastupdated).
 Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments).
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] codecov-io edited a comment on issue #6652: [AIRFLOW-5548] [AIRFLOW-5550] REST API enhancement - dag info, task …

2020-02-10 Thread GitBox
codecov-io edited a comment on issue #6652: [AIRFLOW-5548] [AIRFLOW-5550] REST 
API enhancement - dag info, task …
URL: https://github.com/apache/airflow/pull/6652#issuecomment-558277657
 
 
   # [Codecov](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=h1) 
Report
   > Merging 
[#6652](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=desc) into 
[master](https://codecov.io/gh/apache/airflow/commit/b738c9eaca11f4ed7dd680d797335e2eef9817b8?src=pr=desc)
 will **increase** coverage by `53.52%`.
   > The diff coverage is `33.66%`.
   
   [![Impacted file tree 
graph](https://codecov.io/gh/apache/airflow/pull/6652/graphs/tree.svg?width=650=WdLKlKHOAU=150=pr)](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=tree)
   
   ```diff
   @@ Coverage Diff @@
   ##   master#6652   +/-   ##
   ===
   + Coverage   32.93%   86.45%   +53.52% 
   ===
 Files 871  874+3 
 Lines   4066040760  +100 
   ===
   + Hits1339235240+21848 
   + Misses  27268 5520-21748
   ```
   
   
   | [Impacted 
Files](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=tree) | 
Coverage Δ | |
   |---|---|---|
   | 
[airflow/www/api/experimental/endpoints.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy93d3cvYXBpL2V4cGVyaW1lbnRhbC9lbmRwb2ludHMucHk=)
 | `75.54% <23.72%> (+41.75%)` | :arrow_up: |
   | 
[airflow/api/common/experimental/get\_task.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9hcGkvY29tbW9uL2V4cGVyaW1lbnRhbC9nZXRfdGFzay5weQ==)
 | `50% <28.57%> (-16.67%)` | :arrow_down: |
   | 
[airflow/api/common/experimental/get\_tasks.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9hcGkvY29tbW9uL2V4cGVyaW1lbnRhbC9nZXRfdGFza3MucHk=)
 | `50% <50%> (ø)` | |
   | 
[airflow/api/common/experimental/get\_dags.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9hcGkvY29tbW9uL2V4cGVyaW1lbnRhbC9nZXRfZGFncy5weQ==)
 | `58.33% <58.33%> (ø)` | |
   | 
[airflow/api/common/experimental/get\_dag.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9hcGkvY29tbW9uL2V4cGVyaW1lbnRhbC9nZXRfZGFnLnB5)
 | `66.66% <66.66%> (ø)` | |
   | 
[airflow/bin/airflow](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9iaW4vYWlyZmxvdw==)
 | `84.61% <0%> (ø)` | :arrow_up: |
   | 
[airflow/plugins\_manager.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9wbHVnaW5zX21hbmFnZXIucHk=)
 | `86.95% <0%> (+0.72%)` | :arrow_up: |
   | ... and [745 
more](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree-more) 
| |
   
   --
   
   [Continue to review full report at 
Codecov](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=continue).
   > **Legend** - [Click here to learn 
more](https://docs.codecov.io/docs/codecov-delta)
   > `Δ = absolute  (impact)`, `ø = not affected`, `? = missing data`
   > Powered by 
[Codecov](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=footer). 
Last update 
[b738c9e...159dfd3](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=lastupdated).
 Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments).
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] codecov-io edited a comment on issue #6652: [AIRFLOW-5548] [AIRFLOW-5550] REST API enhancement - dag info, task …

2020-02-10 Thread GitBox
codecov-io edited a comment on issue #6652: [AIRFLOW-5548] [AIRFLOW-5550] REST 
API enhancement - dag info, task …
URL: https://github.com/apache/airflow/pull/6652#issuecomment-558277657
 
 
   # [Codecov](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=h1) 
Report
   > Merging 
[#6652](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=desc) into 
[master](https://codecov.io/gh/apache/airflow/commit/b738c9eaca11f4ed7dd680d797335e2eef9817b8?src=pr=desc)
 will **increase** coverage by `53.33%`.
   > The diff coverage is `33.66%`.
   
   [![Impacted file tree 
graph](https://codecov.io/gh/apache/airflow/pull/6652/graphs/tree.svg?width=650=WdLKlKHOAU=150=pr)](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=tree)
   
   ```diff
   @@ Coverage Diff @@
   ##   master#6652   +/-   ##
   ===
   + Coverage   32.93%   86.26%   +53.33% 
   ===
 Files 871  874+3 
 Lines   4066040760  +100 
   ===
   + Hits1339235163+21771 
   + Misses  27268 5597-21671
   ```
   
   
   | [Impacted 
Files](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=tree) | 
Coverage Δ | |
   |---|---|---|
   | 
[airflow/www/api/experimental/endpoints.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy93d3cvYXBpL2V4cGVyaW1lbnRhbC9lbmRwb2ludHMucHk=)
 | `75.54% <23.72%> (+41.75%)` | :arrow_up: |
   | 
[airflow/api/common/experimental/get\_task.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9hcGkvY29tbW9uL2V4cGVyaW1lbnRhbC9nZXRfdGFzay5weQ==)
 | `50% <28.57%> (-16.67%)` | :arrow_down: |
   | 
[airflow/api/common/experimental/get\_tasks.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9hcGkvY29tbW9uL2V4cGVyaW1lbnRhbC9nZXRfdGFza3MucHk=)
 | `50% <50%> (ø)` | |
   | 
[airflow/api/common/experimental/get\_dags.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9hcGkvY29tbW9uL2V4cGVyaW1lbnRhbC9nZXRfZGFncy5weQ==)
 | `58.33% <58.33%> (ø)` | |
   | 
[airflow/api/common/experimental/get\_dag.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9hcGkvY29tbW9uL2V4cGVyaW1lbnRhbC9nZXRfZGFnLnB5)
 | `66.66% <66.66%> (ø)` | |
   | 
[airflow/bin/airflow](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9iaW4vYWlyZmxvdw==)
 | `84.61% <0%> (ø)` | :arrow_up: |
   | 
[airflow/plugins\_manager.py](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree#diff-YWlyZmxvdy9wbHVnaW5zX21hbmFnZXIucHk=)
 | `86.95% <0%> (+0.72%)` | :arrow_up: |
   | ... and [742 
more](https://codecov.io/gh/apache/airflow/pull/6652/diff?src=pr=tree-more) 
| |
   
   --
   
   [Continue to review full report at 
Codecov](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=continue).
   > **Legend** - [Click here to learn 
more](https://docs.codecov.io/docs/codecov-delta)
   > `Δ = absolute  (impact)`, `ø = not affected`, `? = missing data`
   > Powered by 
[Codecov](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=footer). 
Last update 
[b738c9e...159dfd3](https://codecov.io/gh/apache/airflow/pull/6652?src=pr=lastupdated).
 Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments).
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] blcksrx commented on a change in pull request #7358: [AIRFLOW-1202] Create Elasticsearch Hook

2020-02-10 Thread GitBox
blcksrx commented on a change in pull request #7358: [AIRFLOW-1202] Create 
Elasticsearch Hook
URL: https://github.com/apache/airflow/pull/7358#discussion_r377157022
 
 

 ##
 File path: airflow/providers/elasticsearch/hooks/elasticsearch.py
 ##
 @@ -0,0 +1,89 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+from es.elastic.api import connect
+
+from airflow.hooks.dbapi_hook import DbApiHook
+
+
+class ElasticsearchHook(DbApiHook):
+"""
+Interact with Elasticsearch through the elasticsearch-dbapi
+"""
+
+conn_name_attr = 'elasticsearch_conn_id'
+default_conn_name = 'elasticsearch_default'
+
+def __init__(self, *args, **kwargs):
+super().__init__(*args, **kwargs)
+self.schema = kwargs.pop("schema", None)
+self.connection = kwargs.pop("connection", None)
 
 Review comment:
   Actually, `self.connection` can not be as string and I did this:
   ```
def __init__(self,
schema: str = "http",
connection=None,
*args,
**kwargs):
   super().__init__(*args, **kwargs)
   self.schema = schema
   self.connection = connection
   ```


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] bhavika commented on issue #7394: [AIRFLOW-6767] Correct name for default Athena workgroup

2020-02-10 Thread GitBox
bhavika commented on issue #7394: [AIRFLOW-6767] Correct name for default 
Athena workgroup
URL: https://github.com/apache/airflow/pull/7394#issuecomment-584197494
 
 
   :wave: I'm not entirely sure how CodeCov does this but I haven't reduced the 
scope of any tests or deleted them so the reduction in coverage is probably due 
to something else. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] codecov-io edited a comment on issue #6337: [AIRFLOW-5659] - Add support for ephemeral storage on KubernetesPodOp…

2020-02-10 Thread GitBox
codecov-io edited a comment on issue #6337: [AIRFLOW-5659] - Add support for 
ephemeral storage on KubernetesPodOp…
URL: https://github.com/apache/airflow/pull/6337#issuecomment-584135185
 
 
   # [Codecov](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=h1) 
Report
   > Merging 
[#6337](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=desc) into 
[master](https://codecov.io/gh/apache/airflow/commit/53ba6d607c9a3548a996c0c716d1ffb8da0b291b?src=pr=desc)
 will **increase** coverage by `0.28%`.
   > The diff coverage is `50%`.
   
   [![Impacted file tree 
graph](https://codecov.io/gh/apache/airflow/pull/6337/graphs/tree.svg?width=650=WdLKlKHOAU=150=pr)](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=tree)
   
   ```diff
   @@Coverage Diff @@
   ##   master#6337  +/-   ##
   ==
   + Coverage86.3%   86.59%   +0.28% 
   ==
 Files 871  871  
 Lines   4066040662   +2 
   ==
   + Hits3509135210 +119 
   + Misses   5569 5452 -117
   ```
   
   
   | [Impacted 
Files](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=tree) | 
Coverage Δ | |
   |---|---|---|
   | 
[airflow/kubernetes/pod\_generator.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3BvZF9nZW5lcmF0b3IucHk=)
 | `95.95% <ø> (ø)` | :arrow_up: |
   | 
[airflow/kubernetes/pod.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3BvZC5weQ==)
 | `92.5% <50%> (+0.39%)` | :arrow_up: |
   | 
[airflow/utils/dag\_processing.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy91dGlscy9kYWdfcHJvY2Vzc2luZy5weQ==)
 | `88.12% <0%> (+0.19%)` | :arrow_up: |
   | 
[...viders/cncf/kubernetes/operators/kubernetes\_pod.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvY25jZi9rdWJlcm5ldGVzL29wZXJhdG9ycy9rdWJlcm5ldGVzX3BvZC5weQ==)
 | `93.61% <0%> (+23.4%)` | :arrow_up: |
   | 
[airflow/kubernetes/refresh\_config.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3JlZnJlc2hfY29uZmlnLnB5)
 | `74.5% <0%> (+23.52%)` | :arrow_up: |
   | 
[airflow/kubernetes/pod\_launcher.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3BvZF9sYXVuY2hlci5weQ==)
 | `92.25% <0%> (+45.07%)` | :arrow_up: |
   | 
[airflow/kubernetes/volume.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3ZvbHVtZS5weQ==)
 | `100% <0%> (+47.05%)` | :arrow_up: |
   | 
[airflow/kubernetes/volume\_mount.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3ZvbHVtZV9tb3VudC5weQ==)
 | `100% <0%> (+55.55%)` | :arrow_up: |
   
   --
   
   [Continue to review full report at 
Codecov](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=continue).
   > **Legend** - [Click here to learn 
more](https://docs.codecov.io/docs/codecov-delta)
   > `Δ = absolute  (impact)`, `ø = not affected`, `? = missing data`
   > Powered by 
[Codecov](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=footer). 
Last update 
[53ba6d6...0b17c3a](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=lastupdated).
 Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments).
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] blcksrx commented on a change in pull request #7358: [AIRFLOW-1202] Create Elasticsearch Hook

2020-02-10 Thread GitBox
blcksrx commented on a change in pull request #7358: [AIRFLOW-1202] Create 
Elasticsearch Hook
URL: https://github.com/apache/airflow/pull/7358#discussion_r376952688
 
 

 ##
 File path: airflow/providers/elasticsearch/hooks/elasticsearch.py
 ##
 @@ -0,0 +1,89 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+from es.elastic.api import connect
+
+from airflow.hooks.dbapi_hook import DbApiHook
+
+
+class ElasticsearchHook(DbApiHook):
+"""
+Interact with Elasticsearch through the elasticsearch-dbapi
+"""
+
+conn_name_attr = 'elasticsearch_conn_id'
+default_conn_name = 'elasticsearch_default'
+
+def __init__(self, *args, **kwargs):
+super().__init__(*args, **kwargs)
+self.schema = kwargs.pop("schema", None)
+self.connection = kwargs.pop("connection", None)
 
 Review comment:
   Done!


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] codecov-io commented on issue #7394: [AIRFLOW-6767] Correct name for default Athena workgroup

2020-02-10 Thread GitBox
codecov-io commented on issue #7394: [AIRFLOW-6767] Correct name for default 
Athena workgroup
URL: https://github.com/apache/airflow/pull/7394#issuecomment-584195791
 
 
   # [Codecov](https://codecov.io/gh/apache/airflow/pull/7394?src=pr=h1) 
Report
   > Merging 
[#7394](https://codecov.io/gh/apache/airflow/pull/7394?src=pr=desc) into 
[master](https://codecov.io/gh/apache/airflow/commit/53ba6d607c9a3548a996c0c716d1ffb8da0b291b?src=pr=desc)
 will **decrease** coverage by `0.17%`.
   > The diff coverage is `n/a`.
   
   [![Impacted file tree 
graph](https://codecov.io/gh/apache/airflow/pull/7394/graphs/tree.svg?width=650=WdLKlKHOAU=150=pr)](https://codecov.io/gh/apache/airflow/pull/7394?src=pr=tree)
   
   ```diff
   @@Coverage Diff @@
   ##   master#7394  +/-   ##
   ==
   - Coverage86.3%   86.12%   -0.18% 
   ==
 Files 871  871  
 Lines   4066040660  
   ==
   - Hits3509135020  -71 
   - Misses   5569 5640  +71
   ```
   
   
   | [Impacted 
Files](https://codecov.io/gh/apache/airflow/pull/7394?src=pr=tree) | 
Coverage Δ | |
   |---|---|---|
   | 
[airflow/providers/amazon/aws/hooks/athena.py](https://codecov.io/gh/apache/airflow/pull/7394/diff?src=pr=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvYW1hem9uL2F3cy9ob29rcy9hdGhlbmEucHk=)
 | `67.18% <ø> (ø)` | :arrow_up: |
   | 
[airflow/providers/amazon/aws/operators/athena.py](https://codecov.io/gh/apache/airflow/pull/7394/diff?src=pr=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvYW1hem9uL2F3cy9vcGVyYXRvcnMvYXRoZW5hLnB5)
 | `74.5% <ø> (ø)` | :arrow_up: |
   | 
[...w/providers/apache/hive/operators/mysql\_to\_hive.py](https://codecov.io/gh/apache/airflow/pull/7394/diff?src=pr=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvYXBhY2hlL2hpdmUvb3BlcmF0b3JzL215c3FsX3RvX2hpdmUucHk=)
 | `35.84% <0%> (-64.16%)` | :arrow_down: |
   | 
[airflow/security/kerberos.py](https://codecov.io/gh/apache/airflow/pull/7394/diff?src=pr=tree#diff-YWlyZmxvdy9zZWN1cml0eS9rZXJiZXJvcy5weQ==)
 | `30.43% <0%> (-45.66%)` | :arrow_down: |
   | 
[airflow/providers/mysql/operators/mysql.py](https://codecov.io/gh/apache/airflow/pull/7394/diff?src=pr=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvbXlzcWwvb3BlcmF0b3JzL215c3FsLnB5)
 | `55% <0%> (-45%)` | :arrow_down: |
   | 
[airflow/providers/apache/hive/hooks/hive.py](https://codecov.io/gh/apache/airflow/pull/7394/diff?src=pr=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvYXBhY2hlL2hpdmUvaG9va3MvaGl2ZS5weQ==)
 | `76.02% <0%> (-1.54%)` | :arrow_down: |
   | 
[airflow/hooks/dbapi\_hook.py](https://codecov.io/gh/apache/airflow/pull/7394/diff?src=pr=tree#diff-YWlyZmxvdy9ob29rcy9kYmFwaV9ob29rLnB5)
 | `90.9% <0%> (-0.83%)` | :arrow_down: |
   | 
[airflow/jobs/scheduler\_job.py](https://codecov.io/gh/apache/airflow/pull/7394/diff?src=pr=tree#diff-YWlyZmxvdy9qb2JzL3NjaGVkdWxlcl9qb2IucHk=)
 | `89.19% <0%> (-0.15%)` | :arrow_down: |
   | 
[airflow/utils/dag\_processing.py](https://codecov.io/gh/apache/airflow/pull/7394/diff?src=pr=tree#diff-YWlyZmxvdy91dGlscy9kYWdfcHJvY2Vzc2luZy5weQ==)
 | `88.12% <0%> (+0.19%)` | :arrow_up: |
   
   --
   
   [Continue to review full report at 
Codecov](https://codecov.io/gh/apache/airflow/pull/7394?src=pr=continue).
   > **Legend** - [Click here to learn 
more](https://docs.codecov.io/docs/codecov-delta)
   > `Δ = absolute  (impact)`, `ø = not affected`, `? = missing data`
   > Powered by 
[Codecov](https://codecov.io/gh/apache/airflow/pull/7394?src=pr=footer). 
Last update 
[53ba6d6...0fb123a](https://codecov.io/gh/apache/airflow/pull/7394?src=pr=lastupdated).
 Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments).
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Comment Edited] (AIRFLOW-3534) KubernetesPodOperator breaks with active log-collection for long running tasks

2020-02-10 Thread William Lachance (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3534?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033699#comment-17033699
 ] 

William Lachance edited comment on AIRFLOW-3534 at 2/10/20 3:46 PM:


My intuition at this point is that this is an issue with the Kubernetes python 
library, but more investigation would be required to establish that for certain.


was (Author: wlach):
My intuition at this point is that this is an issue with the Kubernetes 
library, but more investigation would be required to establish that for certain.

> KubernetesPodOperator breaks with active log-collection for long running tasks
> --
>
> Key: AIRFLOW-3534
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3534
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: operators
>Affects Versions: 1.10.1
>Reporter: Christian Thiel
>Priority: Major
>  Labels: kubernetes
>
> If a KubernetesPodOperator is started with get_logs=True, the pod breaks if 
> no logs are produced after ~30 seconds due to http client timeout.
> The error occurs in two ways:
> 1. If the script doesn't write anything to stdout there are three WARNINGS of 
> the connectionpool trying to get the logs: 
> {code:python}
> // Some comments here
> 2018-12-17 15:23:15,092] {{logging_mixin.py:95}} WARNING - 2018-12-17 
> 15:23:15,092 WARNING Retrying (Retry(total=2, connect=None, read=None, 
> redirect=None, status=None)) after connection broken by 
> 'ProtocolError('Connection aborted.', RemoteDisconnected('Remote end closed 
> connection without response',))': 
> /k8s/clusters/local/api/v1/namespaces/my-namespace/pods/my-pod/log?container=base=True=10
> {code}
> Followed by a {code:python}http.client.RemoteDisconnected: Remote end closed 
> connection without response{code}
> originating from _monitor_pod in /contrib/kubernetes/pod_launcher.py
> Full Traceback:
> {code:python}
> Traceback (most recent call last):
>   File "/opt/conda/lib/python3.6/site-packages/airflow/models.py", line 1659, 
> in _run_raw_task
> result = task_copy.execute(context=context)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/operators/kubernetes_pod_operator.py",
>  line 123, in execute
> get_logs=self.get_logs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/kubernetes/pod_launcher.py",
>  line 90, in run_pod
> return self._monitor_pod(pod, get_logs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/kubernetes/pod_launcher.py",
>  line 102, in _monitor_pod
> _preload_content=False)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/apis/core_v1_api.py",
>  line 18583, in read_namespaced_pod_log
> (data) = self.read_namespaced_pod_log_with_http_info(name, namespace, 
> **kwargs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/apis/core_v1_api.py",
>  line 18689, in read_namespaced_pod_log_with_http_info
> collection_formats=collection_formats)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 321, in call_api
> _return_http_data_only, collection_formats, _preload_content, 
> _request_timeout)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 155, in __call_api
> _request_timeout=_request_timeout)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 342, in request
> headers=headers)
>   File "/opt/conda/lib/python3.6/site-packages/kubernetes/client/rest.py", 
> line 231, in GET
> query_params=query_params)
>   File "/opt/conda/lib/python3.6/site-packages/kubernetes/client/rest.py", 
> line 205, in request
> headers=headers)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/request.py", line 68, 
> in request
> **urlopen_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/request.py", line 89, 
> in request_encode_url
> return self.urlopen(method, url, **extra_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/poolmanager.py", line 
> 322, in urlopen
> response = conn.urlopen(method, u.request_uri, **kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 638, in urlopen
> _stacktrace=sys.exc_info()[2])
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/util/retry.py", line 
> 398, in increment
> 

[jira] [Commented] (AIRFLOW-3534) KubernetesPodOperator breaks with active log-collection for long running tasks

2020-02-10 Thread William Lachance (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3534?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033699#comment-17033699
 ] 

William Lachance commented on AIRFLOW-3534:
---

My intuition at this point is that this is an issue with the Kubernetes 
library, but more investigation would be required to establish that for certain.

> KubernetesPodOperator breaks with active log-collection for long running tasks
> --
>
> Key: AIRFLOW-3534
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3534
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: operators
>Affects Versions: 1.10.1
>Reporter: Christian Thiel
>Priority: Major
>  Labels: kubernetes
>
> If a KubernetesPodOperator is started with get_logs=True, the pod breaks if 
> no logs are produced after ~30 seconds due to http client timeout.
> The error occurs in two ways:
> 1. If the script doesn't write anything to stdout there are three WARNINGS of 
> the connectionpool trying to get the logs: 
> {code:python}
> // Some comments here
> 2018-12-17 15:23:15,092] {{logging_mixin.py:95}} WARNING - 2018-12-17 
> 15:23:15,092 WARNING Retrying (Retry(total=2, connect=None, read=None, 
> redirect=None, status=None)) after connection broken by 
> 'ProtocolError('Connection aborted.', RemoteDisconnected('Remote end closed 
> connection without response',))': 
> /k8s/clusters/local/api/v1/namespaces/my-namespace/pods/my-pod/log?container=base=True=10
> {code}
> Followed by a {code:python}http.client.RemoteDisconnected: Remote end closed 
> connection without response{code}
> originating from _monitor_pod in /contrib/kubernetes/pod_launcher.py
> Full Traceback:
> {code:python}
> Traceback (most recent call last):
>   File "/opt/conda/lib/python3.6/site-packages/airflow/models.py", line 1659, 
> in _run_raw_task
> result = task_copy.execute(context=context)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/operators/kubernetes_pod_operator.py",
>  line 123, in execute
> get_logs=self.get_logs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/kubernetes/pod_launcher.py",
>  line 90, in run_pod
> return self._monitor_pod(pod, get_logs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/kubernetes/pod_launcher.py",
>  line 102, in _monitor_pod
> _preload_content=False)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/apis/core_v1_api.py",
>  line 18583, in read_namespaced_pod_log
> (data) = self.read_namespaced_pod_log_with_http_info(name, namespace, 
> **kwargs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/apis/core_v1_api.py",
>  line 18689, in read_namespaced_pod_log_with_http_info
> collection_formats=collection_formats)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 321, in call_api
> _return_http_data_only, collection_formats, _preload_content, 
> _request_timeout)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 155, in __call_api
> _request_timeout=_request_timeout)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 342, in request
> headers=headers)
>   File "/opt/conda/lib/python3.6/site-packages/kubernetes/client/rest.py", 
> line 231, in GET
> query_params=query_params)
>   File "/opt/conda/lib/python3.6/site-packages/kubernetes/client/rest.py", 
> line 205, in request
> headers=headers)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/request.py", line 68, 
> in request
> **urlopen_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/request.py", line 89, 
> in request_encode_url
> return self.urlopen(method, url, **extra_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/poolmanager.py", line 
> 322, in urlopen
> response = conn.urlopen(method, u.request_uri, **kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 638, in urlopen
> _stacktrace=sys.exc_info()[2])
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/util/retry.py", line 
> 398, in increment
> raise MaxRetryError(_pool, url, error or ResponseError(cause))
> urllib3.exceptions.MaxRetryError: 
> HTTPSConnectionPool(host='rancher.benteler.net', port=443): Max retries 
> exceeded with url: 
> 

[jira] [Commented] (AIRFLOW-6505) Let emoji encoded properly for json.dumps() -- BaseSQLToGoogleCloudStorageOperator

2020-02-10 Thread Kamil Bregula (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-6505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033698#comment-17033698
 ] 

Kamil Bregula commented on AIRFLOW-6505:


[~Damon09273] Can you create PR in the Apache repository?  I missed this PR 
because I only browse PRs from the Apache repository.

> Let emoji encoded properly for json.dumps() -- 
> BaseSQLToGoogleCloudStorageOperator
> --
>
> Key: AIRFLOW-6505
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6505
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.8
> Environment: 1.10.7
>Reporter: Damon Liao
>Assignee: Damon Liao
>Priority: Blocker
> Fix For: 1.10.7
>
>
> Emoji can't be encoded properly when ` json.dumps()` and 'UTF-8', problem 
> fixed by adding parameter `ensure_ascii=False`. 
>  I've created a [PR|[https://github.com/damon09273/airflow/pull/1]] to fix 
> this problem.
> In [line|#L175]] BaseSQLToGoogleCloudStorageOperato.
>  
> For example
> the emoji  encoded differentially when in use or not use  
> `ensure_ascii=False`.
> In Use (correct UTF-8 encode): "\xf0\x9f\x8d\xbb"
> Not Use (only 2 slash):"\\\ud83c\\\udf7b"
>  
> Ref: 
> [https://stackoverflow.com/questions/51183947/python-json-dumps-doesnt-encode-emojis-properly]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (AIRFLOW-6505) Let emoji encoded properly for json.dumps() -- BaseSQLToGoogleCloudStorageOperator

2020-02-10 Thread Roster (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-6505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033694#comment-17033694
 ] 

Roster commented on AIRFLOW-6505:
-

Ran into this one too and came up with this Jira
[~Damon09273] i think you wanted to PR your code to Airflow... you PRed it to 
your own fork thus no one see it.

> Let emoji encoded properly for json.dumps() -- 
> BaseSQLToGoogleCloudStorageOperator
> --
>
> Key: AIRFLOW-6505
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6505
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: contrib
>Affects Versions: 1.10.8
> Environment: 1.10.7
>Reporter: Damon Liao
>Assignee: Damon Liao
>Priority: Blocker
> Fix For: 1.10.7
>
>
> Emoji can't be encoded properly when ` json.dumps()` and 'UTF-8', problem 
> fixed by adding parameter `ensure_ascii=False`. 
>  I've created a [PR|[https://github.com/damon09273/airflow/pull/1]] to fix 
> this problem.
> In [line|#L175]] BaseSQLToGoogleCloudStorageOperato.
>  
> For example
> the emoji  encoded differentially when in use or not use  
> `ensure_ascii=False`.
> In Use (correct UTF-8 encode): "\xf0\x9f\x8d\xbb"
> Not Use (only 2 slash):"\\\ud83c\\\udf7b"
>  
> Ref: 
> [https://stackoverflow.com/questions/51183947/python-json-dumps-doesnt-encode-emojis-properly]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (AIRFLOW-3534) KubernetesPodOperator breaks with active log-collection for long running tasks

2020-02-10 Thread Kamil Bregula (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3534?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033693#comment-17033693
 ] 

Kamil Bregula commented on AIRFLOW-3534:


Do you think this is somehow caused by Airflow or is it a bug in the Kubernetes 
library? I would like to have clarity.

> KubernetesPodOperator breaks with active log-collection for long running tasks
> --
>
> Key: AIRFLOW-3534
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3534
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: operators
>Affects Versions: 1.10.1
>Reporter: Christian Thiel
>Priority: Major
>  Labels: kubernetes
>
> If a KubernetesPodOperator is started with get_logs=True, the pod breaks if 
> no logs are produced after ~30 seconds due to http client timeout.
> The error occurs in two ways:
> 1. If the script doesn't write anything to stdout there are three WARNINGS of 
> the connectionpool trying to get the logs: 
> {code:python}
> // Some comments here
> 2018-12-17 15:23:15,092] {{logging_mixin.py:95}} WARNING - 2018-12-17 
> 15:23:15,092 WARNING Retrying (Retry(total=2, connect=None, read=None, 
> redirect=None, status=None)) after connection broken by 
> 'ProtocolError('Connection aborted.', RemoteDisconnected('Remote end closed 
> connection without response',))': 
> /k8s/clusters/local/api/v1/namespaces/my-namespace/pods/my-pod/log?container=base=True=10
> {code}
> Followed by a {code:python}http.client.RemoteDisconnected: Remote end closed 
> connection without response{code}
> originating from _monitor_pod in /contrib/kubernetes/pod_launcher.py
> Full Traceback:
> {code:python}
> Traceback (most recent call last):
>   File "/opt/conda/lib/python3.6/site-packages/airflow/models.py", line 1659, 
> in _run_raw_task
> result = task_copy.execute(context=context)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/operators/kubernetes_pod_operator.py",
>  line 123, in execute
> get_logs=self.get_logs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/kubernetes/pod_launcher.py",
>  line 90, in run_pod
> return self._monitor_pod(pod, get_logs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/kubernetes/pod_launcher.py",
>  line 102, in _monitor_pod
> _preload_content=False)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/apis/core_v1_api.py",
>  line 18583, in read_namespaced_pod_log
> (data) = self.read_namespaced_pod_log_with_http_info(name, namespace, 
> **kwargs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/apis/core_v1_api.py",
>  line 18689, in read_namespaced_pod_log_with_http_info
> collection_formats=collection_formats)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 321, in call_api
> _return_http_data_only, collection_formats, _preload_content, 
> _request_timeout)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 155, in __call_api
> _request_timeout=_request_timeout)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 342, in request
> headers=headers)
>   File "/opt/conda/lib/python3.6/site-packages/kubernetes/client/rest.py", 
> line 231, in GET
> query_params=query_params)
>   File "/opt/conda/lib/python3.6/site-packages/kubernetes/client/rest.py", 
> line 205, in request
> headers=headers)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/request.py", line 68, 
> in request
> **urlopen_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/request.py", line 89, 
> in request_encode_url
> return self.urlopen(method, url, **extra_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/poolmanager.py", line 
> 322, in urlopen
> response = conn.urlopen(method, u.request_uri, **kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 638, in urlopen
> _stacktrace=sys.exc_info()[2])
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/util/retry.py", line 
> 398, in increment
> raise MaxRetryError(_pool, url, error or ResponseError(cause))
> urllib3.exceptions.MaxRetryError: 
> HTTPSConnectionPool(host='rancher.benteler.net', port=443): Max retries 
> exceeded with url: 
> 

[jira] [Comment Edited] (AIRFLOW-3534) KubernetesPodOperator breaks with active log-collection for long running tasks

2020-02-10 Thread William Lachance (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3534?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17023334#comment-17023334
 ] 

William Lachance edited comment on AIRFLOW-3534 at 2/10/20 3:33 PM:


I've been looking into this one as it's pretty annoying for us 
(https://github.com/mozilla/telemetry-airflow/issues/844).

It appears that it's possible to override the default timeouts in the method by 
passing a _`_request_timeout` paramater to the `read_namespaced_pod_log` 
method:_

_[https://raw.githubusercontent.com/kubernetes-client/python/master/kubernetes/client/api/core_v1_api.py_]

Reference on how this parameter gets processed:

[https://github.com/kubernetes-client/python/blob/5181d2389eacd92fa0f72237da87286ed0d45a3e/kubernetes/client/rest.py#L141]

I did up a quick prototype of this here:

[https://github.com/wlach/airflow/commit/e7ae01ac608d3f944b691875a9cf90dceb60ebcc]

Unfortunately it doesn't seem to solve the problem (I'm still getting the error 
after 5 minutes of no output in my testing). Will do more research as time 
allows...


was (Author: wlach):
I've been looking into this one as it's pretty annoying for us ().

It appears that it's possible to override the default timeouts in the method by 
passing a _`_request_timeout` paramater to the `read_namespaced_pod_log` 
method:_

_https://raw.githubusercontent.com/kubernetes-client/python/master/kubernetes/client/api/core_v1_api.py_

Reference on how this parameter gets processed:

https://github.com/kubernetes-client/python/blob/5181d2389eacd92fa0f72237da87286ed0d45a3e/kubernetes/client/rest.py#L141

I did up a quick prototype of this here:

[https://github.com/wlach/airflow/commit/e7ae01ac608d3f944b691875a9cf90dceb60ebcc]

Unfortunately it doesn't seem to solve the problem (I'm still getting the error 
after 5 minutes of no output in my testing). Will do more research as time 
allows...

> KubernetesPodOperator breaks with active log-collection for long running tasks
> --
>
> Key: AIRFLOW-3534
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3534
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: operators
>Affects Versions: 1.10.1
>Reporter: Christian Thiel
>Priority: Major
>  Labels: kubernetes
>
> If a KubernetesPodOperator is started with get_logs=True, the pod breaks if 
> no logs are produced after ~30 seconds due to http client timeout.
> The error occurs in two ways:
> 1. If the script doesn't write anything to stdout there are three WARNINGS of 
> the connectionpool trying to get the logs: 
> {code:python}
> // Some comments here
> 2018-12-17 15:23:15,092] {{logging_mixin.py:95}} WARNING - 2018-12-17 
> 15:23:15,092 WARNING Retrying (Retry(total=2, connect=None, read=None, 
> redirect=None, status=None)) after connection broken by 
> 'ProtocolError('Connection aborted.', RemoteDisconnected('Remote end closed 
> connection without response',))': 
> /k8s/clusters/local/api/v1/namespaces/my-namespace/pods/my-pod/log?container=base=True=10
> {code}
> Followed by a {code:python}http.client.RemoteDisconnected: Remote end closed 
> connection without response{code}
> originating from _monitor_pod in /contrib/kubernetes/pod_launcher.py
> Full Traceback:
> {code:python}
> Traceback (most recent call last):
>   File "/opt/conda/lib/python3.6/site-packages/airflow/models.py", line 1659, 
> in _run_raw_task
> result = task_copy.execute(context=context)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/operators/kubernetes_pod_operator.py",
>  line 123, in execute
> get_logs=self.get_logs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/kubernetes/pod_launcher.py",
>  line 90, in run_pod
> return self._monitor_pod(pod, get_logs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/kubernetes/pod_launcher.py",
>  line 102, in _monitor_pod
> _preload_content=False)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/apis/core_v1_api.py",
>  line 18583, in read_namespaced_pod_log
> (data) = self.read_namespaced_pod_log_with_http_info(name, namespace, 
> **kwargs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/apis/core_v1_api.py",
>  line 18689, in read_namespaced_pod_log_with_http_info
> collection_formats=collection_formats)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 321, in call_api
> _return_http_data_only, collection_formats, _preload_content, 
> _request_timeout)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 155, in __call_api
> _request_timeout=_request_timeout)
>   File 
> 

[jira] [Commented] (AIRFLOW-3534) KubernetesPodOperator breaks with active log-collection for long running tasks

2020-02-10 Thread William Lachance (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-3534?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033689#comment-17033689
 ] 

William Lachance commented on AIRFLOW-3534:
---

I was using the latest stable version (10.0.1) but can reproduce this problem 
on the pre-release version (11.0.0b2) as well.

> KubernetesPodOperator breaks with active log-collection for long running tasks
> --
>
> Key: AIRFLOW-3534
> URL: https://issues.apache.org/jira/browse/AIRFLOW-3534
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: operators
>Affects Versions: 1.10.1
>Reporter: Christian Thiel
>Priority: Major
>  Labels: kubernetes
>
> If a KubernetesPodOperator is started with get_logs=True, the pod breaks if 
> no logs are produced after ~30 seconds due to http client timeout.
> The error occurs in two ways:
> 1. If the script doesn't write anything to stdout there are three WARNINGS of 
> the connectionpool trying to get the logs: 
> {code:python}
> // Some comments here
> 2018-12-17 15:23:15,092] {{logging_mixin.py:95}} WARNING - 2018-12-17 
> 15:23:15,092 WARNING Retrying (Retry(total=2, connect=None, read=None, 
> redirect=None, status=None)) after connection broken by 
> 'ProtocolError('Connection aborted.', RemoteDisconnected('Remote end closed 
> connection without response',))': 
> /k8s/clusters/local/api/v1/namespaces/my-namespace/pods/my-pod/log?container=base=True=10
> {code}
> Followed by a {code:python}http.client.RemoteDisconnected: Remote end closed 
> connection without response{code}
> originating from _monitor_pod in /contrib/kubernetes/pod_launcher.py
> Full Traceback:
> {code:python}
> Traceback (most recent call last):
>   File "/opt/conda/lib/python3.6/site-packages/airflow/models.py", line 1659, 
> in _run_raw_task
> result = task_copy.execute(context=context)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/operators/kubernetes_pod_operator.py",
>  line 123, in execute
> get_logs=self.get_logs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/kubernetes/pod_launcher.py",
>  line 90, in run_pod
> return self._monitor_pod(pod, get_logs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/airflow/contrib/kubernetes/pod_launcher.py",
>  line 102, in _monitor_pod
> _preload_content=False)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/apis/core_v1_api.py",
>  line 18583, in read_namespaced_pod_log
> (data) = self.read_namespaced_pod_log_with_http_info(name, namespace, 
> **kwargs)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/apis/core_v1_api.py",
>  line 18689, in read_namespaced_pod_log_with_http_info
> collection_formats=collection_formats)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 321, in call_api
> _return_http_data_only, collection_formats, _preload_content, 
> _request_timeout)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 155, in __call_api
> _request_timeout=_request_timeout)
>   File 
> "/opt/conda/lib/python3.6/site-packages/kubernetes/client/api_client.py", 
> line 342, in request
> headers=headers)
>   File "/opt/conda/lib/python3.6/site-packages/kubernetes/client/rest.py", 
> line 231, in GET
> query_params=query_params)
>   File "/opt/conda/lib/python3.6/site-packages/kubernetes/client/rest.py", 
> line 205, in request
> headers=headers)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/request.py", line 68, 
> in request
> **urlopen_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/request.py", line 89, 
> in request_encode_url
> return self.urlopen(method, url, **extra_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/poolmanager.py", line 
> 322, in urlopen
> response = conn.urlopen(method, u.request_uri, **kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 667, in urlopen
> **response_kw)
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/connectionpool.py", 
> line 638, in urlopen
> _stacktrace=sys.exc_info()[2])
>   File "/opt/conda/lib/python3.6/site-packages/urllib3/util/retry.py", line 
> 398, in increment
> raise MaxRetryError(_pool, url, error or ResponseError(cause))
> urllib3.exceptions.MaxRetryError: 
> HTTPSConnectionPool(host='rancher.benteler.net', port=443): Max retries 
> exceeded with url: 
> 

[GitHub] [airflow] ashb commented on a change in pull request #7391: [AIRFLOW-6663] Prepare backporting packages

2020-02-10 Thread GitBox
ashb commented on a change in pull request #7391: [AIRFLOW-6663] Prepare 
backporting packages
URL: https://github.com/apache/airflow/pull/7391#discussion_r377129979
 
 

 ##
 File path: MANIFEST-packages.in
 ##
 @@ -0,0 +1,24 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+include NOTICE
+include LICENSE
+include CHANGELOG.txt
+include README.md
+graft licenses/
 
 Review comment:
   This was from CI logs


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] codecov-io commented on issue #7321: [AIRFLOW-XXXX] Remove duplication in BaseOperator docstring

2020-02-10 Thread GitBox
codecov-io commented on issue #7321: [AIRFLOW-] Remove duplication in 
BaseOperator docstring
URL: https://github.com/apache/airflow/pull/7321#issuecomment-584171117
 
 
   # [Codecov](https://codecov.io/gh/apache/airflow/pull/7321?src=pr=h1) 
Report
   > Merging 
[#7321](https://codecov.io/gh/apache/airflow/pull/7321?src=pr=desc) into 
[master](https://codecov.io/gh/apache/airflow/commit/53ba6d607c9a3548a996c0c716d1ffb8da0b291b?src=pr=desc)
 will **not change** coverage.
   > The diff coverage is `n/a`.
   
   [![Impacted file tree 
graph](https://codecov.io/gh/apache/airflow/pull/7321/graphs/tree.svg?width=650=WdLKlKHOAU=150=pr)](https://codecov.io/gh/apache/airflow/pull/7321?src=pr=tree)
   
   ```diff
   @@  Coverage Diff   @@
   ##   master   #7321   +/-   ##
   ==
 Coverage86.3%   86.3%   
   ==
 Files 871 871   
 Lines   40660   40660   
   ==
 Hits35091   35091   
 Misses   55695569
   ```
   
   
   | [Impacted 
Files](https://codecov.io/gh/apache/airflow/pull/7321?src=pr=tree) | 
Coverage Δ | |
   |---|---|---|
   | 
[airflow/models/baseoperator.py](https://codecov.io/gh/apache/airflow/pull/7321/diff?src=pr=tree#diff-YWlyZmxvdy9tb2RlbHMvYmFzZW9wZXJhdG9yLnB5)
 | `96.51% <ø> (ø)` | :arrow_up: |
   
   --
   
   [Continue to review full report at 
Codecov](https://codecov.io/gh/apache/airflow/pull/7321?src=pr=continue).
   > **Legend** - [Click here to learn 
more](https://docs.codecov.io/docs/codecov-delta)
   > `Δ = absolute  (impact)`, `ø = not affected`, `? = missing data`
   > Powered by 
[Codecov](https://codecov.io/gh/apache/airflow/pull/7321?src=pr=footer). 
Last update 
[53ba6d6...2d1e37a](https://codecov.io/gh/apache/airflow/pull/7321?src=pr=lastupdated).
 Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments).
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on issue #7391: [AIRFLOW-6663] Prepare backporting packages

2020-02-10 Thread GitBox
potiuk commented on issue #7391: [AIRFLOW-6663] Prepare backporting packages
URL: https://github.com/apache/airflow/pull/7391#issuecomment-584166040
 
 
   Fixed comments


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare backporting packages

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare 
backporting packages
URL: https://github.com/apache/airflow/pull/7391#discussion_r377115621
 
 

 ##
 File path: setup.py
 ##
 @@ -591,5 +594,144 @@ def do_setup():
 )
 
 
+PROVIDERS_DEPENDENCIES = {
+"amazon": [aws],
+"apache.cassandra": [cassandra],
+"apache.druid": [druid],
+"apache.hdfs": [hdfs],
+"apache.hive": [hive],
+"apache.pig": [],
+"apache.pinot": [pinot],
+"apache.spark": [],
+"apache.sqoop": [],
+"celery": [celery],
+"cloudant": [cloudant],
+"cncf.kubernetes": [kubernetes],
+"databricks": [databricks],
+"datadog": [datadog],
+"dingding": [],
+"discord": [],
+"docker": [docker],
+"email": [],
+"ftp": [],
+"google.cloud": [gcp],
+"google.marketing_platform": [gcp],
+"google.suite": [gcp],
+"grpc": [grpc],
+"http": [],
+"imap": [],
+"jdbc": [jdbc],
+"jenkins": [jenkins],
+"jira": [jira],
+"microsoft.azure": [azure],
+"microsoft.mssql": [mssql],
+"microsoft.winrm": [winrm],
+"mongo": [mongo],
+"mysql": [mysql],
+"odbc": [odbc],
+"openfass": [],
+"opsgenie": [],
+"oracle": [oracle],
+"pagerduty": [pagerduty],
+"papermill": [papermill],
+"postgres": [postgres],
+"presto": [presto],
+"qubole": [qds],
+"redis": [redis],
+"salesforce": [salesforce],
+"samba": [samba],
+"segment": [segment],
+"sftp": [],
+"slack": [slack],
+"snowflake": [snowflake],
+"sqlite": [],
+"ssh": [ssh],
+"vertica": [vertica],
+"zendesk": [zendesk],
+}
+
+
+def do_setup_package_providers(provider_module: str, deps: List[str]):
+"""Set up package providers"""
+provider_package_name = provider_module.replace(".", "_")
+package_name = f'apache-airflow-providers-{provider_package_name}' if 
provider_module != "providers" \
+else f'apache-airflow-providers'
+
+package_prefix = f'airflow.providers.{provider_module}' if provider_module 
!= 'providers' \
+else 'airflow.providers'
+found_packages = find_packages()
+found_packages = [package for package in found_packages if 
package.startswith(package_prefix)]
+print(found_packages)
+setup(
+name=package_name,
+description=f'Back-porting ${package_name} package for Airflow 1.10.*',
+long_description=f"""
+Back-ported {package_name} to 1.10.* series of Airflow.
+""",
+long_description_content_type='text/markdown',
+license='Apache License 2.0',
+version='0.0.1',
+packages=found_packages,
+include_package_data=True,
+zip_safe=False,
+install_requires=[
+'apache-airflow~=1.10',
+'pandas>=0.17.1, <1.0.0',
+] + deps,
+classifiers=[
+'Development Status :: 5 - Production/Stable',
+'Environment :: Console',
+'Intended Audience :: Developers',
+'Intended Audience :: System Administrators',
+'License :: OSI Approved :: Apache Software License',
+'Programming Language :: Python :: 3.6',
+'Programming Language :: Python :: 3.7',
+'Topic :: System :: Monitoring',
+],
+python_requires='>=3.6',
+)
+
+
+def find_package_dependencies(package):
+"""Finds dependencies for the packages"""
+if package != 'providers':
+return PROVIDERS_DEPENDENCIES.get(package)
+else:
+return list(itertools.chain(PROVIDERS_DEPENDENCIES.values()))
+
+
+def get_provider_packages():
+"""Returns all packages available in providers"""
+packages = list(PROVIDERS_DEPENDENCIES)
+return ['providers'] + packages
+
+
 if __name__ == "__main__":
-do_setup()
+if "--help" in sys.argv or "-h" in sys.argv:
+print()
+print("You can also build any of the backport packages by "
+  "adding --provider-package PACKAGE as first two")
+print("Of the setup.py arguments")
+print(f"Available packages: {get_provider_packages()}")
+print()
+print("You can see all packages configured by adding 
--list-backport-packages flag")
+
+try:
+rmtree(os.path.join(dirname(__file__), "build"))
 
 Review comment:
   I will do both but I move it to "providers-package if". It happened far too 
often that I forgot about it and build much bigger package than I thought. We 
are anyhow abusing setup.py slightly in this case so cleaning build in case we 
run providers-package is a small performance penalty.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare backporting packages

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare 
backporting packages
URL: https://github.com/apache/airflow/pull/7391#discussion_r377112953
 
 

 ##
 File path: scripts/ci/ci_prepare_backport_packages.sh
 ##
 @@ -0,0 +1,71 @@
+#!/usr/bin/env bash
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+set -euo pipefail
+
+MY_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
+
+export AIRFLOW_CI_SILENT=${AIRFLOW_CI_SILENT:="true"}
+
+export PYTHON_VERSION=${PYTHON_VERSION:-3.6}
+
+# shellcheck source=scripts/ci/_utils.sh
+. "${MY_DIR}/_utils.sh"
+
+basic_sanity_checks
+
+script_start
+
+cd "${AIRFLOW_SOURCES}"
+
+rm -rf dist/*
+rm -rf -- *.egg-info
 
 Review comment:
   Added. No harm in that.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (AIRFLOW-6768) Graph view rendering angular edges

2020-02-10 Thread Nathan Hadfield (Jira)
Nathan Hadfield created AIRFLOW-6768:


 Summary: Graph view rendering angular edges
 Key: AIRFLOW-6768
 URL: https://issues.apache.org/jira/browse/AIRFLOW-6768
 Project: Apache Airflow
  Issue Type: Bug
  Components: ui
Affects Versions: 1.10.9, 1.10.8
Reporter: Nathan Hadfield
 Attachments: Screenshot 2020-02-10 at 08.51.02.png, Screenshot 
2020-02-10 at 08.51.20.png

Since the release of v1.10.8 the DAG graph view is rendering the edges between 
nodes with angular lines rather than nice smooth curves.

Seems to have been caused by a bump of dagre-d3.

[https://github.com/apache/airflow/pull/7280]

[https://github.com/dagrejs/dagre-d3/issues/305]

 

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] nuclearpinguin commented on a change in pull request #7370: [AIRFLOW-6590] Use batch db operations in jobs

2020-02-10 Thread GitBox
nuclearpinguin commented on a change in pull request #7370: [AIRFLOW-6590] Use 
batch db operations in jobs
URL: https://github.com/apache/airflow/pull/7370#discussion_r377109262
 
 

 ##
 File path: scripts/perf/sql_queries.py
 ##
 @@ -0,0 +1,178 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+import os
+from time import sleep, time
+from typing import List, NamedTuple, Optional, Tuple
+
+import pandas as pd
+
+# Setup environment before any Airflow import
+DAG_FOLDER = "/opt/airflow/scripts/perf/dags"
+os.environ["AIRFLOW__CORE__DAGS_FOLDER"] = DAG_FOLDER
+os.environ["AIRFLOW__DEBUG__SQLALCHEMY_STATS"] = "True"
+os.environ["AIRFLOW__CORE__LOAD_EXAMPLES"] = "False"
+
+# Here we setup simpler logger to avoid any code changes in
+# Airflow core code base
+LOG_LEVEL = "INFO"
+LOG_FILE = "/files/sql_stats.log"  # Default to run in Breeze
 
 Review comment:
   Hm, hardcoding this for breeze has an advantage that it works out of the box 
in a single environment. Not hardcoding has no advantages to me. Especially 
that if someone what to play with it and use other places they will have to 
adjust it anyway 路‍♂️


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare backporting packages

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare 
backporting packages
URL: https://github.com/apache/airflow/pull/7391#discussion_r377105680
 
 

 ##
 File path: setup.py
 ##
 @@ -591,5 +594,144 @@ def do_setup():
 )
 
 
+PROVIDERS_DEPENDENCIES = {
+"amazon": [aws],
+"apache.cassandra": [cassandra],
+"apache.druid": [druid],
+"apache.hdfs": [hdfs],
+"apache.hive": [hive],
+"apache.pig": [],
+"apache.pinot": [pinot],
+"apache.spark": [],
+"apache.sqoop": [],
+"celery": [celery],
+"cloudant": [cloudant],
+"cncf.kubernetes": [kubernetes],
+"databricks": [databricks],
+"datadog": [datadog],
+"dingding": [],
+"discord": [],
+"docker": [docker],
+"email": [],
+"ftp": [],
+"google.cloud": [gcp],
+"google.marketing_platform": [gcp],
+"google.suite": [gcp],
+"grpc": [grpc],
+"http": [],
+"imap": [],
+"jdbc": [jdbc],
+"jenkins": [jenkins],
+"jira": [jira],
+"microsoft.azure": [azure],
+"microsoft.mssql": [mssql],
+"microsoft.winrm": [winrm],
+"mongo": [mongo],
+"mysql": [mysql],
+"odbc": [odbc],
+"openfass": [],
+"opsgenie": [],
+"oracle": [oracle],
+"pagerduty": [pagerduty],
+"papermill": [papermill],
+"postgres": [postgres],
+"presto": [presto],
+"qubole": [qds],
+"redis": [redis],
+"salesforce": [salesforce],
+"samba": [samba],
+"segment": [segment],
+"sftp": [],
+"slack": [slack],
+"snowflake": [snowflake],
+"sqlite": [],
+"ssh": [ssh],
+"vertica": [vertica],
+"zendesk": [zendesk],
+}
+
+
+def do_setup_package_providers(provider_module: str, deps: List[str]):
+"""Set up package providers"""
+provider_package_name = provider_module.replace(".", "_")
+package_name = f'apache-airflow-providers-{provider_package_name}' if 
provider_module != "providers" \
+else f'apache-airflow-providers'
+
+package_prefix = f'airflow.providers.{provider_module}' if provider_module 
!= 'providers' \
+else 'airflow.providers'
+found_packages = find_packages()
+found_packages = [package for package in found_packages if 
package.startswith(package_prefix)]
+print(found_packages)
+setup(
+name=package_name,
+description=f'Back-porting ${package_name} package for Airflow 1.10.*',
+long_description=f"""
+Back-ported {package_name} to 1.10.* series of Airflow.
+""",
+long_description_content_type='text/markdown',
+license='Apache License 2.0',
+version='0.0.1',
+packages=found_packages,
+include_package_data=True,
+zip_safe=False,
+install_requires=[
+'apache-airflow~=1.10',
+'pandas>=0.17.1, <1.0.0',
+] + deps,
+classifiers=[
+'Development Status :: 5 - Production/Stable',
+'Environment :: Console',
+'Intended Audience :: Developers',
+'Intended Audience :: System Administrators',
+'License :: OSI Approved :: Apache Software License',
+'Programming Language :: Python :: 3.6',
+'Programming Language :: Python :: 3.7',
+'Topic :: System :: Monitoring',
+],
+python_requires='>=3.6',
+)
+
+
+def find_package_dependencies(package):
+"""Finds dependencies for the packages"""
+if package != 'providers':
+return PROVIDERS_DEPENDENCIES.get(package)
+else:
+return list(itertools.chain(PROVIDERS_DEPENDENCIES.values()))
+
+
+def get_provider_packages():
+"""Returns all packages available in providers"""
+packages = list(PROVIDERS_DEPENDENCIES)
+return ['providers'] + packages
+
+
 if __name__ == "__main__":
-do_setup()
+if "--help" in sys.argv or "-h" in sys.argv:
+print()
+print("You can also build any of the backport packages by "
+  "adding --provider-package PACKAGE as first two")
+print("Of the setup.py arguments")
+print(f"Available packages: {get_provider_packages()}")
+print()
+print("You can see all packages configured by adding 
--list-backport-packages flag")
+
+try:
+rmtree(os.path.join(dirname(__file__), "build"))
+except FileNotFoundError:
+pass
+if len(sys.argv) > 1 and sys.argv[1] == "--provider-package":
+provider_package = sys.argv[2]
+if provider_package not in get_provider_packages():
+raise Exception(f"The package {provider_package} is not a backport 
package. "
+f"Use one of {get_provider_packages()}")
+del sys.argv[1]
+del sys.argv[1]
 
 Review comment:
   yep


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For 

[GitHub] [airflow] potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare backporting packages

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare 
backporting packages
URL: https://github.com/apache/airflow/pull/7391#discussion_r377105026
 
 

 ##
 File path: setup.py
 ##
 @@ -591,5 +594,144 @@ def do_setup():
 )
 
 
+PROVIDERS_DEPENDENCIES = {
+"amazon": [aws],
+"apache.cassandra": [cassandra],
+"apache.druid": [druid],
+"apache.hdfs": [hdfs],
+"apache.hive": [hive],
+"apache.pig": [],
+"apache.pinot": [pinot],
+"apache.spark": [],
+"apache.sqoop": [],
+"celery": [celery],
+"cloudant": [cloudant],
+"cncf.kubernetes": [kubernetes],
+"databricks": [databricks],
+"datadog": [datadog],
+"dingding": [],
+"discord": [],
+"docker": [docker],
+"email": [],
+"ftp": [],
+"google.cloud": [gcp],
+"google.marketing_platform": [gcp],
+"google.suite": [gcp],
+"grpc": [grpc],
+"http": [],
+"imap": [],
+"jdbc": [jdbc],
+"jenkins": [jenkins],
+"jira": [jira],
+"microsoft.azure": [azure],
+"microsoft.mssql": [mssql],
+"microsoft.winrm": [winrm],
+"mongo": [mongo],
+"mysql": [mysql],
+"odbc": [odbc],
+"openfass": [],
+"opsgenie": [],
+"oracle": [oracle],
+"pagerduty": [pagerduty],
+"papermill": [papermill],
+"postgres": [postgres],
+"presto": [presto],
+"qubole": [qds],
+"redis": [redis],
+"salesforce": [salesforce],
+"samba": [samba],
+"segment": [segment],
+"sftp": [],
+"slack": [slack],
+"snowflake": [snowflake],
+"sqlite": [],
+"ssh": [ssh],
+"vertica": [vertica],
+"zendesk": [zendesk],
+}
+
+
+def do_setup_package_providers(provider_module: str, deps: List[str]):
+"""Set up package providers"""
+provider_package_name = provider_module.replace(".", "_")
+package_name = f'apache-airflow-providers-{provider_package_name}' if 
provider_module != "providers" \
+else f'apache-airflow-providers'
+
+package_prefix = f'airflow.providers.{provider_module}' if provider_module 
!= 'providers' \
+else 'airflow.providers'
+found_packages = find_packages()
+found_packages = [package for package in found_packages if 
package.startswith(package_prefix)]
+print(found_packages)
+setup(
+name=package_name,
+description=f'Back-porting ${package_name} package for Airflow 1.10.*',
+long_description=f"""
+Back-ported {package_name} to 1.10.* series of Airflow.
+""",
+long_description_content_type='text/markdown',
+license='Apache License 2.0',
+version='0.0.1',
+packages=found_packages,
+include_package_data=True,
+zip_safe=False,
+install_requires=[
+'apache-airflow~=1.10',
+'pandas>=0.17.1, <1.0.0',
 
 Review comment:
   Ah yeah. Not needed. Removed.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare backporting packages

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare 
backporting packages
URL: https://github.com/apache/airflow/pull/7391#discussion_r377104128
 
 

 ##
 File path: MANIFEST-packages.in
 ##
 @@ -0,0 +1,24 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+include NOTICE
+include LICENSE
+include CHANGELOG.txt
+include README.md
+graft licenses/
 
 Review comment:
   Strange. It works for me.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (AIRFLOW-6767) Default workgroup name for Athena is incorrect

2020-02-10 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-6767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033643#comment-17033643
 ] 

ASF GitHub Bot commented on AIRFLOW-6767:
-

bhavika commented on pull request #7394: [AIRFLOW-6767] Correct name for 
default Athena workgroup
URL: https://github.com/apache/airflow/pull/7394
 
 
   ---
   Issue link: WILL BE INSERTED BY 
[boring-cyborg](https://github.com/kaxil/boring-cyborg)
   
   Make sure to mark the boxes below before creating PR: [x]
   
   - [x] Description above provides context of the change
   - [x] Commit message/PR title starts with `[AIRFLOW-]`. AIRFLOW- = 
JIRA ID*
   - [x] Unit tests coverage for changes (not needed for documentation changes)
   - [x] Commits follow "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)"
   - [x] Relevant documentation is updated including usage instructions.
   - [] I will engage committers as explained in [Contribution Workflow 
Example](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#contribution-workflow-example).
   
   * For document-only changes commit message can start with 
`[AIRFLOW-]`.
   
   ---
   In case of fundamental code change, Airflow Improvement Proposal 
([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals))
 is needed.
   In case of a new dependency, check compliance with the [ASF 3rd Party 
License Policy](https://www.apache.org/legal/resolved.html#category-x).
   In case of backwards incompatible changes please leave a note in 
[UPDATING.md](https://github.com/apache/airflow/blob/master/UPDATING.md).
   Read the [Pull Request 
Guidelines](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#pull-request-guidelines)
 for more information.
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


> Default workgroup name for Athena is incorrect
> --
>
> Key: AIRFLOW-6767
> URL: https://issues.apache.org/jira/browse/AIRFLOW-6767
> Project: Apache Airflow
>  Issue Type: Improvement
>  Components: aws
>Affects Versions: 1.10.9
>Reporter: Bhavika
>Assignee: Bhavika
>Priority: Minor
>  Labels: easy-fix, pull-request-available
> Fix For: 1.10.10
>
>
> The default Athena workgroup name is set to "default" when it should be 
> "primary". See this comment - 
> [https://github.com/apache/airflow/pull/6871#issuecomment-584066252]
> I'll submit a fix for the incorrect name. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] nuclearpinguin commented on a change in pull request #7370: [AIRFLOW-6590] Use batch db operations in jobs

2020-02-10 Thread GitBox
nuclearpinguin commented on a change in pull request #7370: [AIRFLOW-6590] Use 
batch db operations in jobs
URL: https://github.com/apache/airflow/pull/7370#discussion_r377103522
 
 

 ##
 File path: airflow/utils/sqlalchemy.py
 ##
 @@ -20,48 +20,73 @@
 import json
 import logging
 import os
+import time
+import traceback
 
 import pendulum
 from dateutil import relativedelta
 from sqlalchemy import event, exc
 from sqlalchemy.types import DateTime, Text, TypeDecorator
 
+from airflow.configuration import conf
+
 log = logging.getLogger(__name__)
+
 utc = pendulum.timezone('UTC')
 
 
 def setup_event_handlers(engine):
 """
 Setups event handlers.
 """
+# pylint: disable=unused-argument
 @event.listens_for(engine, "connect")
-def connect(dbapi_connection, connection_record):  # pylint: 
disable=unused-argument
+def connect(dbapi_connection, connection_record):
 connection_record.info['pid'] = os.getpid()
 
 if engine.dialect.name == "sqlite":
 @event.listens_for(engine, "connect")
-def set_sqlite_pragma(dbapi_connection, connection_record):  # pylint: 
disable=unused-argument
+def set_sqlite_pragma(dbapi_connection, connection_record):
 cursor = dbapi_connection.cursor()
 cursor.execute("PRAGMA foreign_keys=ON")
 cursor.close()
 
 # this ensures sanity in mysql when storing datetimes (not required for 
postgres)
 if engine.dialect.name == "mysql":
 @event.listens_for(engine, "connect")
-def set_mysql_timezone(dbapi_connection, connection_record):  # 
pylint: disable=unused-argument
+def set_mysql_timezone(dbapi_connection, connection_record):
 cursor = dbapi_connection.cursor()
 cursor.execute("SET time_zone = '+00:00'")
 cursor.close()
 
 @event.listens_for(engine, "checkout")
-def checkout(dbapi_connection, connection_record, connection_proxy):  # 
pylint: disable=unused-argument
+def checkout(dbapi_connection, connection_record, connection_proxy):
 pid = os.getpid()
 if connection_record.info['pid'] != pid:
 connection_record.connection = connection_proxy.connection = None
 raise exc.DisconnectionError(
 "Connection record belongs to pid {}, "
 "attempting to check out in pid 
{}".format(connection_record.info['pid'], pid)
 )
+if conf.getboolean('debug', 'sqlalchemy_stats', fallback=False):
+@event.listens_for(engine, "before_cursor_execute")
+def before_cursor_execute(conn, cursor, statement, parameters, 
context, executemany):
+conn.info.setdefault('query_start_time', []).append(time.time())
 
 Review comment:
   Oh, I haven't known that 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] bhavika opened a new pull request #7394: [AIRFLOW-6767] Correct name for default Athena workgroup

2020-02-10 Thread GitBox
bhavika opened a new pull request #7394: [AIRFLOW-6767] Correct name for 
default Athena workgroup
URL: https://github.com/apache/airflow/pull/7394
 
 
   ---
   Issue link: WILL BE INSERTED BY 
[boring-cyborg](https://github.com/kaxil/boring-cyborg)
   
   Make sure to mark the boxes below before creating PR: [x]
   
   - [x] Description above provides context of the change
   - [x] Commit message/PR title starts with `[AIRFLOW-]`. AIRFLOW- = 
JIRA ID*
   - [x] Unit tests coverage for changes (not needed for documentation changes)
   - [x] Commits follow "[How to write a good git commit 
message](http://chris.beams.io/posts/git-commit/)"
   - [x] Relevant documentation is updated including usage instructions.
   - [] I will engage committers as explained in [Contribution Workflow 
Example](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#contribution-workflow-example).
   
   * For document-only changes commit message can start with 
`[AIRFLOW-]`.
   
   ---
   In case of fundamental code change, Airflow Improvement Proposal 
([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals))
 is needed.
   In case of a new dependency, check compliance with the [ASF 3rd Party 
License Policy](https://www.apache.org/legal/resolved.html#category-x).
   In case of backwards incompatible changes please leave a note in 
[UPDATING.md](https://github.com/apache/airflow/blob/master/UPDATING.md).
   Read the [Pull Request 
Guidelines](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#pull-request-guidelines)
 for more information.
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] nuclearpinguin commented on a change in pull request #7370: [AIRFLOW-6590] Use batch db operations in jobs

2020-02-10 Thread GitBox
nuclearpinguin commented on a change in pull request #7370: [AIRFLOW-6590] Use 
batch db operations in jobs
URL: https://github.com/apache/airflow/pull/7370#discussion_r377101936
 
 

 ##
 File path: airflow/models/dagrun.py
 ##
 @@ -338,15 +339,30 @@ def update_state(self, session=None):
 def _get_ready_tis(self, scheduleable_tasks, finished_tasks, session):
 ready_tis = []
 changed_tis = False
-for st in scheduleable_tasks:
-st_old_state = st.state
-if st.are_dependencies_met(
+
+if not scheduleable_tasks:
+return ready_tis, changed_tis
+
+# Refresh states
+filter_for_tis = TI.filter_for_tis(scheduleable_tasks)
+fresh_tis = session.query(TI).filter(filter_for_tis).order_by(
+TI.dag_id.desc(),
+TI.task_id.desc(),
+TI.execution_date.desc(),
 
 Review comment:
   No more now


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare backporting packages

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare 
backporting packages
URL: https://github.com/apache/airflow/pull/7391#discussion_r377101053
 
 

 ##
 File path: setup.py
 ##
 @@ -591,5 +594,144 @@ def do_setup():
 )
 
 
+PROVIDERS_DEPENDENCIES = {
+"amazon": [aws],
+"apache.cassandra": [cassandra],
+"apache.druid": [druid],
+"apache.hdfs": [hdfs],
+"apache.hive": [hive],
+"apache.pig": [],
+"apache.pinot": [pinot],
+"apache.spark": [],
+"apache.sqoop": [],
+"celery": [celery],
+"cloudant": [cloudant],
+"cncf.kubernetes": [kubernetes],
+"databricks": [databricks],
+"datadog": [datadog],
+"dingding": [],
+"discord": [],
+"docker": [docker],
+"email": [],
+"ftp": [],
+"google.cloud": [gcp],
+"google.marketing_platform": [gcp],
+"google.suite": [gcp],
+"grpc": [grpc],
+"http": [],
+"imap": [],
+"jdbc": [jdbc],
+"jenkins": [jenkins],
+"jira": [jira],
+"microsoft.azure": [azure],
+"microsoft.mssql": [mssql],
+"microsoft.winrm": [winrm],
+"mongo": [mongo],
+"mysql": [mysql],
+"odbc": [odbc],
+"openfass": [],
+"opsgenie": [],
+"oracle": [oracle],
+"pagerduty": [pagerduty],
+"papermill": [papermill],
+"postgres": [postgres],
+"presto": [presto],
+"qubole": [qds],
+"redis": [redis],
+"salesforce": [salesforce],
+"samba": [samba],
+"segment": [segment],
+"sftp": [],
 
 Review comment:
   Right!


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare backporting packages

2020-02-10 Thread GitBox
potiuk commented on a change in pull request #7391: [AIRFLOW-6663] Prepare 
backporting packages
URL: https://github.com/apache/airflow/pull/7391#discussion_r377100501
 
 

 ##
 File path: scripts/ci/ci_prepare_backport_packages.sh
 ##
 @@ -0,0 +1,71 @@
+#!/usr/bin/env bash
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+set -euo pipefail
+
+MY_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
+
+export AIRFLOW_CI_SILENT=${AIRFLOW_CI_SILENT:="true"}
+
+export PYTHON_VERSION=${PYTHON_VERSION:-3.6}
+
+# shellcheck source=scripts/ci/_utils.sh
+. "${MY_DIR}/_utils.sh"
+
+basic_sanity_checks
+
+script_start
+
+cd "${AIRFLOW_SOURCES}"
+
+rm -rf dist/*
+rm -rf -- *.egg-info
+
+BACKPORT_PACKAGES=$(python setup.py --list-backport-packages)
+
+cp -v MANIFEST.in MANIFEST.in.bak
+cp MANIFEST-packages.in MANIFEST.in
+
+function cleanup {
+cp MANIFEST.in.bak MANIFEST.in
+rm MANIFEST.in.bak
+}
+
+trap cleanup EXIT
+
+for BACKPORT_PACKAGE in ${BACKPORT_PACKAGES}
+do
+echo
+echo 
"---"
+echo " Preparing backporting package ${BACKPORT_PACKAGE}"
+echo 
"---"
+echo
+python setup.py --provider-package "${BACKPORT_PACKAGE}" sdist bdist 
bdist_wheel >/dev/null
+done
+
+python setup.py --provider-package providers sdist bdist bdist_wheel >/dev/null
+
+DUMP_FILE="/tmp/airflow_provider_packages_$(date +"%Y%m%d-%H%M%S").tar.gz"
+
+tar -cvzf "${DUMP_FILE}" "dist"
+
+echo "Packages are prepared in ${DUMP_FILE}"
+
+curl -F "file=@${DUMP_FILE}" https://file.io
 
 Review comment:
   I noticed file.io is not super-stable. Some links start to return 404 
earlier than after 14 days.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] potiuk commented on issue #7391: [AIRFLOW-6663] Prepare backporting packages

2020-02-10 Thread GitBox
potiuk commented on issue #7391: [AIRFLOW-6663] Prepare backporting packages
URL: https://github.com/apache/airflow/pull/7391#issuecomment-584152145
 
 
   > Something slightly odd is going on with the naming -- some are using `-`, 
others are using `_`
   > 
   Yeah. That's default naming conventions used by sdist/bdist :(. They are not 
consistent. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (AIRFLOW-6767) Default workgroup name for Athena is incorrect

2020-02-10 Thread Bhavika (Jira)
Bhavika created AIRFLOW-6767:


 Summary: Default workgroup name for Athena is incorrect
 Key: AIRFLOW-6767
 URL: https://issues.apache.org/jira/browse/AIRFLOW-6767
 Project: Apache Airflow
  Issue Type: Improvement
  Components: aws
Affects Versions: 1.10.9
Reporter: Bhavika
Assignee: Bhavika
 Fix For: 1.10.10


The default Athena workgroup name is set to "default" when it should be 
"primary". See this comment - 
[https://github.com/apache/airflow/pull/6871#issuecomment-584066252]

I'll submit a fix for the incorrect name. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [airflow] mrshu commented on issue #6871: [AIRFLOW-6319] Add support for AWS Athena workgroups

2020-02-10 Thread GitBox
mrshu commented on issue #6871: [AIRFLOW-6319] Add support for AWS Athena 
workgroups
URL: https://github.com/apache/airflow/pull/6871#issuecomment-584142723
 
 
   @bhavika Nice -- thanks a lot for the reply!
   
   Please feel free to go ahead and update the `workgroup` default -- it would 
take me much longer to setup :)
   
   Thanks again!


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] codecov-io commented on issue #6337: [AIRFLOW-5659] - Add support for ephemeral storage on KubernetesPodOp…

2020-02-10 Thread GitBox
codecov-io commented on issue #6337: [AIRFLOW-5659] - Add support for ephemeral 
storage on KubernetesPodOp…
URL: https://github.com/apache/airflow/pull/6337#issuecomment-584135185
 
 
   # [Codecov](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=h1) 
Report
   > Merging 
[#6337](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=desc) into 
[master](https://codecov.io/gh/apache/airflow/commit/53ba6d607c9a3548a996c0c716d1ffb8da0b291b?src=pr=desc)
 will **increase** coverage by `0.28%`.
   > The diff coverage is `50%`.
   
   [![Impacted file tree 
graph](https://codecov.io/gh/apache/airflow/pull/6337/graphs/tree.svg?width=650=WdLKlKHOAU=150=pr)](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=tree)
   
   ```diff
   @@Coverage Diff @@
   ##   master#6337  +/-   ##
   ==
   + Coverage86.3%   86.59%   +0.28% 
   ==
 Files 871  871  
 Lines   4066040662   +2 
   ==
   + Hits3509135210 +119 
   + Misses   5569 5452 -117
   ```
   
   
   | [Impacted 
Files](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=tree) | 
Coverage Δ | |
   |---|---|---|
   | 
[airflow/kubernetes/pod\_generator.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3BvZF9nZW5lcmF0b3IucHk=)
 | `95.95% <ø> (ø)` | :arrow_up: |
   | 
[airflow/kubernetes/pod.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3BvZC5weQ==)
 | `92.5% <50%> (+0.39%)` | :arrow_up: |
   | 
[airflow/utils/dag\_processing.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy91dGlscy9kYWdfcHJvY2Vzc2luZy5weQ==)
 | `88.12% <0%> (+0.19%)` | :arrow_up: |
   | 
[...viders/cncf/kubernetes/operators/kubernetes\_pod.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvY25jZi9rdWJlcm5ldGVzL29wZXJhdG9ycy9rdWJlcm5ldGVzX3BvZC5weQ==)
 | `93.61% <0%> (+23.4%)` | :arrow_up: |
   | 
[airflow/kubernetes/refresh\_config.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3JlZnJlc2hfY29uZmlnLnB5)
 | `74.5% <0%> (+23.52%)` | :arrow_up: |
   | 
[airflow/kubernetes/pod\_launcher.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3BvZF9sYXVuY2hlci5weQ==)
 | `92.25% <0%> (+45.07%)` | :arrow_up: |
   | 
[airflow/kubernetes/volume.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3ZvbHVtZS5weQ==)
 | `100% <0%> (+47.05%)` | :arrow_up: |
   | 
[airflow/kubernetes/volume\_mount.py](https://codecov.io/gh/apache/airflow/pull/6337/diff?src=pr=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3ZvbHVtZV9tb3VudC5weQ==)
 | `100% <0%> (+55.55%)` | :arrow_up: |
   
   --
   
   [Continue to review full report at 
Codecov](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=continue).
   > **Legend** - [Click here to learn 
more](https://docs.codecov.io/docs/codecov-delta)
   > `Δ = absolute  (impact)`, `ø = not affected`, `? = missing data`
   > Powered by 
[Codecov](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=footer). 
Last update 
[53ba6d6...594ea16](https://codecov.io/gh/apache/airflow/pull/6337?src=pr=lastupdated).
 Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments).
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] bhavika commented on issue #6871: [AIRFLOW-6319] Add support for AWS Athena workgroups

2020-02-10 Thread GitBox
bhavika commented on issue #6871: [AIRFLOW-6319] Add support for AWS Athena 
workgroups
URL: https://github.com/apache/airflow/pull/6871#issuecomment-584132576
 
 
   @mrshu just wanted to flag that someone else already fixed the WorkGroup 
param name here - https://github.com/apache/airflow/pull/7386
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] ashb commented on a change in pull request #7370: [AIRFLOW-6590] Use batch db operations in jobs

2020-02-10 Thread GitBox
ashb commented on a change in pull request #7370: [AIRFLOW-6590] Use batch db 
operations in jobs
URL: https://github.com/apache/airflow/pull/7370#discussion_r377061966
 
 

 ##
 File path: scripts/perf/sql_queries.py
 ##
 @@ -0,0 +1,178 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+import os
+from time import sleep, time
+from typing import List, NamedTuple, Optional, Tuple
+
+import pandas as pd
+
+# Setup environment before any Airflow import
+DAG_FOLDER = "/opt/airflow/scripts/perf/dags"
+os.environ["AIRFLOW__CORE__DAGS_FOLDER"] = DAG_FOLDER
+os.environ["AIRFLOW__DEBUG__SQLALCHEMY_STATS"] = "True"
+os.environ["AIRFLOW__CORE__LOAD_EXAMPLES"] = "False"
+
+# Here we setup simpler logger to avoid any code changes in
+# Airflow core code base
+LOG_LEVEL = "INFO"
+LOG_FILE = "/files/sql_stats.log"  # Default to run in Breeze
+
+os.environ[
+"AIRFLOW__LOGGING__LOGGING_CONFIG_CLASS"
+] = "scripts.perf.sql_queries.DEBUG_LOGGING_CONFIG"
+
+DEBUG_LOGGING_CONFIG = {
+"version": 1,
+"disable_existing_loggers": False,
+"formatters": {"airflow": {"format": "%(message)s"}},
+"handlers": {
+"console": {"class": "logging.StreamHandler"},
+"task": {
+"class": "logging.FileHandler",
+"formatter": "airflow",
+"filename": LOG_FILE,
+},
+"processor": {
+"class": "logging.FileHandler",
+"formatter": "airflow",
+"filename": LOG_FILE,
+},
+},
+"loggers": {
+"airflow.processor": {
+"handlers": ["processor"],
+"level": LOG_LEVEL,
+"propagate": False,
+},
+"airflow.task": {"handlers": ["task"], "level": LOG_LEVEL, 
"propagate": False},
+"flask_appbuilder": {
+"handler": ["console"],
+"level": LOG_LEVEL,
+"propagate": True,
+},
+},
+"root": {"handlers": ["console", "task"], "level": LOG_LEVEL},
+}
+
+
+class Query(NamedTuple):
+function: str
+file: str
+location: int
+sql: str
+stack: str
+time: float
+
+def __str__(self):
+sql = self.sql if len(self.sql) < 110 else f"{self.sql[:111]}..."
+return f"{self.function} in {self.file}:{self.location}: {sql}"
+
+def __eq__(self, other):
+return (
+self.function == other.function
+and self.sql == other.sql
+and self.location == other.location
+and self.file == other.file
+)
+
+def to_dict(self):
+return dict(zip(("function", "file", "location", "sql", "stack", 
"time"), self))
+
+
+def reset_db():
+from airflow.utils.db import resetdb
+
+resetdb()
+
+
+def run_scheduler_job(with_db_reset=False) -> None:
+from airflow.jobs.scheduler_job import SchedulerJob
+
+if with_db_reset:
+reset_db()
+SchedulerJob(subdir=DAG_FOLDER, do_pickle=False, num_runs=3).run()
+
+
+def is_query(line: str) -> bool:
+return "@SQLALCHEMY" in line and "|$" in line
 
 Review comment:
   How about instead writing the SQLAlchemy logs to a different file? (or a 
different file _and_ the normal log file) -- we've already got a custom logging 
config so setting up an `airflow.debug.sql` logger isn't much more work.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] ashb commented on a change in pull request #7370: [AIRFLOW-6590] Use batch db operations in jobs

2020-02-10 Thread GitBox
ashb commented on a change in pull request #7370: [AIRFLOW-6590] Use batch db 
operations in jobs
URL: https://github.com/apache/airflow/pull/7370#discussion_r377070248
 
 

 ##
 File path: tests/models/test_dagrun.py
 ##
 @@ -227,15 +227,14 @@ def test_dagrun_deadlock(self):
execution_date=now,
start_date=now)
 
+dr.update_state()
+self.assertEqual(dr.state, State.RUNNING)
+
 ti_op1 = dr.get_task_instance(task_id=op1.task_id)
 ti_op1.set_state(state=State.SUCCESS, session=session)
 ti_op2 = dr.get_task_instance(task_id=op2.task_id)
 ti_op2.set_state(state=State.NONE, session=session)
 
-dr.update_state()
-self.assertEqual(dr.state, State.RUNNING)
-
-ti_op2.set_state(state=State.NONE, session=session)
 
 Review comment:
   What was the reason for this change?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [airflow] ashb commented on a change in pull request #7370: [AIRFLOW-6590] Use batch db operations in jobs

2020-02-10 Thread GitBox
ashb commented on a change in pull request #7370: [AIRFLOW-6590] Use batch db 
operations in jobs
URL: https://github.com/apache/airflow/pull/7370#discussion_r377034240
 
 

 ##
 File path: airflow/config_templates/config.yml
 ##
 @@ -495,6 +495,13 @@
   type: string
   example: ~
   default: "False"
+- name: sqlalchemy_stats
+  description: |
+Set to True to enable logging of SQLAlchemy queries and execution 
times.
 
 Review comment:
   Is this something we think people will want to turn on in their clusters? If 
we think it's only for developing airflow itself then perhaps it makes sense to 
not put it in the config file (and use `fallback=False`)


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


  1   2   >