Thanks for picking this up. Your fix should stop the 500 error, but there's 
another problem (which is ultimately user misconfiguration about) 
https://issues.apache.org/jira/browse/AIRFLOW-1796 - the fix for that to update 
a doc somewhere, and probably validate this settings is correct at start time.


I've found another issue related to arg names of S3Hook. In 1.8.2 it was 
`s3_conn_id` but the move to boto3/basing off AWSHook now expects 
`aws_conn_id`, and various places in Airflow code base (and a few places in our 
dags/operators code base) still pass it as s3_conn_id. I've created 
https://issues.apache.org/jira/browse/AIRFLOW-1795 for that issue.

-ash


> On 8 Nov 2017, at 18:54, Daniel Huang <[email protected]> wrote:
> 
> Still testing this out.
> 
> Put up a small fix for Ash's second exception
> https://github.com/apache/incubator-airflow/pull/2766
> 
> On Wed, Nov 8, 2017 at 10:48 AM, Bolke de Bruin <[email protected]> wrote:
> 
>> Hi Chris,
>> 
>> Actively testing here: we found an issue in the SSHOperator introduced in
>> 1.9.0 (fix already merged for RC2, but blocking I as it stops us from
>> running SSH properly), some minor fixes by Airbnb should also be in RC2.
>> There is some logging “weirdness”, that might warrant a small patch here in
>> there and could be squeezed into RC2, but I don’t consider them blocking.
>> 
>> So almost there, but we need an RC2 imho.
>> 
>> -1, binding.
>> 
>> Bolke
>> 
>>> On 8 Nov 2017, at 19:00, Ash Berlin-Taylor <ash_airflowlist@firemirror.
>> com> wrote:
>>> 
>>> -1 (for now. Non binding. Is that how this process works?)
>>> 
>>> We've built a test env for this RC and are testing, but have run into an
>> issue reading task logs. (See below)
>>> 
>>> We haven't gotten very far with this yet, we will dig more tomorrow
>> (it's the end of the UK work day now). I suspect this might be how we've
>> misconfigured our logging. We will see tomorrow.
>>> 
>>> -ash
>>> 
>>> 
>>> 
>>> 
>>> File "/usr/local/lib/python3.5/dist-packages/airflow/www/views.py",
>> line 712, in log
>>>   logs = handler.read(ti)
>>> AttributeError: 'NoneType' object has no attribute 'read'
>>> 
>>> During handling of the above exception, another exception occurred:
>>> 
>>> Traceback (most recent call last):
>>> File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1988,
>> in wsgi_app
>>>   response = self.full_dispatch_request()
>>> File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1641,
>> in full_dispatch_request
>>>   rv = self.handle_user_exception(e)
>>> File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1544,
>> in handle_user_exception
>>>   reraise(exc_type, exc_value, tb)
>>> File "/usr/local/lib/python3.5/dist-packages/flask/_compat.py", line
>> 33, in reraise
>>>   raise value
>>> File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1639,
>> in full_dispatch_request
>>>   rv = self.dispatch_request()
>>> File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1625,
>> in dispatch_request
>>>   return self.view_functions[rule.endpoint](**req.view_args)
>>> File "/usr/local/lib/python3.5/dist-packages/flask_admin/base.py",
>> line 69, in inner
>>>   return self._run_view(f, *args, **kwargs)
>>> File "/usr/local/lib/python3.5/dist-packages/flask_admin/base.py",
>> line 368, in _run_view
>>>   return fn(self, *args, **kwargs)
>>> File "/usr/local/lib/python3.5/dist-packages/flask_login.py", line
>> 758, in decorated_view
>>>   return func(*args, **kwargs)
>>> File "/usr/local/lib/python3.5/dist-packages/airflow/www/utils.py",
>> line 262, in wrapper
>>>   return f(*args, **kwargs)
>>> File "/usr/local/lib/python3.5/dist-packages/airflow/www/views.py",
>> line 715, in log
>>>   .format(task_log_reader, e.message)]
>>> AttributeError: 'AttributeError' object has no attribute 'message'
>>> 
>>> 
>>>> On 8 Nov 2017, at 17:46, Chris Riccomini <[email protected]> wrote:
>>>> 
>>>> Anyone? :/
>>>> 
>>>> On Mon, Nov 6, 2017 at 1:22 PM, Chris Riccomini <[email protected]>
>>>> wrote:
>>>> 
>>>>> Hey all,
>>>>> 
>>>>> I have cut Airflow 1.9.0 RC1. This email is calling a vote on the
>> release,
>>>>> which will last fo 72 hours. Consider this my (binding) +1.
>>>>> 
>>>>> Airflow 1.9.0 RC1 is available at:
>>>>> 
>>>>> https://dist.apache.org/repos/dist/dev/incubator/airflow/1.9.0rc1/
>>>>> 
>>>>> apache-airflow-1.9.0rc1+incubating-source.tar.gz is a source release
>> that
>>>>> comes with INSTALL instructions.
>>>>> apache-airflow-1.9.0rc1+incubating-bin.tar.gz is the binary Python
>>>>> "sdist" release.
>>>>> 
>>>>> Public keys are available at:
>>>>> 
>>>>> https://dist.apache.org/repos/dist/release/incubator/airflow/
>>>>> 
>>>>> The release contains the following JIRAs:
>>>>> 
>>>>> ISSUE ID    |DESCRIPTION                                       |PR
>>>>> |COMMIT
>>>>> AIRFLOW-1779|Add keepalive packets to ssh hook                 |#2749
>>>>> |d2f9d1
>>>>> AIRFLOW-1776|stdout/stderr logging not captured                |#2745
>>>>> |590d9f
>>>>> AIRFLOW-1771|Change heartbeat text from boom to heartbeat      |-
>> |-
>>>>> 
>>>>> AIRFLOW-1767|Airflow Scheduler no longer schedules DAGs        |-
>> |-
>>>>> 
>>>>> AIRFLOW-1765|Default API auth backed should deny all.          |#2737
>>>>> |6ecdac
>>>>> AIRFLOW-1764|Web Interface should not use experimental api     |#2738
>>>>> |6bed1d
>>>>> AIRFLOW-1757|Contrib.SparkSubmitOperator should allow --package|#2725
>>>>> |4e06ee
>>>>> AIRFLOW-1745|BashOperator ignores SIGPIPE in subprocess        |#2714
>>>>> |e021c9
>>>>> AIRFLOW-1744|task.retries can be False                         |#2713
>>>>> |6144c6
>>>>> AIRFLOW-1743|Default config template should not contain ldap fi|#2712
>>>>> |270684
>>>>> AIRFLOW-1741|Task Duration shows two charts on first page load.|#2711
>>>>> |974b49
>>>>> AIRFLOW-1734|Sqoop Operator contains logic errors & needs optio|#2703
>>>>> |f6810c
>>>>> AIRFLOW-1731|Import custom config on PYTHONPATH                |#2721
>>>>> |f07eb3
>>>>> AIRFLOW-1726|Copy Expert command for Postgres Hook             |#2698
>>>>> |8a4ad3
>>>>> AIRFLOW-1719|Fix small typo - your vs you                      |-
>> |-
>>>>> 
>>>>> AIRFLOW-1712|Log SSHOperator output                            |-
>> |-
>>>>> 
>>>>> AIRFLOW-1711|Ldap Attributes not always a "list" part 2        |#2731
>>>>> |40a936
>>>>> AIRFLOW-1706|Scheduler is failed on startup with MS SQL Server |#2733
>>>>> |9e209b
>>>>> AIRFLOW-1698|Remove confusing SCHEDULER_RUNS env var from syste|#2677
>>>>> |00dd06
>>>>> AIRFLOW-1695|Redshift Hook using boto3 & AWS Hook              |#2717
>>>>> |bfddae
>>>>> AIRFLOW-1694|Hive Hooks: Python 3 does not have an `itertools.i|#2674
>>>>> |c6e5ae
>>>>> AIRFLOW-1692|Master cannot be checked out on windows           |#2673
>>>>> |31805e
>>>>> AIRFLOW-1691|Add better documentation for Google cloud storage |#2671
>>>>> |ace2b1
>>>>> AIRFLOW-1690|Error messages regarding gcs log commits are spars|#2670
>>>>> |5fb5cd
>>>>> AIRFLOW-1682|S3 task handler never writes to S3                |#2664
>>>>> |0080f0
>>>>> AIRFLOW-1678|Fix docstring errors for `set_upstream` and `set_d|-
>> |-
>>>>> 
>>>>> AIRFLOW-1677|Fix typo in example_qubole_operator               |-
>> |-
>>>>> 
>>>>> AIRFLOW-1676|GCS task handler never writes to GCS              |#2659
>>>>> |781fa4
>>>>> AIRFLOW-1675|Fix API docstrings to be properly rendered        |#2667
>>>>> |f12381
>>>>> AIRFLOW-1671|Missing @apply_defaults annotation for gcs downloa|#2655
>>>>> |97666b
>>>>> AIRFLOW-1669|Fix Docker import in Master                       |#na
>>>>> |f7f2a8
>>>>> AIRFLOW-1668|Redhsift requires a keep alive of < 300s          |#2650
>>>>> |f2bb77
>>>>> AIRFLOW-1664|Make MySqlToGoogleCloudStorageOperator support bin|#2649
>>>>> |95813d
>>>>> AIRFLOW-1660|Change webpage width to full-width                |#2646
>>>>> |8ee3d9
>>>>> AIRFLOW-1659|Fix invalid attribute bug in FileTaskHandler      |#2645
>>>>> |bee823
>>>>> AIRFLOW-1658|Kill (possibly) still running Druid indexing job a|#2644
>>>>> |cbf7ad
>>>>> AIRFLOW-1657|Handle failure of Qubole Operator for s3distcp had|-
>> |-
>>>>> 
>>>>> AIRFLOW-1654|Show tooltips for link icons in DAGs view         |#2642
>>>>> |ada7b2
>>>>> AIRFLOW-1647|Fix Spark-sql hook                                |#2637
>>>>> |b1e5c6
>>>>> AIRFLOW-1641|Task gets stuck in queued state                   |#2715
>>>>> |735497
>>>>> AIRFLOW-1640|Add Qubole default connection in connection table |-
>> |-
>>>>> 
>>>>> AIRFLOW-1639|ValueError does not have .message attribute       |#2629
>>>>> |87df67
>>>>> AIRFLOW-1637|readme not tracking master branch for travis      |-
>> |-
>>>>> 
>>>>> AIRFLOW-1636|aws and emr connection types get cleared          |#2626
>>>>> |540e04
>>>>> AIRFLOW-1635|Allow creating Google Cloud Platform connection wi|#2640
>>>>> |6dec7a
>>>>> AIRFLOW-1629|make extra a textarea in edit connections form    |#2623
>>>>> |f5d46f
>>>>> AIRFLOW-1628|Docstring of sqlsensor is incorrect               |#2621
>>>>> |9ba73d
>>>>> AIRFLOW-1627|SubDagOperator initialization should only query po|#2620
>>>>> |516ace
>>>>> AIRFLOW-1621|Add tests for logic added on server side dag list |#2614
>>>>> |8de9fd
>>>>> AIRFLOW-1614|Improve performance of DAG parsing when there are |#2610
>>>>> |a95adb
>>>>> AIRFLOW-1611|Customize logging in Airflow                      |#2631
>>>>> |8b4a50
>>>>> AIRFLOW-1609|Ignore all venvs in gitignore                     |#2608
>>>>> |f1f9b4
>>>>> AIRFLOW-1608|GCP Dataflow hook missing pending job state       |#2607
>>>>> |653562
>>>>> AIRFLOW-1606|DAG.sync_to_db is static, but takes a DAG as first|#2606
>>>>> |6ac296
>>>>> AIRFLOW-1605|Fix log source of local loggers                   |-
>> |-
>>>>> 
>>>>> AIRFLOW-1604|Rename the logger to log                          |#2604
>>>>> |af4050
>>>>> AIRFLOW-1602|Use LoggingMixin for the DAG class                |#2602
>>>>> |956699
>>>>> AIRFLOW-1601|Add configurable time between SIGTERM and SIGKILL |#2601
>>>>> |48a95e
>>>>> AIRFLOW-1600|Uncaught exceptions in get_fernet if cryptography |#2600
>>>>> |ad963e
>>>>> AIRFLOW-1597|Add GameWisp as Airflow user                      |#2599
>>>>> |26b747
>>>>> AIRFLOW-1594|Installing via pip copies test files into python l|#2597
>>>>> |a6b23a
>>>>> AIRFLOW-1593|Expose load_string in WasbHook                    |#2596
>>>>> |7ece95
>>>>> AIRFLOW-1591|Exception: 'TaskInstance' object has no attribute |#2578
>>>>> |f4653e
>>>>> AIRFLOW-1590|Small fix for dates util                          |#2652
>>>>> |31946e
>>>>> AIRFLOW-1587|fix `ImportError: cannot import name 'CeleryExecut|#2590
>>>>> |34c73b
>>>>> AIRFLOW-1586|MySQL to GCS to BigQuery fails for tables with dat|#2589
>>>>> |e83012
>>>>> AIRFLOW-1584|Remove the insecure /headers endpoints            |#2588
>>>>> |17ac07
>>>>> AIRFLOW-1582|Improve logging structure of Airflow              |#2592
>>>>> |a7a518
>>>>> AIRFLOW-1580|Error in string formatter when throwing an excepti|#2583
>>>>> |ea9ab9
>>>>> AIRFLOW-1579|Allow jagged rows in BQ Hook.                     |#2582
>>>>> |5b978b
>>>>> AIRFLOW-1577|Add token support to DatabricksHook               |#2579
>>>>> |c2c515
>>>>> AIRFLOW-1573|Remove `thrift < 0.10.0` requirement              |#2574
>>>>> |aa95f2
>>>>> AIRFLOW-1571|Add AWS Lambda Hook for invoking Lambda Function  |#2718
>>>>> |017f18
>>>>> AIRFLOW-1568|Add datastore import/export operator              |#2568
>>>>> |86063b
>>>>> AIRFLOW-1567|Clean up ML Engine operators                      |#2567
>>>>> |af91e2
>>>>> AIRFLOW-1564|Default logging filename contains a colon         |#2565
>>>>> |4c674c
>>>>> AIRFLOW-1560|Add AWS DynamoDB hook for inserting batch items   |#2587
>>>>> |71400b
>>>>> AIRFLOW-1556|BigQueryBaseCursor should support SQL parameters  |#2557
>>>>> |9df0ac
>>>>> AIRFLOW-1546| add Zymergen to org list in README               |#2512
>>>>> |7cc346
>>>>> AIRFLOW-1535|Add support for Dataproc serviceAccountScopes in D|#2546
>>>>> |b1f902
>>>>> AIRFLOW-1529|Support quoted newlines in Google BigQuery load jo|#2545
>>>>> |4a4b02
>>>>> AIRFLOW-1527|Refactor celery config to make use of template    |#2542
>>>>> |f4437b
>>>>> AIRFLOW-1522|Increase size of val column for variable table in |#2535
>>>>> |8a2d24
>>>>> AIRFLOW-1521|Template fields definition for bigquery_table_dele|#2534
>>>>> |f1a7c0
>>>>> AIRFLOW-1520|S3Hook uses boto2                                 |#2532
>>>>> |386583
>>>>> AIRFLOW-1519|Main DAG list page does not scale using client sid|#2531
>>>>> |d7d7ce
>>>>> AIRFLOW-1512|Add operator for running Python functions in a vir|#2446
>>>>> |14e6d7
>>>>> AIRFLOW-1507|Make src, dst and bucket parameters as templated i|#2516
>>>>> |d295cf
>>>>> AIRFLOW-1505|Document when Jinja substitution occurs           |#2523
>>>>> |984a87
>>>>> AIRFLOW-1504|Log Cluster Name on Dataproc Operator When Execute|#2517
>>>>> |1cd6c4
>>>>> AIRFLOW-1499s|Eliminate duplicate and unneeded code             |-
>> |-
>>>>> 
>>>>> AIRFLOW-1497|Hidden fields in connection form aren't reset when|#2507
>>>>> |d8da8b
>>>>> AIRFLOW-1493|Fix race condition with airflow run               |#2505
>>>>> |b2e175
>>>>> AIRFLOW-1492|Add metric for task success/failure               |#2504
>>>>> |fa84d4
>>>>> AIRFLOW-1489|Docs: Typo in BigQueryCheckOperator               |#2501
>>>>> |111ce5
>>>>> AIRFLOW-1483|Page size on model views is to large to render qui|#2497
>>>>> |04bfba
>>>>> AIRFLOW-1478|Chart -> Owner column should be sortable          |#2493
>>>>> |651e60
>>>>> AIRFLOW-1476|Add INSTALL file for source releases              |#2492
>>>>> |da76ac
>>>>> AIRFLOW-1474|Add dag_id regex for 'airflow clear' CLI command  |#2486
>>>>> |18f849
>>>>> AIRFLOW-1470s|BashSensor Implementation                         |-
>> |-
>>>>> 
>>>>> AIRFLOW-1459|integration rst doc is broken in github view      |#2481
>>>>> |322ec9
>>>>> AIRFLOW-1438|Scheduler batch queries should have a limit       |#2462
>>>>> |3547cb
>>>>> AIRFLOW-1437|BigQueryTableDeleteOperator should define deletion|#2459
>>>>> |b87903
>>>>> AIRFLOW-1432|NVD3 Charts do not have labeled axes and units cha|#2710
>>>>> |70ffa4
>>>>> AIRFLOW-1402|Cleanup SafeConfigParser DeprecationWarning       |#2435
>>>>> |38c86b
>>>>> AIRFLOW-1401|Standardize GCP project, region, and zone argument|#2439
>>>>> |b6d363
>>>>> AIRFLOW-1397|Airflow 1.8.1 - No data displays in Last Run Colum|-
>> |-
>>>>> 
>>>>> AIRFLOW-1394|Add quote_character parameter to GoogleCloudStorag|#2428
>>>>> |9fd0be
>>>>> AIRFLOW-1389|BigQueryOperator should support `createDisposition|#2470
>>>>> |6e2640
>>>>> AIRFLOW-1384|Add ARGO/CaDC                                     |#2434
>>>>> |715947
>>>>> AIRFLOW-1368|Automatically remove the container when it exits  |#2653
>>>>> |d42d23
>>>>> AIRFLOW-1359|Provide GoogleCloudML operator for model evaluatio|#2407
>>>>> |194d1d
>>>>> AIRFLOW-1356|add `--celery_hostname` to `airflow worker`       |#2405
>>>>> |b9d7d1
>>>>> AIRFLOW-1352|Revert bad logging Handler                        |-
>> |-
>>>>> 
>>>>> AIRFLOW-1350|Add "query_uri" parameter for Google DataProc oper|#2402
>>>>> |d32c72
>>>>> AIRFLOW-1348|Paginated UI has broken toggles after first page  |-
>> |-
>>>>> 
>>>>> AIRFLOW-1345|Don't commit on each loop                         |#2397
>>>>> |0dd002
>>>>> AIRFLOW-1344|Builds failing on Python 3.5 with AttributeError  |#2394
>>>>> |2a5883
>>>>> AIRFLOW-1343|Add airflow default label to the dataproc operator|#2396
>>>>> |e4b240
>>>>> AIRFLOW-1338|gcp_dataflow_hook is incompatible with the recent |#2388
>>>>> |cf2605
>>>>> AIRFLOW-1337|Customize log format via config file              |#2392
>>>>> |4841e3
>>>>> AIRFLOW-1335|Use buffered logger                               |#2386
>>>>> |0d23d3
>>>>> AIRFLOW-1333|Enable copy function for Google Cloud Storage Hook|#2385
>>>>> |e2c383
>>>>> AIRFLOW-1331|Contrib.SparkSubmitOperator should allow --package|#2622
>>>>> |fbca8f
>>>>> AIRFLOW-1330|Connection.parse_from_uri doesn't work for google_|#2525
>>>>> |6e5e9d
>>>>> AIRFLOW-1324|Make the Druid operator/hook more general         |#2378
>>>>> |de99aa
>>>>> AIRFLOW-1323|Operators related to Dataproc should keep some par|#2636
>>>>> |ed248d
>>>>> AIRFLOW-1315|Add Qubole File and Partition Sensors             |-
>> |-
>>>>> 
>>>>> AIRFLOW-1309|Add optional hive_tblproperties in HiveToDruidTran|-
>> |-
>>>>> 
>>>>> AIRFLOW-1301|Add New Relic to Airflow user list                |#2359
>>>>> |355fc9
>>>>> AIRFLOW-1299|Google Dataproc cluster creation operator should s|#2358
>>>>> |c2b80e
>>>>> AIRFLOW-1289|Don't restrict scheduler threads to CPU cores     |#2353
>>>>> |8e23d2
>>>>> AIRFLOW-1286|BaseTaskRunner - Exception TypeError: a bytes-like|#2363
>>>>> |d8891d
>>>>> AIRFLOW-1277|Forbid creation of a known event with empty fields|#na
>>>>> |65184a
>>>>> AIRFLOW-1276|Forbid event creation with end_data earlier than s|#na
>>>>> |d5d02f
>>>>> AIRFLOW-1275|Fix `airflow pool` command exception              |#2346
>>>>> |9958aa
>>>>> AIRFLOW-1273|Google Cloud ML Version and Model CRUD Operator   |#2379
>>>>> |534a0e
>>>>> AIRFLOW-1272|Google Cloud ML Batch Prediction Operator         |#2390
>>>>> |e92d6b
>>>>> AIRFLOW-1271|Google Cloud ML Training Operator                 |#2408
>>>>> |0fc450
>>>>> AIRFLOW-1256|Add United Airlines as Airflow user               |#2332
>>>>> |d3484a
>>>>> AIRFLOW-1251|Add eRevalue as an Airflow user                   |#2331
>>>>> |8d5160
>>>>> AIRFLOW-1248|Fix inconsistent configuration name for worker tim|#2328
>>>>> |92314f
>>>>> AIRFLOW-1247|CLI: ignore all dependencies argument ignored     |#2441
>>>>> |e88ecf
>>>>> AIRFLOW-1245|Fix random failure of test_trigger_dag_for_date un|#2325
>>>>> |cef01b
>>>>> AIRFLOW-1244|Forbid creation of a pool with empty name         |#2324
>>>>> |df9a10
>>>>> AIRFLOW-1242|BigQueryHook assumes that a valid project_id can't|#2335
>>>>> |ffe616
>>>>> AIRFLOW-1237|Fix IN-predicate sqlalchemy warning               |#2320
>>>>> |a1f422
>>>>> AIRFLOW-1234|Cover utils.operator_helpers with unit tests      |#2317
>>>>> |d16537
>>>>> AIRFLOW-1233|Cover utils.json with unit tests                  |#2316
>>>>> |502410
>>>>> AIRFLOW-1232|Remove deprecated readfp warning                  |#2315
>>>>> |6ffaaf
>>>>> AIRFLOW-1231|Use flask_wtf.CSRFProtect instead of flask_wtf.Csr|#2313
>>>>> |cac49e
>>>>> AIRFLOW-1221|Fix DatabricksSubmitRunOperator Templating        |#2308
>>>>> |0fa104
>>>>> AIRFLOW-1217|Enable logging in Sqoop hook                      |#2307
>>>>> |4f459b
>>>>> AIRFLOW-1213|Add hcatalog parameters to the sqoop operator/hook|#2305
>>>>> |857850
>>>>> AIRFLOW-1208|Speed-up cli tests                                |#2301
>>>>> |21c142
>>>>> AIRFLOW-1207|Enable utils.helpers unit tests                   |#2300
>>>>> |8ac87b
>>>>> AIRFLOW-1203|Tests failing after oauth upgrade                 |#2296
>>>>> |3e9c66
>>>>> AIRFLOW-1201|Update deprecated 'nose-parameterized' library to |#2298
>>>>> |d2d3e4
>>>>> AIRFLOW-1193|Add Checkr to Airflow user list                   |#2276
>>>>> |707238
>>>>> AIRFLOW-1189|Get pandas DataFrame using BigQueryHook fails     |#2287
>>>>> |93666f
>>>>> AIRFLOW-1188|Add max_bad_records param to GoogleCloudStorageToB|#2286
>>>>> |443e6b
>>>>> AIRFLOW-1187|Obsolete package names in documentation           |-
>> |-
>>>>> 
>>>>> AIRFLOW-1185|Incorrect url to PyPi                             |#2283
>>>>> |829755
>>>>> AIRFLOW-1181|Enable delete and list function for Google Cloud S|#2281
>>>>> |24f73c
>>>>> AIRFLOW-1179|Pandas 0.20 broke Google BigQuery hook            |#2279
>>>>> |ac9ccb
>>>>> AIRFLOW-1177|variable json deserialize does not work at set def|#2540
>>>>> |65319a
>>>>> AIRFLOW-1175|Add Pronto Tools to Airflow user list             |#2277
>>>>> |86aafa
>>>>> AIRFLOW-1173|Add Robinhood to list of Airflow users            |#2271
>>>>> |379115
>>>>> AIRFLOW-1165|airflow webservice crashes on ubuntu16 - python3  |-
>> |-
>>>>> 
>>>>> AIRFLOW-1160|Upadte SparkSubmitOperator parameters             |#2265
>>>>> |2e3f07
>>>>> AIRFLOW-1155|Add Tails.com to community                        |#2261
>>>>> |2fa690
>>>>> AIRFLOW-1149|Allow custom filters to be added to jinja2        |#2258
>>>>> |48135a
>>>>> AIRFLOW-1141|Remove DAG.crawl_for_tasks method                 |#2275
>>>>> |a30fee
>>>>> AIRFLOW-1140|DatabricksSubmitRunOperator should template the "j|#2255
>>>>> |e6d316
>>>>> AIRFLOW-1136|Invalid parameters are not captured for Sqoop oper|#2252
>>>>> |2ef4db
>>>>> AIRFLOW-1125|Clarify documentation regarding fernet_key        |#2251
>>>>> |831f8d
>>>>> AIRFLOW-1122|Node strokes are too thin for people with color vi|#2246
>>>>> |a08761
>>>>> AIRFLOW-1121|airflow webserver --pid no longer write out pid fi|-
>> |-
>>>>> 
>>>>> AIRFLOW-1118|Add evo.company to Airflow users                  |#2243
>>>>> |f16914
>>>>> AIRFLOW-1112|Log which pool is full in scheduler when pool slot|#2242
>>>>> |74c1ce
>>>>> AIRFLOW-1107|Add support for ftps non-default port             |#2240
>>>>> |4d0c2f
>>>>> AIRFLOW-1106|Add Groupalia/Letsbonus                           |#2239
>>>>> |945b42
>>>>> AIRFLOW-1095|ldap_auth memberOf should come from configuration |#2232
>>>>> |6b1c32
>>>>> AIRFLOW-1094|Invalid unit tests under `contrib/`               |#2234
>>>>> |219c50
>>>>> AIRFLOW-1091|As a release manager I want to be able to compare |#2231
>>>>> |bfae42
>>>>> AIRFLOW-1090|Add HBO                                           |#2230
>>>>> |177d34
>>>>> AIRFLOW-1089|Add Spark application arguments to SparkSubmitOper|#2229
>>>>> |e5b914
>>>>> AIRFLOW-1081|Task duration page is slow                        |#2226
>>>>> |0da512
>>>>> AIRFLOW-1075|Cleanup security docs                             |#2222
>>>>> |5a6f18
>>>>> AIRFLOW-1065|Add functionality for Azure Blob Storage          |#2216
>>>>> |f1bc5f
>>>>> AIRFLOW-1059|Reset_state_for_orphaned_task should operate in ba|#2205
>>>>> |e05d3b
>>>>> AIRFLOW-1058|Improvements for SparkSubmitOperator              |-
>> |-
>>>>> 
>>>>> AIRFLOW-1051|Add a test for resetdb to CliTests                |#2198
>>>>> |15aee0
>>>>> AIRFLOW-1047|Airflow logs vulnerable to XSS                    |#2193
>>>>> |fe9ebe
>>>>> AIRFLOW-1045|Make log level configurable via airflow.cfg       |#2191
>>>>> |e739a5
>>>>> AIRFLOW-1043|Documentation issues for operators                |#2188
>>>>> |b55f41
>>>>> AIRFLOW-1041|DockerOperator replaces its xcom_push method with |#2274
>>>>> |03704c
>>>>> AIRFLOW-1040|Fix typos in comments/docstrings in models.py     |#2174
>>>>> |d8c0f5
>>>>> AIRFLOW-1036|Exponential backoff should use randomization      |#2262
>>>>> |66168e
>>>>> AIRFLOW-1035|Exponential backoff retry logic should use 2 as ba|#2196
>>>>> |4ec932
>>>>> AIRFLOW-1034|Make it possible to connect to S3 in sigv4 regions|#2181
>>>>> |4c0905
>>>>> AIRFLOW-1031|'scheduled__' may replace with DagRun.ID_PREFIX in|#2613
>>>>> |aa3844
>>>>> AIRFLOW-1030|HttpHook error when creating HttpSensor           |-
>> |-
>>>>> 
>>>>> AIRFLOW-1028|Databricks Operator for Airflow                   |#2202
>>>>> |53ca50
>>>>> AIRFLOW-1024|Handle CeleryExecutor errors gracefully           |#2355
>>>>> |7af20f
>>>>> AIRFLOW-1018|Scheduler DAG processes can not log to stdout     |#2728
>>>>> |ef775d
>>>>> AIRFLOW-1016|Allow HTTP HEAD request method on HTTPSensor      |#2175
>>>>> |4c41f6
>>>>> AIRFLOW-1010|Add a convenience script for signing              |#2169
>>>>> |a2b65a
>>>>> AIRFLOW-1009|Remove SQLOperator from Concepts page             |#2168
>>>>> |7d1144
>>>>> AIRFLOW-1007|Jinja sandbox is vulnerable to RCE                |#2184
>>>>> |daa281
>>>>> AIRFLOW-1005|Speed up Airflow startup time                     |#na
>>>>> |996dd3
>>>>> AIRFLOW-999 |Support for Redis database                        |#2165
>>>>> |8de850
>>>>> AIRFLOW-997 |Change setup.cfg to point to Apache instead of Max|#na
>>>>> |75cd46
>>>>> AIRFLOW-995 |Update Github PR template                         |#2163
>>>>> |b62485
>>>>> AIRFLOW-994 |Add MiNODES to the AIRFLOW Active Users List      |#2159
>>>>> |ca1623
>>>>> AIRFLOW-991 |Mark_success while a task is running leads to fail|-
>> |-
>>>>> 
>>>>> AIRFLOW-990 |DockerOperator fails when logging unicode string  |#2155
>>>>> |6bbf54
>>>>> AIRFLOW-988 |SLA Miss Callbacks Are Repeated if Email is Not be|#2415
>>>>> |6e74d4
>>>>> AIRFLOW-985 |Extend the sqoop operator/hook with additional par|#2177
>>>>> |82eb20
>>>>> AIRFLOW-984 |Subdags unrecognized when subclassing SubDagOperat|#2152
>>>>> |a8bd16
>>>>> AIRFLOW-979 |Add GovTech GDS                                   |#2149
>>>>> |b17bd3
>>>>> AIRFLOW-976 |Mark success running task causes it to fail       |-
>> |-
>>>>> 
>>>>> AIRFLOW-969 |Catch bad python_callable argument at DAG construc|#2142
>>>>> |12901d
>>>>> AIRFLOW-963 |Some code examples are not rendered in the airflow|#2139
>>>>> |f69c1b
>>>>> AIRFLOW-960 |Add support for .editorconfig                     |#na
>>>>> |f5cacc
>>>>> AIRFLOW-959 |.gitignore file is disorganized and incomplete    |#na
>>>>> |3d3c14
>>>>> AIRFLOW-958 |Improve tooltip readability                       |#2134
>>>>> |b3c3eb
>>>>> AIRFLOW-950 |Missing AWS integrations on documentation::integra|#2552
>>>>> |01be02
>>>>> AIRFLOW-947 |Make PrestoHook surface better messages when the P|#na
>>>>> |6dd4b3
>>>>> AIRFLOW-945 |Revert psycopg2 workaround when psycopg2 2.7.1 is |-
>> |-
>>>>> 
>>>>> AIRFLOW-943 |Add Digital First Media to the Airflow users list |#2115
>>>>> |2cfe28
>>>>> AIRFLOW-942 |Add mytaxi to Airflow Users                       |#2111
>>>>> |d579e6
>>>>> AIRFLOW-935 |Impossible to use plugin executors                |#2120
>>>>> |08a784
>>>>> AIRFLOW-926 |jdbc connector is broken due to jaydebeapi api upd|#2651
>>>>> |07ed29
>>>>> AIRFLOW-917 |Incorrectly formatted failure status message      |#2109
>>>>> |b8164c
>>>>> AIRFLOW-916 |Fix ConfigParser deprecation warning              |#2108
>>>>> |ef6dd1
>>>>> AIRFLOW-911 |Add colouring and profiling info on tests         |#2106
>>>>> |4f52db
>>>>> AIRFLOW-903 |Add configuration setting for default DAG view.   |#2103
>>>>> |cadfae
>>>>> AIRFLOW-896 |BigQueryOperator fails to execute with certain inp|#2097
>>>>> |2bceee
>>>>> AIRFLOW-891 |Webserver Clock Should Include Day                |-
>> |-
>>>>> 
>>>>> AIRFLOW-889 |Minor error in the docstrings for BaseOperator.   |#2084
>>>>> |50702d
>>>>> AIRFLOW-887 |Add compatibility with future v0.16               |#na
>>>>> |50902d
>>>>> AIRFLOW-886 |Pass Operator result to post_execute hook         |#na
>>>>> |4da361
>>>>> AIRFLOW-885 |Add Change.org to the list of Airflow users       |#2089
>>>>> |a279be
>>>>> AIRFLOW-882 |Code example in docs has unnecessary DAG>>Operator|#2088
>>>>> |baa4cd
>>>>> AIRFLOW-881 |Create SubDagOperator within DAG context manager w|#2087
>>>>> |0ed608
>>>>> AIRFLOW-880 |Fix remote log functionality inconsistencies for W|#2086
>>>>> |974b75
>>>>> AIRFLOW-877 |GoogleCloudStorageDownloadOperator: template_ext c|#2083
>>>>> |debc69
>>>>> AIRFLOW-875 |Allow HttpSensor params to be templated           |#2080
>>>>> |62f503
>>>>> AIRFLOW-871 |multiple places use logging.warn() instead of warn|#2082
>>>>> |21d775
>>>>> AIRFLOW-866 |Add FTPSensor                                     |#2070
>>>>> |5f87f8
>>>>> AIRFLOW-863 |Example DAG start dates should be recent to avoid |#2068
>>>>> |bbfd43
>>>>> AIRFLOW-862 |Add DaskExecutor                                  |#2067
>>>>> |6e2210
>>>>> AIRFLOW-860 |Circular module dependency prevents loading of cus|-
>> |-
>>>>> 
>>>>> AIRFLOW-854 |Add Open Knowledge International to Airflow users |#2061
>>>>> |51a311
>>>>> AIRFLOW-842 |scheduler.clean_dirty raises warning: SAWarning: T|#2072
>>>>> |485280
>>>>> AIRFLOW-840 |Python3 encoding issue in Kerberos                |#2158
>>>>> |639336
>>>>> AIRFLOW-836 |The paused and queryview endpoints are vulnerable |#2054
>>>>> |6aca2c
>>>>> AIRFLOW-831 |Fix broken unit tests                             |#2050
>>>>> |b86194
>>>>> AIRFLOW-830 |Plugin manager should log to debug, not info      |-
>> |-
>>>>> 
>>>>> AIRFLOW-829 |Reduce verbosity of successful Travis unit tests  |-
>> |-
>>>>> 
>>>>> AIRFLOW-826 |Add Zendesk Hook                                  |#2066
>>>>> |a09762
>>>>> AIRFLOW-823 |Make task instance details available via API      |#2045
>>>>> |3f546e
>>>>> AIRFLOW-822 |Close the connection before throwing exception in |#2038
>>>>> |4b6c38
>>>>> AIRFLOW-821 |Scheduler dagbag importing not Py3 compatible     |#2039
>>>>> |fbb59b
>>>>> AIRFLOW-809 |SqlAlchemy is_ ColumnOperator Causing Errors in MS|-
>> |-
>>>>> 
>>>>> AIRFLOW-802 |Integration of spark-submit                       |-
>> |-
>>>>> 
>>>>> AIRFLOW-781 |Allow DataFlowJavaOperator to accept jar file stor|#2037
>>>>> |259c86
>>>>> AIRFLOW-770 |HDFS hooks should support alternative ways of gett|#2056
>>>>> |261b65
>>>>> AIRFLOW-756 |Refactor ssh_hook and ssh_operator                |-
>> |-
>>>>> 
>>>>> AIRFLOW-751 |SFTP file transfer functionality                  |#1999
>>>>> |fe0ede
>>>>> AIRFLOW-725 |Make merge tool use OS' keyring for password stora|#1966
>>>>> |8c1695
>>>>> AIRFLOW-706 |Configuration shell commands are not split properl|#2053
>>>>> |0bb6f2
>>>>> AIRFLOW-705 |airflow.configuration.run_command output does not |-
>> |-
>>>>> 
>>>>> AIRFLOW-681 |homepage doc link should pointing to apache's repo|#2164
>>>>> |a8027a
>>>>> AIRFLOW-654 |SSL for AMQP w/ Celery(Executor)                  |#2333
>>>>> |868bfe
>>>>> AIRFLOW-645 |HttpHook ignores https                            |#2311
>>>>> |fd381a
>>>>> AIRFLOW-365 |Code view in subdag trigger exception             |#2043
>>>>> |cf102c
>>>>> AIRFLOW-300 |Add Google Pubsub hook and operator               |#2036
>>>>> |d231dc
>>>>> AIRFLOW-289 |Use datetime.utcnow() to keep airflow system indep|#2618
>>>>> |20c83e
>>>>> AIRFLOW-71  |docker_operator - pulling from private repositorie|#na
>>>>> |d4406c
>>>>> 
>>>>> Cheers,
>>>>> Chris
>>>>> 
>>> 
>> 
>> 

Reply via email to