Still testing this out.

Put up a small fix for Ash's second exception
https://github.com/apache/incubator-airflow/pull/2766

On Wed, Nov 8, 2017 at 10:48 AM, Bolke de Bruin <bdbr...@gmail.com> wrote:

> Hi Chris,
>
> Actively testing here: we found an issue in the SSHOperator introduced in
> 1.9.0 (fix already merged for RC2, but blocking I as it stops us from
> running SSH properly), some minor fixes by Airbnb should also be in RC2.
> There is some logging “weirdness”, that might warrant a small patch here in
> there and could be squeezed into RC2, but I don’t consider them blocking.
>
> So almost there, but we need an RC2 imho.
>
> -1, binding.
>
> Bolke
>
> > On 8 Nov 2017, at 19:00, Ash Berlin-Taylor <ash_airflowlist@firemirror.
> com> wrote:
> >
> > -1 (for now. Non binding. Is that how this process works?)
> >
> > We've built a test env for this RC and are testing, but have run into an
> issue reading task logs. (See below)
> >
> > We haven't gotten very far with this yet, we will dig more tomorrow
> (it's the end of the UK work day now). I suspect this might be how we've
> misconfigured our logging. We will see tomorrow.
> >
> > -ash
> >
> >
> >
> >
> > File "/usr/local/lib/python3.5/dist-packages/airflow/www/views.py",
> line 712, in log
> >    logs = handler.read(ti)
> > AttributeError: 'NoneType' object has no attribute 'read'
> >
> > During handling of the above exception, another exception occurred:
> >
> > Traceback (most recent call last):
> >  File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1988,
> in wsgi_app
> >    response = self.full_dispatch_request()
> >  File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1641,
> in full_dispatch_request
> >    rv = self.handle_user_exception(e)
> >  File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1544,
> in handle_user_exception
> >    reraise(exc_type, exc_value, tb)
> >  File "/usr/local/lib/python3.5/dist-packages/flask/_compat.py", line
> 33, in reraise
> >    raise value
> >  File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1639,
> in full_dispatch_request
> >    rv = self.dispatch_request()
> >  File "/usr/local/lib/python3.5/dist-packages/flask/app.py", line 1625,
> in dispatch_request
> >    return self.view_functions[rule.endpoint](**req.view_args)
> >  File "/usr/local/lib/python3.5/dist-packages/flask_admin/base.py",
> line 69, in inner
> >    return self._run_view(f, *args, **kwargs)
> >  File "/usr/local/lib/python3.5/dist-packages/flask_admin/base.py",
> line 368, in _run_view
> >    return fn(self, *args, **kwargs)
> >  File "/usr/local/lib/python3.5/dist-packages/flask_login.py", line
> 758, in decorated_view
> >    return func(*args, **kwargs)
> >  File "/usr/local/lib/python3.5/dist-packages/airflow/www/utils.py",
> line 262, in wrapper
> >    return f(*args, **kwargs)
> >  File "/usr/local/lib/python3.5/dist-packages/airflow/www/views.py",
> line 715, in log
> >    .format(task_log_reader, e.message)]
> > AttributeError: 'AttributeError' object has no attribute 'message'
> >
> >
> >> On 8 Nov 2017, at 17:46, Chris Riccomini <criccom...@apache.org> wrote:
> >>
> >> Anyone? :/
> >>
> >> On Mon, Nov 6, 2017 at 1:22 PM, Chris Riccomini <criccom...@apache.org>
> >> wrote:
> >>
> >>> Hey all,
> >>>
> >>> I have cut Airflow 1.9.0 RC1. This email is calling a vote on the
> release,
> >>> which will last fo 72 hours. Consider this my (binding) +1.
> >>>
> >>> Airflow 1.9.0 RC1 is available at:
> >>>
> >>> https://dist.apache.org/repos/dist/dev/incubator/airflow/1.9.0rc1/
> >>>
> >>> apache-airflow-1.9.0rc1+incubating-source.tar.gz is a source release
> that
> >>> comes with INSTALL instructions.
> >>> apache-airflow-1.9.0rc1+incubating-bin.tar.gz is the binary Python
> >>> "sdist" release.
> >>>
> >>> Public keys are available at:
> >>>
> >>> https://dist.apache.org/repos/dist/release/incubator/airflow/
> >>>
> >>> The release contains the following JIRAs:
> >>>
> >>> ISSUE ID    |DESCRIPTION                                       |PR
> >>> |COMMIT
> >>> AIRFLOW-1779|Add keepalive packets to ssh hook                 |#2749
> >>> |d2f9d1
> >>> AIRFLOW-1776|stdout/stderr logging not captured                |#2745
> >>> |590d9f
> >>> AIRFLOW-1771|Change heartbeat text from boom to heartbeat      |-
>  |-
> >>>
> >>> AIRFLOW-1767|Airflow Scheduler no longer schedules DAGs        |-
>  |-
> >>>
> >>> AIRFLOW-1765|Default API auth backed should deny all.          |#2737
> >>> |6ecdac
> >>> AIRFLOW-1764|Web Interface should not use experimental api     |#2738
> >>> |6bed1d
> >>> AIRFLOW-1757|Contrib.SparkSubmitOperator should allow --package|#2725
> >>> |4e06ee
> >>> AIRFLOW-1745|BashOperator ignores SIGPIPE in subprocess        |#2714
> >>> |e021c9
> >>> AIRFLOW-1744|task.retries can be False                         |#2713
> >>> |6144c6
> >>> AIRFLOW-1743|Default config template should not contain ldap fi|#2712
> >>> |270684
> >>> AIRFLOW-1741|Task Duration shows two charts on first page load.|#2711
> >>> |974b49
> >>> AIRFLOW-1734|Sqoop Operator contains logic errors & needs optio|#2703
> >>> |f6810c
> >>> AIRFLOW-1731|Import custom config on PYTHONPATH                |#2721
> >>> |f07eb3
> >>> AIRFLOW-1726|Copy Expert command for Postgres Hook             |#2698
> >>> |8a4ad3
> >>> AIRFLOW-1719|Fix small typo - your vs you                      |-
>  |-
> >>>
> >>> AIRFLOW-1712|Log SSHOperator output                            |-
>  |-
> >>>
> >>> AIRFLOW-1711|Ldap Attributes not always a "list" part 2        |#2731
> >>> |40a936
> >>> AIRFLOW-1706|Scheduler is failed on startup with MS SQL Server |#2733
> >>> |9e209b
> >>> AIRFLOW-1698|Remove confusing SCHEDULER_RUNS env var from syste|#2677
> >>> |00dd06
> >>> AIRFLOW-1695|Redshift Hook using boto3 & AWS Hook              |#2717
> >>> |bfddae
> >>> AIRFLOW-1694|Hive Hooks: Python 3 does not have an `itertools.i|#2674
> >>> |c6e5ae
> >>> AIRFLOW-1692|Master cannot be checked out on windows           |#2673
> >>> |31805e
> >>> AIRFLOW-1691|Add better documentation for Google cloud storage |#2671
> >>> |ace2b1
> >>> AIRFLOW-1690|Error messages regarding gcs log commits are spars|#2670
> >>> |5fb5cd
> >>> AIRFLOW-1682|S3 task handler never writes to S3                |#2664
> >>> |0080f0
> >>> AIRFLOW-1678|Fix docstring errors for `set_upstream` and `set_d|-
>  |-
> >>>
> >>> AIRFLOW-1677|Fix typo in example_qubole_operator               |-
>  |-
> >>>
> >>> AIRFLOW-1676|GCS task handler never writes to GCS              |#2659
> >>> |781fa4
> >>> AIRFLOW-1675|Fix API docstrings to be properly rendered        |#2667
> >>> |f12381
> >>> AIRFLOW-1671|Missing @apply_defaults annotation for gcs downloa|#2655
> >>> |97666b
> >>> AIRFLOW-1669|Fix Docker import in Master                       |#na
> >>> |f7f2a8
> >>> AIRFLOW-1668|Redhsift requires a keep alive of < 300s          |#2650
> >>> |f2bb77
> >>> AIRFLOW-1664|Make MySqlToGoogleCloudStorageOperator support bin|#2649
> >>> |95813d
> >>> AIRFLOW-1660|Change webpage width to full-width                |#2646
> >>> |8ee3d9
> >>> AIRFLOW-1659|Fix invalid attribute bug in FileTaskHandler      |#2645
> >>> |bee823
> >>> AIRFLOW-1658|Kill (possibly) still running Druid indexing job a|#2644
> >>> |cbf7ad
> >>> AIRFLOW-1657|Handle failure of Qubole Operator for s3distcp had|-
>  |-
> >>>
> >>> AIRFLOW-1654|Show tooltips for link icons in DAGs view         |#2642
> >>> |ada7b2
> >>> AIRFLOW-1647|Fix Spark-sql hook                                |#2637
> >>> |b1e5c6
> >>> AIRFLOW-1641|Task gets stuck in queued state                   |#2715
> >>> |735497
> >>> AIRFLOW-1640|Add Qubole default connection in connection table |-
>  |-
> >>>
> >>> AIRFLOW-1639|ValueError does not have .message attribute       |#2629
> >>> |87df67
> >>> AIRFLOW-1637|readme not tracking master branch for travis      |-
>  |-
> >>>
> >>> AIRFLOW-1636|aws and emr connection types get cleared          |#2626
> >>> |540e04
> >>> AIRFLOW-1635|Allow creating Google Cloud Platform connection wi|#2640
> >>> |6dec7a
> >>> AIRFLOW-1629|make extra a textarea in edit connections form    |#2623
> >>> |f5d46f
> >>> AIRFLOW-1628|Docstring of sqlsensor is incorrect               |#2621
> >>> |9ba73d
> >>> AIRFLOW-1627|SubDagOperator initialization should only query po|#2620
> >>> |516ace
> >>> AIRFLOW-1621|Add tests for logic added on server side dag list |#2614
> >>> |8de9fd
> >>> AIRFLOW-1614|Improve performance of DAG parsing when there are |#2610
> >>> |a95adb
> >>> AIRFLOW-1611|Customize logging in Airflow                      |#2631
> >>> |8b4a50
> >>> AIRFLOW-1609|Ignore all venvs in gitignore                     |#2608
> >>> |f1f9b4
> >>> AIRFLOW-1608|GCP Dataflow hook missing pending job state       |#2607
> >>> |653562
> >>> AIRFLOW-1606|DAG.sync_to_db is static, but takes a DAG as first|#2606
> >>> |6ac296
> >>> AIRFLOW-1605|Fix log source of local loggers                   |-
>  |-
> >>>
> >>> AIRFLOW-1604|Rename the logger to log                          |#2604
> >>> |af4050
> >>> AIRFLOW-1602|Use LoggingMixin for the DAG class                |#2602
> >>> |956699
> >>> AIRFLOW-1601|Add configurable time between SIGTERM and SIGKILL |#2601
> >>> |48a95e
> >>> AIRFLOW-1600|Uncaught exceptions in get_fernet if cryptography |#2600
> >>> |ad963e
> >>> AIRFLOW-1597|Add GameWisp as Airflow user                      |#2599
> >>> |26b747
> >>> AIRFLOW-1594|Installing via pip copies test files into python l|#2597
> >>> |a6b23a
> >>> AIRFLOW-1593|Expose load_string in WasbHook                    |#2596
> >>> |7ece95
> >>> AIRFLOW-1591|Exception: 'TaskInstance' object has no attribute |#2578
> >>> |f4653e
> >>> AIRFLOW-1590|Small fix for dates util                          |#2652
> >>> |31946e
> >>> AIRFLOW-1587|fix `ImportError: cannot import name 'CeleryExecut|#2590
> >>> |34c73b
> >>> AIRFLOW-1586|MySQL to GCS to BigQuery fails for tables with dat|#2589
> >>> |e83012
> >>> AIRFLOW-1584|Remove the insecure /headers endpoints            |#2588
> >>> |17ac07
> >>> AIRFLOW-1582|Improve logging structure of Airflow              |#2592
> >>> |a7a518
> >>> AIRFLOW-1580|Error in string formatter when throwing an excepti|#2583
> >>> |ea9ab9
> >>> AIRFLOW-1579|Allow jagged rows in BQ Hook.                     |#2582
> >>> |5b978b
> >>> AIRFLOW-1577|Add token support to DatabricksHook               |#2579
> >>> |c2c515
> >>> AIRFLOW-1573|Remove `thrift < 0.10.0` requirement              |#2574
> >>> |aa95f2
> >>> AIRFLOW-1571|Add AWS Lambda Hook for invoking Lambda Function  |#2718
> >>> |017f18
> >>> AIRFLOW-1568|Add datastore import/export operator              |#2568
> >>> |86063b
> >>> AIRFLOW-1567|Clean up ML Engine operators                      |#2567
> >>> |af91e2
> >>> AIRFLOW-1564|Default logging filename contains a colon         |#2565
> >>> |4c674c
> >>> AIRFLOW-1560|Add AWS DynamoDB hook for inserting batch items   |#2587
> >>> |71400b
> >>> AIRFLOW-1556|BigQueryBaseCursor should support SQL parameters  |#2557
> >>> |9df0ac
> >>> AIRFLOW-1546| add Zymergen to org list in README               |#2512
> >>> |7cc346
> >>> AIRFLOW-1535|Add support for Dataproc serviceAccountScopes in D|#2546
> >>> |b1f902
> >>> AIRFLOW-1529|Support quoted newlines in Google BigQuery load jo|#2545
> >>> |4a4b02
> >>> AIRFLOW-1527|Refactor celery config to make use of template    |#2542
> >>> |f4437b
> >>> AIRFLOW-1522|Increase size of val column for variable table in |#2535
> >>> |8a2d24
> >>> AIRFLOW-1521|Template fields definition for bigquery_table_dele|#2534
> >>> |f1a7c0
> >>> AIRFLOW-1520|S3Hook uses boto2                                 |#2532
> >>> |386583
> >>> AIRFLOW-1519|Main DAG list page does not scale using client sid|#2531
> >>> |d7d7ce
> >>> AIRFLOW-1512|Add operator for running Python functions in a vir|#2446
> >>> |14e6d7
> >>> AIRFLOW-1507|Make src, dst and bucket parameters as templated i|#2516
> >>> |d295cf
> >>> AIRFLOW-1505|Document when Jinja substitution occurs           |#2523
> >>> |984a87
> >>> AIRFLOW-1504|Log Cluster Name on Dataproc Operator When Execute|#2517
> >>> |1cd6c4
> >>> AIRFLOW-1499s|Eliminate duplicate and unneeded code             |-
>  |-
> >>>
> >>> AIRFLOW-1497|Hidden fields in connection form aren't reset when|#2507
> >>> |d8da8b
> >>> AIRFLOW-1493|Fix race condition with airflow run               |#2505
> >>> |b2e175
> >>> AIRFLOW-1492|Add metric for task success/failure               |#2504
> >>> |fa84d4
> >>> AIRFLOW-1489|Docs: Typo in BigQueryCheckOperator               |#2501
> >>> |111ce5
> >>> AIRFLOW-1483|Page size on model views is to large to render qui|#2497
> >>> |04bfba
> >>> AIRFLOW-1478|Chart -> Owner column should be sortable          |#2493
> >>> |651e60
> >>> AIRFLOW-1476|Add INSTALL file for source releases              |#2492
> >>> |da76ac
> >>> AIRFLOW-1474|Add dag_id regex for 'airflow clear' CLI command  |#2486
> >>> |18f849
> >>> AIRFLOW-1470s|BashSensor Implementation                         |-
>  |-
> >>>
> >>> AIRFLOW-1459|integration rst doc is broken in github view      |#2481
> >>> |322ec9
> >>> AIRFLOW-1438|Scheduler batch queries should have a limit       |#2462
> >>> |3547cb
> >>> AIRFLOW-1437|BigQueryTableDeleteOperator should define deletion|#2459
> >>> |b87903
> >>> AIRFLOW-1432|NVD3 Charts do not have labeled axes and units cha|#2710
> >>> |70ffa4
> >>> AIRFLOW-1402|Cleanup SafeConfigParser DeprecationWarning       |#2435
> >>> |38c86b
> >>> AIRFLOW-1401|Standardize GCP project, region, and zone argument|#2439
> >>> |b6d363
> >>> AIRFLOW-1397|Airflow 1.8.1 - No data displays in Last Run Colum|-
>  |-
> >>>
> >>> AIRFLOW-1394|Add quote_character parameter to GoogleCloudStorag|#2428
> >>> |9fd0be
> >>> AIRFLOW-1389|BigQueryOperator should support `createDisposition|#2470
> >>> |6e2640
> >>> AIRFLOW-1384|Add ARGO/CaDC                                     |#2434
> >>> |715947
> >>> AIRFLOW-1368|Automatically remove the container when it exits  |#2653
> >>> |d42d23
> >>> AIRFLOW-1359|Provide GoogleCloudML operator for model evaluatio|#2407
> >>> |194d1d
> >>> AIRFLOW-1356|add `--celery_hostname` to `airflow worker`       |#2405
> >>> |b9d7d1
> >>> AIRFLOW-1352|Revert bad logging Handler                        |-
>  |-
> >>>
> >>> AIRFLOW-1350|Add "query_uri" parameter for Google DataProc oper|#2402
> >>> |d32c72
> >>> AIRFLOW-1348|Paginated UI has broken toggles after first page  |-
>  |-
> >>>
> >>> AIRFLOW-1345|Don't commit on each loop                         |#2397
> >>> |0dd002
> >>> AIRFLOW-1344|Builds failing on Python 3.5 with AttributeError  |#2394
> >>> |2a5883
> >>> AIRFLOW-1343|Add airflow default label to the dataproc operator|#2396
> >>> |e4b240
> >>> AIRFLOW-1338|gcp_dataflow_hook is incompatible with the recent |#2388
> >>> |cf2605
> >>> AIRFLOW-1337|Customize log format via config file              |#2392
> >>> |4841e3
> >>> AIRFLOW-1335|Use buffered logger                               |#2386
> >>> |0d23d3
> >>> AIRFLOW-1333|Enable copy function for Google Cloud Storage Hook|#2385
> >>> |e2c383
> >>> AIRFLOW-1331|Contrib.SparkSubmitOperator should allow --package|#2622
> >>> |fbca8f
> >>> AIRFLOW-1330|Connection.parse_from_uri doesn't work for google_|#2525
> >>> |6e5e9d
> >>> AIRFLOW-1324|Make the Druid operator/hook more general         |#2378
> >>> |de99aa
> >>> AIRFLOW-1323|Operators related to Dataproc should keep some par|#2636
> >>> |ed248d
> >>> AIRFLOW-1315|Add Qubole File and Partition Sensors             |-
>  |-
> >>>
> >>> AIRFLOW-1309|Add optional hive_tblproperties in HiveToDruidTran|-
>  |-
> >>>
> >>> AIRFLOW-1301|Add New Relic to Airflow user list                |#2359
> >>> |355fc9
> >>> AIRFLOW-1299|Google Dataproc cluster creation operator should s|#2358
> >>> |c2b80e
> >>> AIRFLOW-1289|Don't restrict scheduler threads to CPU cores     |#2353
> >>> |8e23d2
> >>> AIRFLOW-1286|BaseTaskRunner - Exception TypeError: a bytes-like|#2363
> >>> |d8891d
> >>> AIRFLOW-1277|Forbid creation of a known event with empty fields|#na
> >>> |65184a
> >>> AIRFLOW-1276|Forbid event creation with end_data earlier than s|#na
> >>> |d5d02f
> >>> AIRFLOW-1275|Fix `airflow pool` command exception              |#2346
> >>> |9958aa
> >>> AIRFLOW-1273|Google Cloud ML Version and Model CRUD Operator   |#2379
> >>> |534a0e
> >>> AIRFLOW-1272|Google Cloud ML Batch Prediction Operator         |#2390
> >>> |e92d6b
> >>> AIRFLOW-1271|Google Cloud ML Training Operator                 |#2408
> >>> |0fc450
> >>> AIRFLOW-1256|Add United Airlines as Airflow user               |#2332
> >>> |d3484a
> >>> AIRFLOW-1251|Add eRevalue as an Airflow user                   |#2331
> >>> |8d5160
> >>> AIRFLOW-1248|Fix inconsistent configuration name for worker tim|#2328
> >>> |92314f
> >>> AIRFLOW-1247|CLI: ignore all dependencies argument ignored     |#2441
> >>> |e88ecf
> >>> AIRFLOW-1245|Fix random failure of test_trigger_dag_for_date un|#2325
> >>> |cef01b
> >>> AIRFLOW-1244|Forbid creation of a pool with empty name         |#2324
> >>> |df9a10
> >>> AIRFLOW-1242|BigQueryHook assumes that a valid project_id can't|#2335
> >>> |ffe616
> >>> AIRFLOW-1237|Fix IN-predicate sqlalchemy warning               |#2320
> >>> |a1f422
> >>> AIRFLOW-1234|Cover utils.operator_helpers with unit tests      |#2317
> >>> |d16537
> >>> AIRFLOW-1233|Cover utils.json with unit tests                  |#2316
> >>> |502410
> >>> AIRFLOW-1232|Remove deprecated readfp warning                  |#2315
> >>> |6ffaaf
> >>> AIRFLOW-1231|Use flask_wtf.CSRFProtect instead of flask_wtf.Csr|#2313
> >>> |cac49e
> >>> AIRFLOW-1221|Fix DatabricksSubmitRunOperator Templating        |#2308
> >>> |0fa104
> >>> AIRFLOW-1217|Enable logging in Sqoop hook                      |#2307
> >>> |4f459b
> >>> AIRFLOW-1213|Add hcatalog parameters to the sqoop operator/hook|#2305
> >>> |857850
> >>> AIRFLOW-1208|Speed-up cli tests                                |#2301
> >>> |21c142
> >>> AIRFLOW-1207|Enable utils.helpers unit tests                   |#2300
> >>> |8ac87b
> >>> AIRFLOW-1203|Tests failing after oauth upgrade                 |#2296
> >>> |3e9c66
> >>> AIRFLOW-1201|Update deprecated 'nose-parameterized' library to |#2298
> >>> |d2d3e4
> >>> AIRFLOW-1193|Add Checkr to Airflow user list                   |#2276
> >>> |707238
> >>> AIRFLOW-1189|Get pandas DataFrame using BigQueryHook fails     |#2287
> >>> |93666f
> >>> AIRFLOW-1188|Add max_bad_records param to GoogleCloudStorageToB|#2286
> >>> |443e6b
> >>> AIRFLOW-1187|Obsolete package names in documentation           |-
>  |-
> >>>
> >>> AIRFLOW-1185|Incorrect url to PyPi                             |#2283
> >>> |829755
> >>> AIRFLOW-1181|Enable delete and list function for Google Cloud S|#2281
> >>> |24f73c
> >>> AIRFLOW-1179|Pandas 0.20 broke Google BigQuery hook            |#2279
> >>> |ac9ccb
> >>> AIRFLOW-1177|variable json deserialize does not work at set def|#2540
> >>> |65319a
> >>> AIRFLOW-1175|Add Pronto Tools to Airflow user list             |#2277
> >>> |86aafa
> >>> AIRFLOW-1173|Add Robinhood to list of Airflow users            |#2271
> >>> |379115
> >>> AIRFLOW-1165|airflow webservice crashes on ubuntu16 - python3  |-
>  |-
> >>>
> >>> AIRFLOW-1160|Upadte SparkSubmitOperator parameters             |#2265
> >>> |2e3f07
> >>> AIRFLOW-1155|Add Tails.com to community                        |#2261
> >>> |2fa690
> >>> AIRFLOW-1149|Allow custom filters to be added to jinja2        |#2258
> >>> |48135a
> >>> AIRFLOW-1141|Remove DAG.crawl_for_tasks method                 |#2275
> >>> |a30fee
> >>> AIRFLOW-1140|DatabricksSubmitRunOperator should template the "j|#2255
> >>> |e6d316
> >>> AIRFLOW-1136|Invalid parameters are not captured for Sqoop oper|#2252
> >>> |2ef4db
> >>> AIRFLOW-1125|Clarify documentation regarding fernet_key        |#2251
> >>> |831f8d
> >>> AIRFLOW-1122|Node strokes are too thin for people with color vi|#2246
> >>> |a08761
> >>> AIRFLOW-1121|airflow webserver --pid no longer write out pid fi|-
>  |-
> >>>
> >>> AIRFLOW-1118|Add evo.company to Airflow users                  |#2243
> >>> |f16914
> >>> AIRFLOW-1112|Log which pool is full in scheduler when pool slot|#2242
> >>> |74c1ce
> >>> AIRFLOW-1107|Add support for ftps non-default port             |#2240
> >>> |4d0c2f
> >>> AIRFLOW-1106|Add Groupalia/Letsbonus                           |#2239
> >>> |945b42
> >>> AIRFLOW-1095|ldap_auth memberOf should come from configuration |#2232
> >>> |6b1c32
> >>> AIRFLOW-1094|Invalid unit tests under `contrib/`               |#2234
> >>> |219c50
> >>> AIRFLOW-1091|As a release manager I want to be able to compare |#2231
> >>> |bfae42
> >>> AIRFLOW-1090|Add HBO                                           |#2230
> >>> |177d34
> >>> AIRFLOW-1089|Add Spark application arguments to SparkSubmitOper|#2229
> >>> |e5b914
> >>> AIRFLOW-1081|Task duration page is slow                        |#2226
> >>> |0da512
> >>> AIRFLOW-1075|Cleanup security docs                             |#2222
> >>> |5a6f18
> >>> AIRFLOW-1065|Add functionality for Azure Blob Storage          |#2216
> >>> |f1bc5f
> >>> AIRFLOW-1059|Reset_state_for_orphaned_task should operate in ba|#2205
> >>> |e05d3b
> >>> AIRFLOW-1058|Improvements for SparkSubmitOperator              |-
>  |-
> >>>
> >>> AIRFLOW-1051|Add a test for resetdb to CliTests                |#2198
> >>> |15aee0
> >>> AIRFLOW-1047|Airflow logs vulnerable to XSS                    |#2193
> >>> |fe9ebe
> >>> AIRFLOW-1045|Make log level configurable via airflow.cfg       |#2191
> >>> |e739a5
> >>> AIRFLOW-1043|Documentation issues for operators                |#2188
> >>> |b55f41
> >>> AIRFLOW-1041|DockerOperator replaces its xcom_push method with |#2274
> >>> |03704c
> >>> AIRFLOW-1040|Fix typos in comments/docstrings in models.py     |#2174
> >>> |d8c0f5
> >>> AIRFLOW-1036|Exponential backoff should use randomization      |#2262
> >>> |66168e
> >>> AIRFLOW-1035|Exponential backoff retry logic should use 2 as ba|#2196
> >>> |4ec932
> >>> AIRFLOW-1034|Make it possible to connect to S3 in sigv4 regions|#2181
> >>> |4c0905
> >>> AIRFLOW-1031|'scheduled__' may replace with DagRun.ID_PREFIX in|#2613
> >>> |aa3844
> >>> AIRFLOW-1030|HttpHook error when creating HttpSensor           |-
>  |-
> >>>
> >>> AIRFLOW-1028|Databricks Operator for Airflow                   |#2202
> >>> |53ca50
> >>> AIRFLOW-1024|Handle CeleryExecutor errors gracefully           |#2355
> >>> |7af20f
> >>> AIRFLOW-1018|Scheduler DAG processes can not log to stdout     |#2728
> >>> |ef775d
> >>> AIRFLOW-1016|Allow HTTP HEAD request method on HTTPSensor      |#2175
> >>> |4c41f6
> >>> AIRFLOW-1010|Add a convenience script for signing              |#2169
> >>> |a2b65a
> >>> AIRFLOW-1009|Remove SQLOperator from Concepts page             |#2168
> >>> |7d1144
> >>> AIRFLOW-1007|Jinja sandbox is vulnerable to RCE                |#2184
> >>> |daa281
> >>> AIRFLOW-1005|Speed up Airflow startup time                     |#na
> >>> |996dd3
> >>> AIRFLOW-999 |Support for Redis database                        |#2165
> >>> |8de850
> >>> AIRFLOW-997 |Change setup.cfg to point to Apache instead of Max|#na
> >>> |75cd46
> >>> AIRFLOW-995 |Update Github PR template                         |#2163
> >>> |b62485
> >>> AIRFLOW-994 |Add MiNODES to the AIRFLOW Active Users List      |#2159
> >>> |ca1623
> >>> AIRFLOW-991 |Mark_success while a task is running leads to fail|-
>  |-
> >>>
> >>> AIRFLOW-990 |DockerOperator fails when logging unicode string  |#2155
> >>> |6bbf54
> >>> AIRFLOW-988 |SLA Miss Callbacks Are Repeated if Email is Not be|#2415
> >>> |6e74d4
> >>> AIRFLOW-985 |Extend the sqoop operator/hook with additional par|#2177
> >>> |82eb20
> >>> AIRFLOW-984 |Subdags unrecognized when subclassing SubDagOperat|#2152
> >>> |a8bd16
> >>> AIRFLOW-979 |Add GovTech GDS                                   |#2149
> >>> |b17bd3
> >>> AIRFLOW-976 |Mark success running task causes it to fail       |-
>  |-
> >>>
> >>> AIRFLOW-969 |Catch bad python_callable argument at DAG construc|#2142
> >>> |12901d
> >>> AIRFLOW-963 |Some code examples are not rendered in the airflow|#2139
> >>> |f69c1b
> >>> AIRFLOW-960 |Add support for .editorconfig                     |#na
> >>> |f5cacc
> >>> AIRFLOW-959 |.gitignore file is disorganized and incomplete    |#na
> >>> |3d3c14
> >>> AIRFLOW-958 |Improve tooltip readability                       |#2134
> >>> |b3c3eb
> >>> AIRFLOW-950 |Missing AWS integrations on documentation::integra|#2552
> >>> |01be02
> >>> AIRFLOW-947 |Make PrestoHook surface better messages when the P|#na
> >>> |6dd4b3
> >>> AIRFLOW-945 |Revert psycopg2 workaround when psycopg2 2.7.1 is |-
>  |-
> >>>
> >>> AIRFLOW-943 |Add Digital First Media to the Airflow users list |#2115
> >>> |2cfe28
> >>> AIRFLOW-942 |Add mytaxi to Airflow Users                       |#2111
> >>> |d579e6
> >>> AIRFLOW-935 |Impossible to use plugin executors                |#2120
> >>> |08a784
> >>> AIRFLOW-926 |jdbc connector is broken due to jaydebeapi api upd|#2651
> >>> |07ed29
> >>> AIRFLOW-917 |Incorrectly formatted failure status message      |#2109
> >>> |b8164c
> >>> AIRFLOW-916 |Fix ConfigParser deprecation warning              |#2108
> >>> |ef6dd1
> >>> AIRFLOW-911 |Add colouring and profiling info on tests         |#2106
> >>> |4f52db
> >>> AIRFLOW-903 |Add configuration setting for default DAG view.   |#2103
> >>> |cadfae
> >>> AIRFLOW-896 |BigQueryOperator fails to execute with certain inp|#2097
> >>> |2bceee
> >>> AIRFLOW-891 |Webserver Clock Should Include Day                |-
>  |-
> >>>
> >>> AIRFLOW-889 |Minor error in the docstrings for BaseOperator.   |#2084
> >>> |50702d
> >>> AIRFLOW-887 |Add compatibility with future v0.16               |#na
> >>> |50902d
> >>> AIRFLOW-886 |Pass Operator result to post_execute hook         |#na
> >>> |4da361
> >>> AIRFLOW-885 |Add Change.org to the list of Airflow users       |#2089
> >>> |a279be
> >>> AIRFLOW-882 |Code example in docs has unnecessary DAG>>Operator|#2088
> >>> |baa4cd
> >>> AIRFLOW-881 |Create SubDagOperator within DAG context manager w|#2087
> >>> |0ed608
> >>> AIRFLOW-880 |Fix remote log functionality inconsistencies for W|#2086
> >>> |974b75
> >>> AIRFLOW-877 |GoogleCloudStorageDownloadOperator: template_ext c|#2083
> >>> |debc69
> >>> AIRFLOW-875 |Allow HttpSensor params to be templated           |#2080
> >>> |62f503
> >>> AIRFLOW-871 |multiple places use logging.warn() instead of warn|#2082
> >>> |21d775
> >>> AIRFLOW-866 |Add FTPSensor                                     |#2070
> >>> |5f87f8
> >>> AIRFLOW-863 |Example DAG start dates should be recent to avoid |#2068
> >>> |bbfd43
> >>> AIRFLOW-862 |Add DaskExecutor                                  |#2067
> >>> |6e2210
> >>> AIRFLOW-860 |Circular module dependency prevents loading of cus|-
>  |-
> >>>
> >>> AIRFLOW-854 |Add Open Knowledge International to Airflow users |#2061
> >>> |51a311
> >>> AIRFLOW-842 |scheduler.clean_dirty raises warning: SAWarning: T|#2072
> >>> |485280
> >>> AIRFLOW-840 |Python3 encoding issue in Kerberos                |#2158
> >>> |639336
> >>> AIRFLOW-836 |The paused and queryview endpoints are vulnerable |#2054
> >>> |6aca2c
> >>> AIRFLOW-831 |Fix broken unit tests                             |#2050
> >>> |b86194
> >>> AIRFLOW-830 |Plugin manager should log to debug, not info      |-
>  |-
> >>>
> >>> AIRFLOW-829 |Reduce verbosity of successful Travis unit tests  |-
>  |-
> >>>
> >>> AIRFLOW-826 |Add Zendesk Hook                                  |#2066
> >>> |a09762
> >>> AIRFLOW-823 |Make task instance details available via API      |#2045
> >>> |3f546e
> >>> AIRFLOW-822 |Close the connection before throwing exception in |#2038
> >>> |4b6c38
> >>> AIRFLOW-821 |Scheduler dagbag importing not Py3 compatible     |#2039
> >>> |fbb59b
> >>> AIRFLOW-809 |SqlAlchemy is_ ColumnOperator Causing Errors in MS|-
>  |-
> >>>
> >>> AIRFLOW-802 |Integration of spark-submit                       |-
>  |-
> >>>
> >>> AIRFLOW-781 |Allow DataFlowJavaOperator to accept jar file stor|#2037
> >>> |259c86
> >>> AIRFLOW-770 |HDFS hooks should support alternative ways of gett|#2056
> >>> |261b65
> >>> AIRFLOW-756 |Refactor ssh_hook and ssh_operator                |-
>  |-
> >>>
> >>> AIRFLOW-751 |SFTP file transfer functionality                  |#1999
> >>> |fe0ede
> >>> AIRFLOW-725 |Make merge tool use OS' keyring for password stora|#1966
> >>> |8c1695
> >>> AIRFLOW-706 |Configuration shell commands are not split properl|#2053
> >>> |0bb6f2
> >>> AIRFLOW-705 |airflow.configuration.run_command output does not |-
>  |-
> >>>
> >>> AIRFLOW-681 |homepage doc link should pointing to apache's repo|#2164
> >>> |a8027a
> >>> AIRFLOW-654 |SSL for AMQP w/ Celery(Executor)                  |#2333
> >>> |868bfe
> >>> AIRFLOW-645 |HttpHook ignores https                            |#2311
> >>> |fd381a
> >>> AIRFLOW-365 |Code view in subdag trigger exception             |#2043
> >>> |cf102c
> >>> AIRFLOW-300 |Add Google Pubsub hook and operator               |#2036
> >>> |d231dc
> >>> AIRFLOW-289 |Use datetime.utcnow() to keep airflow system indep|#2618
> >>> |20c83e
> >>> AIRFLOW-71  |docker_operator - pulling from private repositorie|#na
> >>> |d4406c
> >>>
> >>> Cheers,
> >>> Chris
> >>>
> >
>
>

Reply via email to