Anyone? :/ On Mon, Nov 6, 2017 at 1:22 PM, Chris Riccomini <[email protected]> wrote:
> Hey all, > > I have cut Airflow 1.9.0 RC1. This email is calling a vote on the release, > which will last fo 72 hours. Consider this my (binding) +1. > > Airflow 1.9.0 RC1 is available at: > > https://dist.apache.org/repos/dist/dev/incubator/airflow/1.9.0rc1/ > > apache-airflow-1.9.0rc1+incubating-source.tar.gz is a source release that > comes with INSTALL instructions. > apache-airflow-1.9.0rc1+incubating-bin.tar.gz is the binary Python > "sdist" release. > > Public keys are available at: > > https://dist.apache.org/repos/dist/release/incubator/airflow/ > > The release contains the following JIRAs: > > ISSUE ID |DESCRIPTION |PR > |COMMIT > AIRFLOW-1779|Add keepalive packets to ssh hook |#2749 > |d2f9d1 > AIRFLOW-1776|stdout/stderr logging not captured |#2745 > |590d9f > AIRFLOW-1771|Change heartbeat text from boom to heartbeat |- |- > > AIRFLOW-1767|Airflow Scheduler no longer schedules DAGs |- |- > > AIRFLOW-1765|Default API auth backed should deny all. |#2737 > |6ecdac > AIRFLOW-1764|Web Interface should not use experimental api |#2738 > |6bed1d > AIRFLOW-1757|Contrib.SparkSubmitOperator should allow --package|#2725 > |4e06ee > AIRFLOW-1745|BashOperator ignores SIGPIPE in subprocess |#2714 > |e021c9 > AIRFLOW-1744|task.retries can be False |#2713 > |6144c6 > AIRFLOW-1743|Default config template should not contain ldap fi|#2712 > |270684 > AIRFLOW-1741|Task Duration shows two charts on first page load.|#2711 > |974b49 > AIRFLOW-1734|Sqoop Operator contains logic errors & needs optio|#2703 > |f6810c > AIRFLOW-1731|Import custom config on PYTHONPATH |#2721 > |f07eb3 > AIRFLOW-1726|Copy Expert command for Postgres Hook |#2698 > |8a4ad3 > AIRFLOW-1719|Fix small typo - your vs you |- |- > > AIRFLOW-1712|Log SSHOperator output |- |- > > AIRFLOW-1711|Ldap Attributes not always a "list" part 2 |#2731 > |40a936 > AIRFLOW-1706|Scheduler is failed on startup with MS SQL Server |#2733 > |9e209b > AIRFLOW-1698|Remove confusing SCHEDULER_RUNS env var from syste|#2677 > |00dd06 > AIRFLOW-1695|Redshift Hook using boto3 & AWS Hook |#2717 > |bfddae > AIRFLOW-1694|Hive Hooks: Python 3 does not have an `itertools.i|#2674 > |c6e5ae > AIRFLOW-1692|Master cannot be checked out on windows |#2673 > |31805e > AIRFLOW-1691|Add better documentation for Google cloud storage |#2671 > |ace2b1 > AIRFLOW-1690|Error messages regarding gcs log commits are spars|#2670 > |5fb5cd > AIRFLOW-1682|S3 task handler never writes to S3 |#2664 > |0080f0 > AIRFLOW-1678|Fix docstring errors for `set_upstream` and `set_d|- |- > > AIRFLOW-1677|Fix typo in example_qubole_operator |- |- > > AIRFLOW-1676|GCS task handler never writes to GCS |#2659 > |781fa4 > AIRFLOW-1675|Fix API docstrings to be properly rendered |#2667 > |f12381 > AIRFLOW-1671|Missing @apply_defaults annotation for gcs downloa|#2655 > |97666b > AIRFLOW-1669|Fix Docker import in Master |#na > |f7f2a8 > AIRFLOW-1668|Redhsift requires a keep alive of < 300s |#2650 > |f2bb77 > AIRFLOW-1664|Make MySqlToGoogleCloudStorageOperator support bin|#2649 > |95813d > AIRFLOW-1660|Change webpage width to full-width |#2646 > |8ee3d9 > AIRFLOW-1659|Fix invalid attribute bug in FileTaskHandler |#2645 > |bee823 > AIRFLOW-1658|Kill (possibly) still running Druid indexing job a|#2644 > |cbf7ad > AIRFLOW-1657|Handle failure of Qubole Operator for s3distcp had|- |- > > AIRFLOW-1654|Show tooltips for link icons in DAGs view |#2642 > |ada7b2 > AIRFLOW-1647|Fix Spark-sql hook |#2637 > |b1e5c6 > AIRFLOW-1641|Task gets stuck in queued state |#2715 > |735497 > AIRFLOW-1640|Add Qubole default connection in connection table |- |- > > AIRFLOW-1639|ValueError does not have .message attribute |#2629 > |87df67 > AIRFLOW-1637|readme not tracking master branch for travis |- |- > > AIRFLOW-1636|aws and emr connection types get cleared |#2626 > |540e04 > AIRFLOW-1635|Allow creating Google Cloud Platform connection wi|#2640 > |6dec7a > AIRFLOW-1629|make extra a textarea in edit connections form |#2623 > |f5d46f > AIRFLOW-1628|Docstring of sqlsensor is incorrect |#2621 > |9ba73d > AIRFLOW-1627|SubDagOperator initialization should only query po|#2620 > |516ace > AIRFLOW-1621|Add tests for logic added on server side dag list |#2614 > |8de9fd > AIRFLOW-1614|Improve performance of DAG parsing when there are |#2610 > |a95adb > AIRFLOW-1611|Customize logging in Airflow |#2631 > |8b4a50 > AIRFLOW-1609|Ignore all venvs in gitignore |#2608 > |f1f9b4 > AIRFLOW-1608|GCP Dataflow hook missing pending job state |#2607 > |653562 > AIRFLOW-1606|DAG.sync_to_db is static, but takes a DAG as first|#2606 > |6ac296 > AIRFLOW-1605|Fix log source of local loggers |- |- > > AIRFLOW-1604|Rename the logger to log |#2604 > |af4050 > AIRFLOW-1602|Use LoggingMixin for the DAG class |#2602 > |956699 > AIRFLOW-1601|Add configurable time between SIGTERM and SIGKILL |#2601 > |48a95e > AIRFLOW-1600|Uncaught exceptions in get_fernet if cryptography |#2600 > |ad963e > AIRFLOW-1597|Add GameWisp as Airflow user |#2599 > |26b747 > AIRFLOW-1594|Installing via pip copies test files into python l|#2597 > |a6b23a > AIRFLOW-1593|Expose load_string in WasbHook |#2596 > |7ece95 > AIRFLOW-1591|Exception: 'TaskInstance' object has no attribute |#2578 > |f4653e > AIRFLOW-1590|Small fix for dates util |#2652 > |31946e > AIRFLOW-1587|fix `ImportError: cannot import name 'CeleryExecut|#2590 > |34c73b > AIRFLOW-1586|MySQL to GCS to BigQuery fails for tables with dat|#2589 > |e83012 > AIRFLOW-1584|Remove the insecure /headers endpoints |#2588 > |17ac07 > AIRFLOW-1582|Improve logging structure of Airflow |#2592 > |a7a518 > AIRFLOW-1580|Error in string formatter when throwing an excepti|#2583 > |ea9ab9 > AIRFLOW-1579|Allow jagged rows in BQ Hook. |#2582 > |5b978b > AIRFLOW-1577|Add token support to DatabricksHook |#2579 > |c2c515 > AIRFLOW-1573|Remove `thrift < 0.10.0` requirement |#2574 > |aa95f2 > AIRFLOW-1571|Add AWS Lambda Hook for invoking Lambda Function |#2718 > |017f18 > AIRFLOW-1568|Add datastore import/export operator |#2568 > |86063b > AIRFLOW-1567|Clean up ML Engine operators |#2567 > |af91e2 > AIRFLOW-1564|Default logging filename contains a colon |#2565 > |4c674c > AIRFLOW-1560|Add AWS DynamoDB hook for inserting batch items |#2587 > |71400b > AIRFLOW-1556|BigQueryBaseCursor should support SQL parameters |#2557 > |9df0ac > AIRFLOW-1546| add Zymergen to org list in README |#2512 > |7cc346 > AIRFLOW-1535|Add support for Dataproc serviceAccountScopes in D|#2546 > |b1f902 > AIRFLOW-1529|Support quoted newlines in Google BigQuery load jo|#2545 > |4a4b02 > AIRFLOW-1527|Refactor celery config to make use of template |#2542 > |f4437b > AIRFLOW-1522|Increase size of val column for variable table in |#2535 > |8a2d24 > AIRFLOW-1521|Template fields definition for bigquery_table_dele|#2534 > |f1a7c0 > AIRFLOW-1520|S3Hook uses boto2 |#2532 > |386583 > AIRFLOW-1519|Main DAG list page does not scale using client sid|#2531 > |d7d7ce > AIRFLOW-1512|Add operator for running Python functions in a vir|#2446 > |14e6d7 > AIRFLOW-1507|Make src, dst and bucket parameters as templated i|#2516 > |d295cf > AIRFLOW-1505|Document when Jinja substitution occurs |#2523 > |984a87 > AIRFLOW-1504|Log Cluster Name on Dataproc Operator When Execute|#2517 > |1cd6c4 > AIRFLOW-1499s|Eliminate duplicate and unneeded code |- |- > > AIRFLOW-1497|Hidden fields in connection form aren't reset when|#2507 > |d8da8b > AIRFLOW-1493|Fix race condition with airflow run |#2505 > |b2e175 > AIRFLOW-1492|Add metric for task success/failure |#2504 > |fa84d4 > AIRFLOW-1489|Docs: Typo in BigQueryCheckOperator |#2501 > |111ce5 > AIRFLOW-1483|Page size on model views is to large to render qui|#2497 > |04bfba > AIRFLOW-1478|Chart -> Owner column should be sortable |#2493 > |651e60 > AIRFLOW-1476|Add INSTALL file for source releases |#2492 > |da76ac > AIRFLOW-1474|Add dag_id regex for 'airflow clear' CLI command |#2486 > |18f849 > AIRFLOW-1470s|BashSensor Implementation |- |- > > AIRFLOW-1459|integration rst doc is broken in github view |#2481 > |322ec9 > AIRFLOW-1438|Scheduler batch queries should have a limit |#2462 > |3547cb > AIRFLOW-1437|BigQueryTableDeleteOperator should define deletion|#2459 > |b87903 > AIRFLOW-1432|NVD3 Charts do not have labeled axes and units cha|#2710 > |70ffa4 > AIRFLOW-1402|Cleanup SafeConfigParser DeprecationWarning |#2435 > |38c86b > AIRFLOW-1401|Standardize GCP project, region, and zone argument|#2439 > |b6d363 > AIRFLOW-1397|Airflow 1.8.1 - No data displays in Last Run Colum|- |- > > AIRFLOW-1394|Add quote_character parameter to GoogleCloudStorag|#2428 > |9fd0be > AIRFLOW-1389|BigQueryOperator should support `createDisposition|#2470 > |6e2640 > AIRFLOW-1384|Add ARGO/CaDC |#2434 > |715947 > AIRFLOW-1368|Automatically remove the container when it exits |#2653 > |d42d23 > AIRFLOW-1359|Provide GoogleCloudML operator for model evaluatio|#2407 > |194d1d > AIRFLOW-1356|add `--celery_hostname` to `airflow worker` |#2405 > |b9d7d1 > AIRFLOW-1352|Revert bad logging Handler |- |- > > AIRFLOW-1350|Add "query_uri" parameter for Google DataProc oper|#2402 > |d32c72 > AIRFLOW-1348|Paginated UI has broken toggles after first page |- |- > > AIRFLOW-1345|Don't commit on each loop |#2397 > |0dd002 > AIRFLOW-1344|Builds failing on Python 3.5 with AttributeError |#2394 > |2a5883 > AIRFLOW-1343|Add airflow default label to the dataproc operator|#2396 > |e4b240 > AIRFLOW-1338|gcp_dataflow_hook is incompatible with the recent |#2388 > |cf2605 > AIRFLOW-1337|Customize log format via config file |#2392 > |4841e3 > AIRFLOW-1335|Use buffered logger |#2386 > |0d23d3 > AIRFLOW-1333|Enable copy function for Google Cloud Storage Hook|#2385 > |e2c383 > AIRFLOW-1331|Contrib.SparkSubmitOperator should allow --package|#2622 > |fbca8f > AIRFLOW-1330|Connection.parse_from_uri doesn't work for google_|#2525 > |6e5e9d > AIRFLOW-1324|Make the Druid operator/hook more general |#2378 > |de99aa > AIRFLOW-1323|Operators related to Dataproc should keep some par|#2636 > |ed248d > AIRFLOW-1315|Add Qubole File and Partition Sensors |- |- > > AIRFLOW-1309|Add optional hive_tblproperties in HiveToDruidTran|- |- > > AIRFLOW-1301|Add New Relic to Airflow user list |#2359 > |355fc9 > AIRFLOW-1299|Google Dataproc cluster creation operator should s|#2358 > |c2b80e > AIRFLOW-1289|Don't restrict scheduler threads to CPU cores |#2353 > |8e23d2 > AIRFLOW-1286|BaseTaskRunner - Exception TypeError: a bytes-like|#2363 > |d8891d > AIRFLOW-1277|Forbid creation of a known event with empty fields|#na > |65184a > AIRFLOW-1276|Forbid event creation with end_data earlier than s|#na > |d5d02f > AIRFLOW-1275|Fix `airflow pool` command exception |#2346 > |9958aa > AIRFLOW-1273|Google Cloud ML Version and Model CRUD Operator |#2379 > |534a0e > AIRFLOW-1272|Google Cloud ML Batch Prediction Operator |#2390 > |e92d6b > AIRFLOW-1271|Google Cloud ML Training Operator |#2408 > |0fc450 > AIRFLOW-1256|Add United Airlines as Airflow user |#2332 > |d3484a > AIRFLOW-1251|Add eRevalue as an Airflow user |#2331 > |8d5160 > AIRFLOW-1248|Fix inconsistent configuration name for worker tim|#2328 > |92314f > AIRFLOW-1247|CLI: ignore all dependencies argument ignored |#2441 > |e88ecf > AIRFLOW-1245|Fix random failure of test_trigger_dag_for_date un|#2325 > |cef01b > AIRFLOW-1244|Forbid creation of a pool with empty name |#2324 > |df9a10 > AIRFLOW-1242|BigQueryHook assumes that a valid project_id can't|#2335 > |ffe616 > AIRFLOW-1237|Fix IN-predicate sqlalchemy warning |#2320 > |a1f422 > AIRFLOW-1234|Cover utils.operator_helpers with unit tests |#2317 > |d16537 > AIRFLOW-1233|Cover utils.json with unit tests |#2316 > |502410 > AIRFLOW-1232|Remove deprecated readfp warning |#2315 > |6ffaaf > AIRFLOW-1231|Use flask_wtf.CSRFProtect instead of flask_wtf.Csr|#2313 > |cac49e > AIRFLOW-1221|Fix DatabricksSubmitRunOperator Templating |#2308 > |0fa104 > AIRFLOW-1217|Enable logging in Sqoop hook |#2307 > |4f459b > AIRFLOW-1213|Add hcatalog parameters to the sqoop operator/hook|#2305 > |857850 > AIRFLOW-1208|Speed-up cli tests |#2301 > |21c142 > AIRFLOW-1207|Enable utils.helpers unit tests |#2300 > |8ac87b > AIRFLOW-1203|Tests failing after oauth upgrade |#2296 > |3e9c66 > AIRFLOW-1201|Update deprecated 'nose-parameterized' library to |#2298 > |d2d3e4 > AIRFLOW-1193|Add Checkr to Airflow user list |#2276 > |707238 > AIRFLOW-1189|Get pandas DataFrame using BigQueryHook fails |#2287 > |93666f > AIRFLOW-1188|Add max_bad_records param to GoogleCloudStorageToB|#2286 > |443e6b > AIRFLOW-1187|Obsolete package names in documentation |- |- > > AIRFLOW-1185|Incorrect url to PyPi |#2283 > |829755 > AIRFLOW-1181|Enable delete and list function for Google Cloud S|#2281 > |24f73c > AIRFLOW-1179|Pandas 0.20 broke Google BigQuery hook |#2279 > |ac9ccb > AIRFLOW-1177|variable json deserialize does not work at set def|#2540 > |65319a > AIRFLOW-1175|Add Pronto Tools to Airflow user list |#2277 > |86aafa > AIRFLOW-1173|Add Robinhood to list of Airflow users |#2271 > |379115 > AIRFLOW-1165|airflow webservice crashes on ubuntu16 - python3 |- |- > > AIRFLOW-1160|Upadte SparkSubmitOperator parameters |#2265 > |2e3f07 > AIRFLOW-1155|Add Tails.com to community |#2261 > |2fa690 > AIRFLOW-1149|Allow custom filters to be added to jinja2 |#2258 > |48135a > AIRFLOW-1141|Remove DAG.crawl_for_tasks method |#2275 > |a30fee > AIRFLOW-1140|DatabricksSubmitRunOperator should template the "j|#2255 > |e6d316 > AIRFLOW-1136|Invalid parameters are not captured for Sqoop oper|#2252 > |2ef4db > AIRFLOW-1125|Clarify documentation regarding fernet_key |#2251 > |831f8d > AIRFLOW-1122|Node strokes are too thin for people with color vi|#2246 > |a08761 > AIRFLOW-1121|airflow webserver --pid no longer write out pid fi|- |- > > AIRFLOW-1118|Add evo.company to Airflow users |#2243 > |f16914 > AIRFLOW-1112|Log which pool is full in scheduler when pool slot|#2242 > |74c1ce > AIRFLOW-1107|Add support for ftps non-default port |#2240 > |4d0c2f > AIRFLOW-1106|Add Groupalia/Letsbonus |#2239 > |945b42 > AIRFLOW-1095|ldap_auth memberOf should come from configuration |#2232 > |6b1c32 > AIRFLOW-1094|Invalid unit tests under `contrib/` |#2234 > |219c50 > AIRFLOW-1091|As a release manager I want to be able to compare |#2231 > |bfae42 > AIRFLOW-1090|Add HBO |#2230 > |177d34 > AIRFLOW-1089|Add Spark application arguments to SparkSubmitOper|#2229 > |e5b914 > AIRFLOW-1081|Task duration page is slow |#2226 > |0da512 > AIRFLOW-1075|Cleanup security docs |#2222 > |5a6f18 > AIRFLOW-1065|Add functionality for Azure Blob Storage |#2216 > |f1bc5f > AIRFLOW-1059|Reset_state_for_orphaned_task should operate in ba|#2205 > |e05d3b > AIRFLOW-1058|Improvements for SparkSubmitOperator |- |- > > AIRFLOW-1051|Add a test for resetdb to CliTests |#2198 > |15aee0 > AIRFLOW-1047|Airflow logs vulnerable to XSS |#2193 > |fe9ebe > AIRFLOW-1045|Make log level configurable via airflow.cfg |#2191 > |e739a5 > AIRFLOW-1043|Documentation issues for operators |#2188 > |b55f41 > AIRFLOW-1041|DockerOperator replaces its xcom_push method with |#2274 > |03704c > AIRFLOW-1040|Fix typos in comments/docstrings in models.py |#2174 > |d8c0f5 > AIRFLOW-1036|Exponential backoff should use randomization |#2262 > |66168e > AIRFLOW-1035|Exponential backoff retry logic should use 2 as ba|#2196 > |4ec932 > AIRFLOW-1034|Make it possible to connect to S3 in sigv4 regions|#2181 > |4c0905 > AIRFLOW-1031|'scheduled__' may replace with DagRun.ID_PREFIX in|#2613 > |aa3844 > AIRFLOW-1030|HttpHook error when creating HttpSensor |- |- > > AIRFLOW-1028|Databricks Operator for Airflow |#2202 > |53ca50 > AIRFLOW-1024|Handle CeleryExecutor errors gracefully |#2355 > |7af20f > AIRFLOW-1018|Scheduler DAG processes can not log to stdout |#2728 > |ef775d > AIRFLOW-1016|Allow HTTP HEAD request method on HTTPSensor |#2175 > |4c41f6 > AIRFLOW-1010|Add a convenience script for signing |#2169 > |a2b65a > AIRFLOW-1009|Remove SQLOperator from Concepts page |#2168 > |7d1144 > AIRFLOW-1007|Jinja sandbox is vulnerable to RCE |#2184 > |daa281 > AIRFLOW-1005|Speed up Airflow startup time |#na > |996dd3 > AIRFLOW-999 |Support for Redis database |#2165 > |8de850 > AIRFLOW-997 |Change setup.cfg to point to Apache instead of Max|#na > |75cd46 > AIRFLOW-995 |Update Github PR template |#2163 > |b62485 > AIRFLOW-994 |Add MiNODES to the AIRFLOW Active Users List |#2159 > |ca1623 > AIRFLOW-991 |Mark_success while a task is running leads to fail|- |- > > AIRFLOW-990 |DockerOperator fails when logging unicode string |#2155 > |6bbf54 > AIRFLOW-988 |SLA Miss Callbacks Are Repeated if Email is Not be|#2415 > |6e74d4 > AIRFLOW-985 |Extend the sqoop operator/hook with additional par|#2177 > |82eb20 > AIRFLOW-984 |Subdags unrecognized when subclassing SubDagOperat|#2152 > |a8bd16 > AIRFLOW-979 |Add GovTech GDS |#2149 > |b17bd3 > AIRFLOW-976 |Mark success running task causes it to fail |- |- > > AIRFLOW-969 |Catch bad python_callable argument at DAG construc|#2142 > |12901d > AIRFLOW-963 |Some code examples are not rendered in the airflow|#2139 > |f69c1b > AIRFLOW-960 |Add support for .editorconfig |#na > |f5cacc > AIRFLOW-959 |.gitignore file is disorganized and incomplete |#na > |3d3c14 > AIRFLOW-958 |Improve tooltip readability |#2134 > |b3c3eb > AIRFLOW-950 |Missing AWS integrations on documentation::integra|#2552 > |01be02 > AIRFLOW-947 |Make PrestoHook surface better messages when the P|#na > |6dd4b3 > AIRFLOW-945 |Revert psycopg2 workaround when psycopg2 2.7.1 is |- |- > > AIRFLOW-943 |Add Digital First Media to the Airflow users list |#2115 > |2cfe28 > AIRFLOW-942 |Add mytaxi to Airflow Users |#2111 > |d579e6 > AIRFLOW-935 |Impossible to use plugin executors |#2120 > |08a784 > AIRFLOW-926 |jdbc connector is broken due to jaydebeapi api upd|#2651 > |07ed29 > AIRFLOW-917 |Incorrectly formatted failure status message |#2109 > |b8164c > AIRFLOW-916 |Fix ConfigParser deprecation warning |#2108 > |ef6dd1 > AIRFLOW-911 |Add colouring and profiling info on tests |#2106 > |4f52db > AIRFLOW-903 |Add configuration setting for default DAG view. |#2103 > |cadfae > AIRFLOW-896 |BigQueryOperator fails to execute with certain inp|#2097 > |2bceee > AIRFLOW-891 |Webserver Clock Should Include Day |- |- > > AIRFLOW-889 |Minor error in the docstrings for BaseOperator. |#2084 > |50702d > AIRFLOW-887 |Add compatibility with future v0.16 |#na > |50902d > AIRFLOW-886 |Pass Operator result to post_execute hook |#na > |4da361 > AIRFLOW-885 |Add Change.org to the list of Airflow users |#2089 > |a279be > AIRFLOW-882 |Code example in docs has unnecessary DAG>>Operator|#2088 > |baa4cd > AIRFLOW-881 |Create SubDagOperator within DAG context manager w|#2087 > |0ed608 > AIRFLOW-880 |Fix remote log functionality inconsistencies for W|#2086 > |974b75 > AIRFLOW-877 |GoogleCloudStorageDownloadOperator: template_ext c|#2083 > |debc69 > AIRFLOW-875 |Allow HttpSensor params to be templated |#2080 > |62f503 > AIRFLOW-871 |multiple places use logging.warn() instead of warn|#2082 > |21d775 > AIRFLOW-866 |Add FTPSensor |#2070 > |5f87f8 > AIRFLOW-863 |Example DAG start dates should be recent to avoid |#2068 > |bbfd43 > AIRFLOW-862 |Add DaskExecutor |#2067 > |6e2210 > AIRFLOW-860 |Circular module dependency prevents loading of cus|- |- > > AIRFLOW-854 |Add Open Knowledge International to Airflow users |#2061 > |51a311 > AIRFLOW-842 |scheduler.clean_dirty raises warning: SAWarning: T|#2072 > |485280 > AIRFLOW-840 |Python3 encoding issue in Kerberos |#2158 > |639336 > AIRFLOW-836 |The paused and queryview endpoints are vulnerable |#2054 > |6aca2c > AIRFLOW-831 |Fix broken unit tests |#2050 > |b86194 > AIRFLOW-830 |Plugin manager should log to debug, not info |- |- > > AIRFLOW-829 |Reduce verbosity of successful Travis unit tests |- |- > > AIRFLOW-826 |Add Zendesk Hook |#2066 > |a09762 > AIRFLOW-823 |Make task instance details available via API |#2045 > |3f546e > AIRFLOW-822 |Close the connection before throwing exception in |#2038 > |4b6c38 > AIRFLOW-821 |Scheduler dagbag importing not Py3 compatible |#2039 > |fbb59b > AIRFLOW-809 |SqlAlchemy is_ ColumnOperator Causing Errors in MS|- |- > > AIRFLOW-802 |Integration of spark-submit |- |- > > AIRFLOW-781 |Allow DataFlowJavaOperator to accept jar file stor|#2037 > |259c86 > AIRFLOW-770 |HDFS hooks should support alternative ways of gett|#2056 > |261b65 > AIRFLOW-756 |Refactor ssh_hook and ssh_operator |- |- > > AIRFLOW-751 |SFTP file transfer functionality |#1999 > |fe0ede > AIRFLOW-725 |Make merge tool use OS' keyring for password stora|#1966 > |8c1695 > AIRFLOW-706 |Configuration shell commands are not split properl|#2053 > |0bb6f2 > AIRFLOW-705 |airflow.configuration.run_command output does not |- |- > > AIRFLOW-681 |homepage doc link should pointing to apache's repo|#2164 > |a8027a > AIRFLOW-654 |SSL for AMQP w/ Celery(Executor) |#2333 > |868bfe > AIRFLOW-645 |HttpHook ignores https |#2311 > |fd381a > AIRFLOW-365 |Code view in subdag trigger exception |#2043 > |cf102c > AIRFLOW-300 |Add Google Pubsub hook and operator |#2036 > |d231dc > AIRFLOW-289 |Use datetime.utcnow() to keep airflow system indep|#2618 > |20c83e > AIRFLOW-71 |docker_operator - pulling from private repositorie|#na > |d4406c > > Cheers, > Chris >
