+1 (binding) verified files, licenses, signatures & checksum

On Wed, 10 Aug 2022 at 14:40, Jarek Potiuk <ja...@potiuk.com> wrote:

> Hey all,
>
> > Yes, this is an RC3 (!) - don't ask :). Forgot to run `git pull` when
> preparing rc2  :facepalm: :)..
>
>
> I have just cut the new wave Airflow Providers packages. This email is
> calling a vote on the release,
> which will last for 72 hours - which means that it will end on Sat 13 Aug
> 4pm CEST 2022.
>
> Consider this my (binding) +1.
>
> *Summary of this wave*
>
> Apart from "regular" bugfix/features release, there are few notable things.
>
> * There is a new "major" release of Amazon provider with some backwards
> incompatibilities and a lot of deprecations removed as well as getting rid
> of a lot of duplications and refactorings that will make Amazon integration
> easier to extend and test in automated way when we get to AIP-47 (big
> shout-out to Vinc, Dennis and the whole Amazon team here).
>
> * Similarly Google provider had a lot of refactorings and changes - a lot
> of them under the hood and preparing AIP-47 to be released but also a
> number of improvements, fixes and new operators (Wojciech and Lukasz are
> leading the pack there)
>
> * Also Databricks provider got a bunch of new features and fixes (here
> Alex is the main contributor).
>
> * We have a lot more functionality moved to the relatively new
> "common.sql" provider (which now becomes our "everything-sql", including a
> bug fix that made it more difficult to use new SQL-providers in generic
> data transfer operators. The "common.sql" provider is a key to get Lineage
> information faster in the hands of all Airflow version users. Dmytro and
> Geoff(?) (@FanatoniQ) and Benji (@denimaalpaca) are the key players here.
> Finally we also have Release notes added to it (and I am entirely the one
> to blame for it).  Based on the learning with Common SQL we might introduce
> more "common" providers in the future.
>
> * Connected to that we have major (slightly backwards-incompatible)
> releases of presto, trino, exasol and hive providers as we are
> standardizing the SQL interface, method, parameter names and reusing more
> of the common functionality of the "common.sql" across multiple providers -
> less duplicated code, more common functionality is the main theme there.
>
> ( There are lot more smaller changes (see the status issue for details)
>
> --------------------------------------------------------------------------
>
> Airflow Providers are available at:
> https://dist.apache.org/repos/dist/dev/airflow/providers/
>
> *apache-airflow-providers-<PROVIDER>-*.tar.gz* are the binary
>  Python "sdist" release - they are also official "sources" for the
> provider packages.
>
> *apache_airflow_providers_<PROVIDER>-*.whl are the binary
>  Python "wheel" release.
>
> The test procedure for PMC members who would like to test the RC
> candidates are described in
>
> https://github.com/apache/airflow/blob/main/dev/README_RELEASE_PROVIDER_PACKAGES.md#verify-the-release-by-pmc-members
>
> and for Contributors:
>
>
> https://github.com/apache/airflow/blob/main/dev/README_RELEASE_PROVIDER_PACKAGES.md#verify-by-contributors
>
>
> Public keys are available at:
> https://dist.apache.org/repos/dist/release/airflow/KEYS
>
> Please vote accordingly:
>
> [ ] +1 approve
> [ ] +0 no opinion
> [ ] -1 disapprove with the reason
>
>
> Only votes from PMC members are binding, but members of the community are
> encouraged to test the release and vote with "(non-binding)".
>
> Please note that the version number excludes the 'rcX' string.
> This will allow us to rename the artifact without modifying
> the artifact checksums when we actually release.
>
> The status of testing the providers by the community is kept here:
>
> https://github.com/apache/airflow/issues/25640
>
> You can find packages as well as detailed changelog following the below
> links:
>
> https://pypi.org/project/apache-airflow-providers-amazon/5.0.0rc3/
> https://pypi.org/project/apache-airflow-providers-apache-drill/2.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-apache-druid/3.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-apache-hdfs/3.1.0rc3/
> https://pypi.org/project/apache-airflow-providers-apache-hive/4.0.0rc3/
> https://pypi.org/project/apache-airflow-providers-apache-livy/3.1.0rc3/
> https://pypi.org/project/apache-airflow-providers-apache-pinot/3.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-cncf-kubernetes/4.3.0rc3/
> https://pypi.org/project/apache-airflow-providers-common-sql/1.1.0rc3/
> https://pypi.org/project/apache-airflow-providers-databricks/3.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-dbt-cloud/2.1.0rc3/
> https://pypi.org/project/apache-airflow-providers-elasticsearch/4.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-exasol/4.0.0rc3/
> https://pypi.org/project/apache-airflow-providers-google/8.3.0rc3/
> https://pypi.org/project/apache-airflow-providers-hashicorp/3.1.0rc3/
> https://pypi.org/project/apache-airflow-providers-jdbc/3.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-microsoft-azure/4.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-microsoft-mssql/3.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-mysql/3.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-neo4j/3.1.0rc3/
> https://pypi.org/project/apache-airflow-providers-odbc/3.1.1rc3/
> https://pypi.org/project/apache-airflow-providers-oracle/3.3.0rc3/
> https://pypi.org/project/apache-airflow-providers-postgres/5.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-presto/4.0.0rc3/
> https://pypi.org/project/apache-airflow-providers-qubole/3.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-salesforce/5.1.0rc3/
> https://pypi.org/project/apache-airflow-providers-snowflake/3.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-sqlite/3.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-trino/4.0.0rc3/
> https://pypi.org/project/apache-airflow-providers-vertica/3.2.0rc3/
> https://pypi.org/project/apache-airflow-providers-yandex/3.1.0rc3/
>
> J.
>
>
>

Reply via email to