+1 (non-binding) verified signature and checksum, build and test passed.

On Fri, Nov 5, 2021 at 6:10 PM Sam Redai <s...@tabular.io> wrote:

> +1 (non-binding) signature, checksum, license, build and test
>
> On Fri, Nov 5, 2021 at 12:36 AM OpenInx <open...@gmail.com> wrote:
>
>> +1  (binding)
>>
>> 1. Download the source tarball, signature (.asc), and checksum
>> (.sha512):   OK
>> 2. Import gpg keys: download KEYS and run gpg --import
>> /path/to/downloaded/KEYS (optional if this hasn’t changed) :  OK
>> 3. Verify the signature by running: gpg --verify
>> apache-iceberg-xx-incubating.tar.gz.asc:  OK
>> 4. Verify the checksum by running: shasum -a 256 -c
>> apache-iceberg-0.12.1.tar.gz.sha512 apache-iceberg-0.12.1.tar.gz :  OK
>> 5. Untar the archive and go into the source directory: tar xzf
>> apache-iceberg-xx-incubating.tar.gz && cd apache-iceberg-xx-incubating:  OK
>> 6. Run RAT checks to validate license headers: dev/check-license: OK
>> 7. Build and test the project: ./gradlew build (use Java 8) :   OK
>> 8. Check the flink works fine by the following command line:
>>
>> ./bin/sql-client.sh embedded -j
>> /Users/openinx/Downloads/apache-iceberg-0.12.1/flink-runtime/build/libs/iceberg-flink-runtime-0.12.1.jar
>> shell
>>
>> CREATE CATALOG hadoop_prod WITH (
>> 'type'='iceberg',
>> 'catalog-type'='hadoop',
>> 'warehouse'='file:///Users/openinx/test/iceberg-warehouse'
>> );
>>
>> CREATE TABLE `hadoop_prod`.`default`.`flink_table` (
>> id BIGINT,
>> data STRING
>> );
>>
>> INSERT INTO `hadoop_prod`.`default`.`flink_table` VALUES (1, 'AAA');
>> SELECT * FROM `hadoop_prod`.`default`.`flink_table`;
>> +----+------+
>> | id | data |
>> +----+------+
>> | 1 | AAA |
>> +----+------+
>> 1 row in set
>>
>> Thanks all for the work.
>>
>> On Fri, Nov 5, 2021 at 2:20 PM Cheng Pan <cheng...@apache.org> wrote:
>>
>>> +1 (non-binding)
>>>
>>> The integration test based on the master branch of Apache Kyuubi
>>> (Incubating) passed.
>>>
>>> https://github.com/apache/incubator-kyuubi/pull/1338
>>>
>>> Thanks,
>>> Cheng Pan
>>>
>>> On Fri, Nov 5, 2021 at 1:19 PM Kyle Bendickson <k...@tabular.io> wrote:
>>> >
>>> >
>>> > +1 (binding)
>>> >
>>> > - Validated checksums, signatures, and licenses
>>> > -  Ran all of the unit tests
>>> > - Imported Files from Orc tables via Spark stored procedure, with
>>> floating point type columns and inspected the metrics afterwards
>>> > - Registered and used bucketed UDFs for various types such as integer
>>> and byte
>>> > - Created and dropped tables
>>> > - Ran MERGE INTO queries using Spark DDL
>>> > - Verified ability to read tables with parquet files with nested map
>>> type schema from various versions (both before and after Parquet 1.11.0 ->
>>> 1.11.1 upgrade)
>>> > - Tried to set a tblproperty to null (received error as expected)
>>> > - Full unit test suite
>>> > - Ran several Flink queries, both batch and streaming.
>>> > - Tested against a custom catalog
>>> >
>>> > My spark configuration was very similar to Ryan’s. I used Flink 1.12.1
>>> on a docker-compose setup via the Flink SQL client with 2 task managers.
>>> >
>>> > In addition to testing with a custom catalog, I also tested with HMS /
>>> Hive catalog with HDFS as storage as well as Hadoop Catalog with data on
>>> (local) HDFS.
>>> >
>>> > I’ve not gotten the Hive3 errors despite running unit tests several
>>> times.
>>> >
>>> > - Kyle (@kbendick)
>>> >
>>> >
>>> > On Thu, Nov 4, 2021 at 9:57 PM Daniel Weeks <dwe...@apache.org> wrote:
>>> >>
>>> >> +1 (binding)
>>> >>
>>> >> Verified sigs, sums, license, build and test.
>>> >>
>>> >> -Dan
>>> >>
>>> >> On Thu, Nov 4, 2021 at 4:30 PM Ryan Blue <b...@tabular.io> wrote:
>>> >>>
>>> >>> +1 (binding)
>>> >>>
>>> >>> Validated checksums, checked signature, ran tests (still a couple
>>> failing in Hive3)
>>> >>> Staged binaries from the release tarball
>>> >>> Tested Spark metadata tables
>>> >>> Used rewrite_manifests stored procedure in Spark
>>> >>> Updated to v2 using SET TBLPROPERTIES
>>> >>> Dropped and added partition fields
>>> >>> Replaced a table with itself using INSERT OVERWRITE
>>> >>> Tested custom catalogs
>>> >>>
>>> >>> Here’s my Spark config script in case anyone else wants to validate:
>>> >>>
>>> >>> /home/blue/Apps/spark-3.1.1-bin-hadoop3.2/bin/spark-shell \
>>> >>>     --conf spark.jars.repositories=
>>> https://repository.apache.org/content/repositories/orgapacheiceberg-1019/
>>> \
>>> >>>     --packages org.apache.iceberg:iceberg-spark3-runtime:0.12.1 \
>>> >>>     --conf
>>> spark.sql.extensions=org.apache.iceberg.spark.extensions.IcebergSparkSessionExtensions
>>> \
>>> >>>     --conf
>>> spark.sql.catalog.local=org.apache.iceberg.spark.SparkCatalog \
>>> >>>     --conf spark.sql.catalog.local.type=hadoop \
>>> >>>     --conf
>>> spark.sql.catalog.local.warehouse=/home/blue/tmp/hadoop-warehouse \
>>> >>>     --conf spark.sql.catalog.local.default-namespace=default \
>>> >>>     --conf
>>> spark.sql.catalog.prodhive=org.apache.iceberg.spark.SparkCatalog \
>>> >>>     --conf spark.sql.catalog.prodhive.type=hive \
>>> >>>     --conf
>>> spark.sql.catalog.prodhive.warehouse=/home/blue/tmp/prod-warehouse \
>>> >>>     --conf spark.sql.catalog.prodhive.default-namespace=default \
>>> >>>     --conf spark.sql.defaultCatalog=local
>>> >>>
>>> >>>
>>> >>> On Thu, Nov 4, 2021 at 1:02 PM Jack Ye <yezhao...@gmail.com> wrote:
>>> >>>>
>>> >>>> +1, non-binding
>>> >>>>
>>> >>>> ran checksum, build, unit tests, AWS integration tests and verified
>>> fixes in EMR 6.4.0.
>>> >>>>
>>> >>>> Best,
>>> >>>> Jack Ye
>>> >>>>
>>> >>>> On Tue, Nov 2, 2021 at 7:16 PM Kyle Bendickson <k...@tabular.io>
>>> wrote:
>>> >>>>>
>>> >>>>> Hi everyone,
>>> >>>>>
>>> >>>>>
>>> >>>>> I propose the following RC to be released as the official Apache
>>> Iceberg 0.12.1 release.
>>> >>>>>
>>> >>>>>
>>> >>>>> The commit id is d4052a73f14b63e1f519aaa722971dc74f8c9796
>>> >>>>>
>>> >>>>> * This corresponds to the tag: apache-iceberg-0.12.1-rc0
>>> >>>>>
>>> >>>>> *
>>> https://github.com/apache/iceberg/commits/apache-iceberg-0.12.1-rc0
>>> >>>>>
>>> >>>>> *
>>> https://github.com/apache/iceberg/tree/d4052a73f14b63e1f519aaa722971dc74f8c9796
>>> >>>>>
>>> >>>>>
>>> >>>>> The release tarball, signature, and checksums are here:
>>> >>>>>
>>> >>>>> *
>>> https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-0.12.1-rc0/
>>> >>>>>
>>> >>>>>
>>> >>>>> You can find the KEYS file here:
>>> >>>>>
>>> >>>>> * https://dist.apache.org/repos/dist/dev/iceberg/KEYS
>>> >>>>>
>>> >>>>>
>>> >>>>> Convenience binary artifacts are staged in Nexus. The Maven
>>> repository URL is:
>>> >>>>>
>>> >>>>> *
>>> https://repository.apache.org/content/repositories/orgapacheiceberg-1019/
>>> >>>>>
>>> >>>>>
>>> >>>>> This release includes the following changes:
>>> >>>>>
>>> >>>>>
>>> https://github.com/apache/iceberg/compare/apache-iceberg-0.12.0...apache-iceberg-0.12.1-rc0
>>> >>>>>
>>> >>>>>
>>> >>>>> Please download, verify, and test.
>>> >>>>>
>>> >>>>>
>>> >>>>> Please vote in the next 72 hours.
>>> >>>>>
>>> >>>>>
>>> >>>>> [ ] +1 Release this as Apache Iceberg <VERSION>
>>> >>>>>
>>> >>>>> [ ] +0
>>> >>>>>
>>> >>>>> [ ] -1 Do not release this because...
>>> >>>>>
>>> >>>>> --
>>> >>>>> Best,
>>> >>>>> Kyle Bendickson
>>> >>>>> Github: @kbendick
>>> >>>
>>> >>>
>>> >>>
>>> >>> --
>>> >>> Ryan Blue
>>> >>> Tabular
>>>
>>

Reply via email to