-1 due to PHOENIX-6377 (wrong HBase version pulled in when external
projects depend on phoenix-client)

Checked sigs and checksums: OK
apache-rat:check: OK
mvn clean install -DskipTests: OK
mvn verify: Never got a clean run, miniclusters kept crashing midway
through the run, but I see others did, so: OK
Examined Jacob's perf runs on PHOENIX-6734, no regressions significant
enough to sink the release: OK
Sanity check locally in sqlline, can create indexes, do aggregations, CRUD
check passes: OK

I was about to +1 when Lars reported 6377. Since Istvan has already
contributed a fix for it, it seems worth it to me to cherry-pick it to 4.16
and spin up another RC.

Geoffrey

On Tue, Feb 9, 2021 at 7:12 PM Jacob Isaac <[email protected]> wrote:

> +1 (non-binding)
>
> mvn clean install -DskipTests -Dhbase.profile=(1.6/1.3) (successful)
>
> Tested various Upserts, Queries on various types of tables. (successful)
> Ran pherf (successful)
> ./bin/pherf-standalone.py  -l -q -z localhost -schemaFile
> <-schema-file-name>.sql -scenarioFile <scenario-file-name>.xml (successful)
>
> Nit: ./bin/phoenix_utils.py gives the following error
> testjar:
> Traceback (most recent call last):
>   File "./bin/phoenix_utils.py", line 215, in <module>
>     print("phoenix_queryserver_jar:", phoenix_queryserver_jar)
> NameError: name 'phoenix_queryserver_jar' is not defined
>
>
> Ran the following MR jobs
> /hbase/bin/hbase org.apache.phoenix.mapreduce.index.IndexTool -op
> /tmp/indexing.log -v AFTER -dt <table-name> -it <index-table-name>
> (successful)
> /hbase/bin/hbase org.apache.hadoop.hbase.mapreduce.RowCounter <table-name>
> (successful)
>
> Had executed perf workloads using the RC1 build
> The results and analysis can be found here -
>
> https://docs.google.com/document/d/19QHG6vvdxwCNkT3nqu8N-ift_1OIn161pqtJx1UcXiY/edit#
>
> Thanks
> Jacob
>
> On Mon, Feb 8, 2021 at 6:48 PM Ankit Singhal <[email protected]> wrote:
>
> > +1 (binding)
> >
> >  * Download source and build - OK
> >  * Ran some DDLs and DMLs on fresh cluster - OK
> >  * Signatures and checksums for src and bin(1.3)- OK
> >  * apache-rat:check - SUCCESS
> >  * CHANGES and RELEASENOTES - OK
> >  * Unit tests( mvn clean install -Dit.test=noITs, though code-coverage
> > check failed for me) - Ok
> >
> > Regards,
> > Ankit Singhal
> >
> > On Mon, Feb 8, 2021 at 1:40 PM Chinmay Kulkarni <
> > [email protected]>
> > wrote:
> >
> > > +1 (Binding)
> > >
> > > Tested against hbase-1.3 and hbase-1.6
> > >
> > > * Build from source (mvn clean install -DskipTests
> > > -Dhbase.profile=1.3/1.6): OK
> > > * Green build: OK (thanks for triggering this Viraj)
> > > * Did some basic DDL, queries, upserts, deletes and everything looked
> > fine:
> > > OK
> > > * Did some upgrade testing: Create tables, views, indices from an old
> > > client, query, upsert. Then upgrade to 4.16 metadata, query, upsert
> from
> > an
> > > old client, then upgrade the client and query, upsert from a new
> client:
> > OK
> > > * Verified checksums: OK
> > > * Verified signatures: OK
> > > * mvn clean apache-rat:check: OK
> > >
> > > On Sun, Feb 7, 2021 at 10:03 PM Viraj Jasani <[email protected]>
> wrote:
> > >
> > > > +1 (non-binding)
> > > >
> > > > Clean build:
> > > >
> > >
> >
> https://ci-hadoop.apache.org/job/Phoenix/job/Phoenix-mulitbranch/job/4.16/29/
> > > >
> > > > Tested against HBase-1.6 profile:
> > > >
> > > > * Checksum : ok
> > > > * Rat check (1.8.0_171): ok
> > > >  - mvn clean apache-rat:check
> > > > * Built from source (1.8.0_171): ok
> > > >  - mvn clean install  -DskipTests
> > > > * Basic testing with mini cluster: ok
> > > > * Unit tests pass (1.8.0_171): failed (passing eventually)
> > > >  - mvn clean package  && mvn verify  -Dskip.embedded
> > > >
> > > >
> > > > [ERROR] Tests run: 23, Failures: 0, Errors: 1, Skipped: 0, Time
> > elapsed:
> > > > 197.428 s <<< FAILURE! - in org.apache.phoenix.end2end.AggregateIT
> > > > [ERROR]
> > > >
> > >
> >
> testOrderByOptimizeForClientAggregatePlanBug4820(org.apache.phoenix.end2end.AggregateIT)
> > > > Time elapsed: 9.055 s  <<< ERROR!
> > > > java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to
> > create
> > > > new native thread
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:239)
> > > >         at
> > > >
> > org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:273)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:434)
> > > >         at
> > > >
> > org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:308)
> > > >
> > > >
> > > > [ERROR] Tests run: 37, Failures: 0, Errors: 1, Skipped: 0, Time
> > elapsed:
> > > > 204.243 s <<< FAILURE! - in
> > > org.apache.phoenix.end2end.ArrayAppendFunctionIT
> > > > [ERROR]
> > > >
> > >
> >
> testUpsertArrayAppendFunctionVarchar(org.apache.phoenix.end2end.ArrayAppendFunctionIT)
> > > > Time elapsed: 4.286 s  <<< ERROR!
> > > > org.apache.phoenix.exception.PhoenixIOException:
> > > > org.apache.hadoop.hbase.DoNotRetryIOException: N000065:
> > > > java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to
> > create
> > > > new native thread
> > > >         at
> > > >
> > org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:122)
> > > >         at
> > > >
> > >
> >
> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:2151)
> > > >         at
> > > >
> > >
> >
> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17317)
> > > >
> > > >
> > > > [ERROR] Tests run: 28, Failures: 0, Errors: 1, Skipped: 0, Time
> > elapsed:
> > > > 147.854 s <<< FAILURE! - in
> > > org.apache.phoenix.end2end.ArrayRemoveFunctionIT
> > > > [ERROR]
> > > >
> > >
> >
> testArrayRemoveFunctionWithNull(org.apache.phoenix.end2end.ArrayRemoveFunctionIT)
> > > > Time elapsed: 2.519 s  <<< ERROR!
> > > > org.apache.phoenix.exception.PhoenixIOException:
> > > > java.util.concurrent.ExecutionException: java.lang.OutOfMemoryError:
> > > unable
> > > > to create new native thread
> > > >         at
> > > >
> > >
> >
> org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:146)
> > > >         at
> > > >
> > >
> >
> org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:1511)
> > > >         at
> > > >
> > >
> >
> org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1901)
> > > >         at
> > > >
> > >
> >
> org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:3065)
> > > >
> > > >
> > > > [ERROR] Tests run: 10, Failures: 0, Errors: 1, Skipped: 0, Time
> > elapsed:
> > > > 5,069.234 s <<< FAILURE! - in
> > > >
> > >
> >
> org.apache.phoenix.end2end.PermissionNSDisabledWithCustomAccessControllerIT
> > > > [ERROR]
> > > >
> > >
> >
> testAutomaticGrantWithIndexAndView(org.apache.phoenix.end2end.PermissionNSDisabledWithCustomAccessControllerIT)
> > > > Time elapsed: 2,572.586 s  <<< ERROR!
> > > > java.lang.reflect.UndeclaredThrowableException
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1862)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:340)
> > > >         at
> > > >
> > >
> >
> org.apache.phoenix.end2end.BasePermissionsIT.verifyAllowed(BasePermissionsIT.java:776)
> > > >         at
> > > >
> > >
> >
> org.apache.phoenix.end2end.BasePermissionsIT.verifyAllowed(BasePermissionsIT.java:769)
> > > >
> > > > Tests are passing in subsequent runs.
> > > >
> > > >
> > > > On 2021/02/06 04:53:48, Xinyi Yan <[email protected]> wrote:
> > > > > Hello Everyone,
> > > > >
> > > > > This is a call for a vote on Apache Phoenix 4.16.0 RC2. This is the
> > > next
> > > > > minor release of Phoenix 4, compatible with Apache HBase 1.3, 1.4,
> > 1.5
> > > > > and 1.6.
> > > > >
> > > > > The VOTE will remain open for at least 72 hours.
> > > > >
> > > > > [ ] +1 Release this package as Apache phoenix 4.16.0
> > > > > [ ] -1 Do not release this package because ...
> > > > >
> > > > > The tag to be voted on is 4.16.0RC2
> > > > > https://github.com/apache/phoenix/tree/4.16.0RC2
> > > > >
> > > > > The release files, including signatures, digests, as well as
> > CHANGES.md
> > > > > and RELEASENOTES.md included in this RC can be found at:
> > > > > https://dist.apache.org/repos/dist/dev/phoenix/phoenix-4.16.0RC2/
> > > > >
> > > > > For a complete list of changes, see:
> > > > >
> > > >
> > >
> >
> https://dist.apache.org/repos/dist/dev/phoenix/phoenix-4.16.0RC2/CHANGES.md
> > > > >
> > > > > Artifacts are signed with my "CODE SIGNING KEY":
> > > > > E4882DD3AB711587
> > > > >
> > > > > KEYS file available here:
> > > > > https://dist.apache.org/repos/dist/dev/phoenix/KEYS
> > > > >
> > > > >
> > > > > Thanks,
> > > > > Xinyi
> > > > >
> > > > > <
> > > >
> > >
> >
> https://dist.apache.org/repos/dist/dev/phoenix/phoenix-4.16.0RC1/CHANGES.md
> > > > >
> > > > >
> > > >
> > >
> > >
> > > --
> > > Chinmay Kulkarni
> > >
> >
>

Reply via email to