vote: 2.2.0 RC4
+1 (non-binding)
// using hbase-vote script
* Signature: ok
* Checksum : ok
* Rat check (1.8.0_212): ok
- mvn clean apache-rat:check
* Built from source (1.8.0_212): ok
- mvn clean install -DskipTests
* Unit tests pass (1.8.0_212): NOK (it's probably due to my environment,
I've seen a lot of test failures, need to validate individually).
- mvn test -P runAllTests
// if I get more time I will try to isolate tests that fail consistently
vs. timeouts.
hbase shell: ok
logs: ok
UI: ok
LTT 1M write/read 20%: ok
Spark Scala 2.3.3: ok // Not hbase-connectors
still see https://issues.apache.org/jira/browse/HBASE-21458
installed on pseudodistributed hadoop 2.9.2
had to modify hbase-site with the following
https://issues.apache.org/jira/browse/HBASE-22465
this is a dev environment so I'm not too concerned but once master came up
and I was able to do most of the vote steps, at some point RS went down
with the stack trace below. Again, I knew the implications and I'm working
on Hadoop 2.9 which according to the docs was not tested.
2019-05-23 18:41:49,141 WARN [Close-WAL-Writer-15] wal.AsyncFSWAL: close
old writer failed
java.io.FileNotFoundException: File does not exist: /apps/hbase/WALs/
hadoop.example.com,16020,1558627341708/hadoop.example.com
%2C16020%2C1558627341708.1558636709617
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:72)
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:62)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLease(FSNamesystem.java:2358)
at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.recoverLease(NameNodeRpcServer.java:790)
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.recoverLease(ClientNamenodeProtocolServerSideTranslatorPB.java:693)
at
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:503)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:989)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:871)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:817)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1893)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2606)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method)
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at
org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:121)
at
org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:88)
at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:867)
at
org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:304)
at
org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:301)
at
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at
org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:301)
at
org.apache.hadoop.hbase.util.FSHDFSUtils.recoverLease(FSHDFSUtils.java:283)
at
org.apache.hadoop.hbase.util.FSHDFSUtils.recoverDFSFileLease(FSHDFSUtils.java:216)
at
org.apache.hadoop.hbase.util.FSHDFSUtils.recoverFileLease(FSHDFSUtils.java:163)
at
org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.recoverAndClose(FanOutOneBlockAsyncDFSOutput.java:555)
at
org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.close(AsyncProtobufLogWriter.java:157)
at
org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$6(AsyncFSWAL.java:643)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by:
org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException): File
does not exist: /apps/hbase/WALs/hadoop.example.com,16020,1558627341708/
hadoop.example.com%2C16020%2C1558627341708.1558636709617
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:72)
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:62)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLease(FSNamesystem.java:2358)
at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.recoverLease(NameNodeRpcServer.java:790)
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.recoverLease(ClientNamenodeProtocolServerSideTranslatorPB.java:693)
at
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:503)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:989)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:871)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:817)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1893)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2606)
at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1489)
at org.apache.hadoop.ipc.Client.call(Client.java:1435)
at org.apache.hadoop.ipc.Client.call(Client.java:1345)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
at com.sun.proxy.$Proxy18.recoverLease(Unknown Source)
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.recoverLease(ClientNamenodeProtocolTranslatorPB.java:626)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:409)
at
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
at
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
at
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:346)
On Tue, May 21, 2019 at 6:51 PM Andrew Purtell <[email protected]> wrote:
> Actually let me change my vote to +1. None of these results are serious
> enough to sink the RC, although if someone besides myself can reproduce the
> TestClusterScopeQuotaThrottle result, we should have a JIRA and the release
> announcement could point to the JIRA as indication that particular feature
> might need a bugfix.
>
> On Tue, May 21, 2019 at 3:43 PM Andrew Purtell <[email protected]>
> wrote:
>
> > +0 at this time
> >
> > Signatures and sums: ok
> > RAT check: ok
> > Build from source: ok
> > Unit tests: some consistent failures
> >
> > This could be a problem with the test:
> >
> > [ERROR]
> >
> org.apache.hadoop.hbase.client.TestAsyncTableRSCrashPublish.test(org.apache.hadoop.hbase.client.TestAsyncTableRSCrashPublish)
> > [ERROR] Run 1: TestAsyncTableRSCrashPublish.test:77 Waiting timed out
> > after [60,000] msec
> > [ERROR] Run 2: TestAsyncTableRSCrashPublish.test:77 Waiting timed out
> > after [60,000] msec
> > [ERROR] Run 3: TestAsyncTableRSCrashPublish.test:77 Waiting timed out
> > after [60,000] msec
> > [ERROR] Run 4: TestAsyncTableRSCrashPublish.test:77 Waiting timed out
> > after [60,000] msec
> >
> > This is an interesting assertion failure:
> >
> > [ERROR] TestClusterRestartFailover.test:89 serverNode should not be
> null
> > when restart whole cluster
> >
> > These failures suggest cluster scope quotas might have an implementation
> > issue, this one is of most concern to me:
> >
> > [ERROR]
> >
> org.apache.hadoop.hbase.quotas.TestClusterScopeQuotaThrottle.testNamespaceClusterScopeQuota(org.apache.hadoop.hbase.quotas.TestClusterScopeQuotaThrottle)
> > [ERROR] Run 1:
> > TestClusterScopeQuotaThrottle.testNamespaceClusterScopeQuota:128
> > expected:<6> but was:<3>
> > [ERROR] Run 2:
> > TestClusterScopeQuotaThrottle.testNamespaceClusterScopeQuota:128
> > expected:<6> but was:<3>
> > [ERROR] Run 3:
> > TestClusterScopeQuotaThrottle.testNamespaceClusterScopeQuota:128
> > expected:<6> but was:<3>
> > [ERROR]
> >
> org.apache.hadoop.hbase.quotas.TestClusterScopeQuotaThrottle.testUserClusterScopeQuota(org.apache.hadoop.hbase.quotas.TestClusterScopeQuotaThrottle)
> > [ERROR] Run 1:
> > TestClusterScopeQuotaThrottle.testUserClusterScopeQuota:177 expected:<6>
> > but was:<0>
> > [ERROR] Run 2:
> > TestClusterScopeQuotaThrottle.testUserClusterScopeQuota:177 expected:<6>
> > but was:<10>
> > [ERROR] Run 3:
> > TestClusterScopeQuotaThrottle.testUserClusterScopeQuota:177 expected:<6>
> > but was:<10>
> >
> >
> > On Fri, May 17, 2019 at 11:12 PM Guanghao Zhang <[email protected]>
> > wrote:
> >
> >> Please vote on this release candidate (RC) for Apache HBase 2.2.0.
> >> This is the first release of the branch-2.2 line.
> >>
> >> The VOTE will remain open for at least 72 hours.
> >> [] +1
> >> [] +0/-0 Because ...
> >> [] -1 Do not release this package because ...
> >>
> >> The tag to be voted on is 2.2.0RC4. The release files, including
> >> signatures, digests, etc. can be found at:
> >> https://dist.apache.org/repos/dist/dev/hbase/2.2.0RC4/
> >>
> >> Maven artifacts are available in a staging repository at:
> >> https://repository.apache.org/content/repositories/orgapachehbase-1312
> >>
> >> Signatures used for HBase RCs can be found in this file:
> >> https://dist.apache.org/repos/dist/release/hbase/KEYS
> >>
> >> The list of bug fixes going into 2.2.0 can be found in included
> >> CHANGES.md and RELEASENOTES.md available here:
> >> https://dist.apache.org/repos/dist/dev/hbase/2.2.0RC4/CHANGES.md
> >> https://dist.apache.org/repos/dist/dev/hbase/2.2.0RC4/RELEASENOTES.md
> >>
> >> A detailed source and binary compatibility report for this release is
> >> available at
> >>
> >>
> https://dist.apache.org/repos/dist/dev/hbase/2.2.0RC4/api_compare_2.2.0RC4_to_2.1.4.html
> >>
> >> To learn more about Apache HBase, please see http://hbase.apache.org/
> >>
> >> Thanks,
> >> Guanghao Zhang
> >>
> >
> >
> > --
> > Best regards,
> > Andrew
> >
> > Words like orphans lost among the crosstalk, meaning torn from truth's
> > decrepit hands
> > - A23, Crosstalk
> >
>
>
> --
> Best regards,
> Andrew
>
> Words like orphans lost among the crosstalk, meaning torn from truth's
> decrepit hands
> - A23, Crosstalk
>