+1 (binding)

* Signature: ok
* Checksum : ok
* Compatibility Report vs 2.3.0: ok
* Rat check (1.8.0_275 + Hadoop2): ok
 - mvn clean apache-rat:check
* Built from source (1.8.0_275+ Hadoop2): ok
 - mvn clean install  -DskipTests
* Unit tests pass (1.8.0_275+ Hadoop2): ok
 - mvn package -P runSmallTests  -Dsurefire.rerunFailingTestsCount=3
* Rat check (1.8.0_275 + Hadoop3): ok
 - mvn clean apache-rat:check -D hadoop.profile=3.0
* Rat check (11.0.8 + Hadoop3): ok
 - mvn clean apache-rat:check -D hadoop.profile=3.0
* Built from source (1.8.0_275 + Hadoop3): ok
 - mvn clean install -D hadoop.profile=3.0 -DskipTests
* Built from source (11.0.8 + Hadoop3): ok
 - mvn clean install -D hadoop.profile=3.0 -DskipTests
* Unit tests pass (1.8.0_275 + Hadoop3): ok
 - mvn package -P runSmallTests -D hadoop.profile=3.0
-Dsurefire.rerunFailingTestsCount=3
* Unit tests pass (11.0.8 + Hadoop3): ok
 - mvn package -P runSmallTests -D hadoop.profile=3.0
-Dsurefire.rerunFailingTestsCount=3
* Unit tests: ok
 - Jenkins is looking pretty good overall.

* saintstack/hbase-downstreamer: had an issue with servlet api. This
looks like a transitive class path problem, but I haven't made time to
investigate in further detail; manual inspection of maven
dependency:tree looks fine, as pertains to the servlet-api jar.
Looking at that project's dependencies, I'm guessing this means that
HBase 2.4.0 is not compatible with Spark 1.6.0. For me, this is
concerning but not a blocker.
 - mvn -Dhbase.2.version=2.4.0
-Dhbase.staging.repository='https://repository.apache.org/content/repositories/orgapachehbase-1420/'
-Dhadoop.version=2.10.0 -Dslf4j.version=1.7.30 clean package
 - had to exclude zookeeper from hadoop-common, which still depends on
zookeeper-3.4.9 and is missing
org/apache/zookeeper/common/X509Exception$SSLContextException (same as
2.3.0 release)

On Wed, Dec 9, 2020 at 5:29 PM Nick Dimiduk <ndimi...@apache.org> wrote:
>
> On Wed, Dec 9, 2020 at 5:23 PM Andrew Purtell <andrew.purt...@gmail.com> 
> wrote:
>>
>> ICYMI, the nightly build passes because the JDK version there has three 
>> components “11.0.6” and yours fails because your JDK version has four 
>> components (“11.0.9.1”).
>
>
> Thanks, I did miss this (and I had to look up ICYMI).
>
>> As this is a release vote you do not need to withdraw the veto, it does not 
>> block, but I would ask that you fix your local environment to conform the 
>> the limitations of Jetty as transitively pulled in via that optional profile 
>> and retest. This is not an HBase issue and it should not be relevant in 
>> voting. (IMHO)
>
>
> Since this is indeed a jetty issue, I agree that this is not an HBase issue 
> and I rescind the -1. Let me grab a slightly different JDK11 build and resume 
> my evaluation.
>
>> > On Dec 9, 2020, at 5:10 PM, Andrew Purtell <andrew.purt...@gmail.com> 
>> > wrote:
>> >
>> > Nick,
>> >
>> > Given the nature of this issue I’d ask you to try Duo’s suggestion and if 
>> > an earlier version of Hadoop 3 succeeds that this be sufficient this time 
>> > around.
>> >
>> > All,
>> >
>> > I will start a DISCUSS thread as follow up as to what should be considered 
>> > required and veto worthy for a RC and what should not, with regard to 
>> > optional build profiles. In my opinion ‘required‘ should be defined as 
>> > what is enabled by default in the release build, and ‘optional’ is 
>> > everything else until we agree to specifically include one or more 
>> > optional build profiles to the required list.
>> >
>> >
>> >> On Dec 9, 2020, at 4:31 PM, 张铎 <palomino...@gmail.com> wrote:
>> >>
>> >> OK, I think the problem is a bug in jetty 9.3, the JavaVersion
>> >> implementation for 9.3 and 9.4 are completely different, there is no
>> >> problem to parse 11.0.9.1 for jetty 9.4 but for jetty 9.3, it can only 
>> >> pass
>> >> version with two dots, i.e, 11.0.9.
>> >>
>> >> I think you could add -Dhadoop-three.version to set the hadoop 3 version
>> >> explicitly to a newer version which uses jetty 9.4 to solve the problem,
>> >> IIRC all the newest release for each active release line has upgrade to
>> >> jetty 9.4 and that's why we need to shade jetty as jetty 9.3 and 9.4 are
>> >> incompatible.
>> >>
>> >> Thanks.
>> >>
>> >> 张铎(Duo Zhang) <palomino...@gmail.com> 于2020年12月10日周四 上午8:21写道:
>> >>
>> >>> On nightly jdk11 build the jdk version is
>> >>>
>> >>> AdoptOpenJDK-11.0.6+10
>> >>>
>> >>> Andrew Purtell <apurt...@apache.org> 于2020年12月10日周四 上午7:21写道:
>> >>>
>> >>>> Let me rephrase.
>> >>>>
>> >>>> I'm all for testing the optional configurations. I'm less supportive of
>> >>>> vetoing releases when an optional configuration has some issue due to a
>> >>>> third party component. I would like to see us veto only on the required
>> >>>> configurations, and otherwise file JIRAs to fix up the nits on the
>> >>>> optional
>> >>>> ones.
>> >>>>
>> >>>>
>> >>>> On Wed, Dec 9, 2020 at 3:19 PM Andrew Purtell <apurt...@apache.org>
>> >>>> wrote:
>> >>>>
>> >>>>>> parseJDK9:71, JavaVersion (org.eclipse.jetty.util)
>> >>>>>
>> >>>>> So unless I am mistaken, some Jetty utility class is not able to parse
>> >>>> the
>> >>>>> version string of your particular JDK/JRE.
>> >>>>>
>> >>>>> We can try to downgrade Jetty but I am not sure in general how we are
>> >>>>> supposed to take on the risk of third party dependencies doing the 
>> >>>>> wrong
>> >>>>> thing in an optional configuration. I for one do not want to deal with 
>> >>>>> a
>> >>>>> combinatorial explosion of transitive dependencies when releasing.
>> >>>>>
>> >>>>>
>> >>>>> On Wed, Dec 9, 2020 at 2:41 PM Nick Dimiduk <ndimi...@apache.org>
>> >>>> wrote:
>> >>>>>
>> >>>>>> This is coming out of Jetty + Hadoop. This build has a regression in
>> >>>> our
>> >>>>>> JDK11 support. Or perhaps there's a regression in the upstream Hadoop
>> >>>>>> against which JDK11 builds.
>> >>>>>>
>> >>>>>> I'm afraid I must vote -1 until we can sort out the issue. I'd
>> >>>> appreciate
>> >>>>>> it if someone else can attempt to repro, help ensure it's not just me.
>> >>>>>>
>> >>>>>> Thanks,
>> >>>>>> Nick
>> >>>>>>
>> >>>>>> (Apologies for the crude stack trace; this is copied out of an 
>> >>>>>> attached
>> >>>>>> debugger)
>> >>>>>>
>> >>>>>> parseJDK9:71, JavaVersion (org.eclipse.jetty.util)
>> >>>>>> parse:49, JavaVersion (org.eclipse.jetty.util)
>> >>>>>> <clinit>:43, JavaVersion (org.eclipse.jetty.util)
>> >>>>>> findAndFilterContainerPaths:185, WebInfConfiguration
>> >>>>>> (org.eclipse.jetty.webapp)
>> >>>>>> preConfigure:155, WebInfConfiguration (org.eclipse.jetty.webapp)
>> >>>>>> preConfigure:485, WebAppContext (org.eclipse.jetty.webapp)
>> >>>>>> doStart:521, WebAppContext (org.eclipse.jetty.webapp)
>> >>>>>> start:68, AbstractLifeCycle (org.eclipse.jetty.util.component)
>> >>>>>> start:131, ContainerLifeCycle (org.eclipse.jetty.util.component)
>> >>>>>> doStart:113, ContainerLifeCycle (org.eclipse.jetty.util.component)
>> >>>>>> doStart:61, AbstractHandler (org.eclipse.jetty.server.handler)
>> >>>>>> start:68, AbstractLifeCycle (org.eclipse.jetty.util.component)
>> >>>>>> start:131, ContainerLifeCycle (org.eclipse.jetty.util.component)
>> >>>>>> start:427, Server (org.eclipse.jetty.server)
>> >>>>>> doStart:105, ContainerLifeCycle (org.eclipse.jetty.util.component)
>> >>>>>> doStart:61, AbstractHandler (org.eclipse.jetty.server.handler)
>> >>>>>> doStart:394, Server (org.eclipse.jetty.server)
>> >>>>>> start:68, AbstractLifeCycle (org.eclipse.jetty.util.component)
>> >>>>>> start:1140, HttpServer2 (org.apache.hadoop.http)
>> >>>>>> start:177, NameNodeHttpServer (org.apache.hadoop.hdfs.server.namenode)
>> >>>>>> startHttpServer:872, NameNode (org.apache.hadoop.hdfs.server.namenode)
>> >>>>>> initialize:694, NameNode (org.apache.hadoop.hdfs.server.namenode)
>> >>>>>> <init>:940, NameNode (org.apache.hadoop.hdfs.server.namenode)
>> >>>>>> <init>:913, NameNode (org.apache.hadoop.hdfs.server.namenode)
>> >>>>>> createNameNode:1646, NameNode (org.apache.hadoop.hdfs.server.namenode)
>> >>>>>> createNameNode:1314, MiniDFSCluster (org.apache.hadoop.hdfs)
>> >>>>>> configureNameService:1083, MiniDFSCluster (org.apache.hadoop.hdfs)
>> >>>>>> createNameNodesAndSetConf:958, MiniDFSCluster (org.apache.hadoop.hdfs)
>> >>>>>> initMiniDFSCluster:890, MiniDFSCluster (org.apache.hadoop.hdfs)
>> >>>>>> <init>:518, MiniDFSCluster (org.apache.hadoop.hdfs)
>> >>>>>> build:477, MiniDFSCluster$Builder (org.apache.hadoop.hdfs)
>> >>>>>> startMiniDFSCluster:108, AsyncFSTestBase
>> >>>>>> (org.apache.hadoop.hbase.io.asyncfs)
>> >>>>>> setUp:87, TestFanOutOneBlockAsyncDFSOutput
>> >>>>>> (org.apache.hadoop.hbase.io.asyncfs)
>> >>>>>> invoke0:-1, NativeMethodAccessorImpl (jdk.internal.reflect)
>> >>>>>> invoke:62, NativeMethodAccessorImpl (jdk.internal.reflect)
>> >>>>>> invoke:43, DelegatingMethodAccessorImpl (jdk.internal.reflect)
>> >>>>>> invoke:566, Method (java.lang.reflect)
>> >>>>>> runReflectiveCall:59, FrameworkMethod$1 (org.junit.runners.model)
>> >>>>>> run:12, ReflectiveCallable (org.junit.internal.runners.model)
>> >>>>>> invokeExplosively:56, FrameworkMethod (org.junit.runners.model)
>> >>>>>> invokeMethod:33, RunBefores (org.junit.internal.runners.statements)
>> >>>>>> evaluate:24, RunBefores (org.junit.internal.runners.statements)
>> >>>>>> evaluate:27, RunAfters (org.junit.internal.runners.statements)
>> >>>>>> evaluate:38, SystemExitRule$1 (org.apache.hadoop.hbase)
>> >>>>>> call:288, FailOnTimeout$CallableStatement
>> >>>>>> (org.junit.internal.runners.statements)
>> >>>>>> call:282, FailOnTimeout$CallableStatement
>> >>>>>> (org.junit.internal.runners.statements)
>> >>>>>> run:264, FutureTask (java.util.concurrent)
>> >>>>>> run:834, Thread (java.lang)
>> >>>>>>
>> >>>>>> On Wed, Dec 9, 2020 at 2:08 PM Nick Dimiduk <ndimi...@apache.org>
>> >>>> wrote:
>> >>>>>>
>> >>>>>>> On Mon, Dec 7, 2020 at 1:51 PM Nick Dimiduk <ndimi...@apache.org>
>> >>>>>> wrote:
>> >>>>>>>
>> >>>>>>>> Has anyone successfully built/run this RC with JDK11 and Hadoop3
>> >>>>>> profile?
>> >>>>>>>> I'm seeing test failures locally in the hbase-asyncfs module.
>> >>>>>>>> Reproducible with:
>> >>>>>>>>
>> >>>>>>>> $
>> >>>>>>>>
>> >>>>>>
>> >>>> JAVA_HOME=/Library/Java/JavaVirtualMachines/adoptopenjdk-11.jdk/Contents/Home
>> >>>>>>>> mvn clean install -Dhadoop.profile=3.0
>> >>>>>>>>
>> >>>>>>
>> >>>> -Dtest=org.apache.hadoop.hbase.io.asyncfs.TestFanOutOneBlockAsyncDFSOutput
>> >>>>>>>> ...
>> >>>>>>>> [INFO] Running
>> >>>>>>>> org.apache.hadoop.hbase.io.asyncfs.TestFanOutOneBlockAsyncDFSOutput
>> >>>>>>>>
>> >>>>>>>>
>> >>>>>>>> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time
>> >>>> elapsed:
>> >>>>>>>> 1.785 s <<< FAILURE! - in
>> >>>>>>>> org.apache.hadoop.hbase.io.asyncfs.TestFanOutOneBlockAsyncDFSOutput
>> >>>>>>>>
>> >>>>>>>> [ERROR]
>> >>>>>>>> org.apache.hadoop.hbase.io.asyncfs.TestFanOutOneBlockAsyncDFSOutput
>> >>>>>> Time
>> >>>>>>>> elapsed: 1.775 s  <<< ERROR!
>> >>>>>>>>
>> >>>>>>>> java.lang.ExceptionInInitializerError
>> >>>>>>>>
>> >>>>>>>>
>> >>>>>>>>       at
>> >>>>>>>> org.apache.hadoop.hbase.io
>> >>>>>>
>> >>>> .asyncfs.TestFanOutOneBlockAsyncDFSOutput.setUp(TestFanOutOneBlockAsyncDFSOutput.java:87)
>> >>>>>>>>
>> >>>>>>>> Caused by: java.lang.IllegalArgumentException: Invalid Java version
>> >>>>>>>> 11.0.9.1
>> >>>>>>>>
>> >>>>>>>>       at
>> >>>>>>>> org.apache.hadoop.hbase.io
>> >>>>>>
>> >>>> .asyncfs.TestFanOutOneBlockAsyncDFSOutput.setUp(TestFanOutOneBlockAsyncDFSOutput.java:87)
>> >>>>>>>>
>> >>>>>>>
>> >>>>>>> This failure is not isolated to macOS. I ran this build on an Ubuntu
>> >>>> VM
>> >>>>>>> with the same AdoptOpenJDK 11.0.9.1. Why don't we see this in
>> >>>> Jenkins?
>> >>>>>>>
>> >>>>>>> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time
>> >>>> elapsed:
>> >>>>>>> 0.011 s <<< FAILURE! - in
>> >>>>>>> org.apache.hadoop.hbase.regionserver.wal.TestAsyncProtobufLog
>> >>>>>>>
>> >>>>>>> [ERROR] org.apache.hadoop.hbase.regionserver.wal.TestAsyncProtobufLog
>> >>>>>>> Time elapsed: 0.003 s  <<< ERROR!
>> >>>>>>>
>> >>>>>>> java.lang.ExceptionInInitializerError
>> >>>>>>>
>> >>>>>>>       at
>> >>>>>>>
>> >>>>>>
>> >>>> org.apache.hadoop.hbase.regionserver.wal.TestAsyncProtobufLog.setUpBeforeClass(TestAsyncProtobufLog.java:56)
>> >>>>>>>
>> >>>>>>>       Caused by: java.lang.IllegalArgumentException: Invalid Java
>> >>>>>> version
>> >>>>>>> 11.0.9.1
>> >>>>>>>       at
>> >>>>>>>
>> >>>>>>
>> >>>> org.apache.hadoop.hbase.regionserver.wal.TestAsyncProtobufLog.setUpBeforeClass(TestAsyncProtobufLog.java:56)
>> >>>>>>>
>> >>>>>>> On Thu, Dec 3, 2020 at 4:05 PM Andrew Purtell <apurt...@apache.org>
>> >>>>>> wrote:
>> >>>>>>>>
>> >>>>>>>>> Please vote on this Apache hbase release candidate, hbase-2.4.0RC1
>> >>>>>>>>>
>> >>>>>>>>> The VOTE will remain open for at least 72 hours.
>> >>>>>>>>>
>> >>>>>>>>> [ ] +1 Release this package as Apache hbase 2.4.0
>> >>>>>>>>> [ ] -1 Do not release this package because ...
>> >>>>>>>>>
>> >>>>>>>>> The tag to be voted on is 2.4.0RC1:
>> >>>>>>>>>
>> >>>>>>>>>   https://github.com/apache/hbase/tree/2.4.0RC1
>> >>>>>>>>>
>> >>>>>>>>> The release files, including signatures, digests, as well as
>> >>>>>> CHANGES.md
>> >>>>>>>>> and RELEASENOTES.md included in this RC can be found at:
>> >>>>>>>>>
>> >>>>>>>>>   https://dist.apache.org/repos/dist/dev/hbase/2.4.0RC1/
>> >>>>>>>>>
>> >>>>>>>>> Customarily Maven artifacts would be available in a staging
>> >>>>>> repository.
>> >>>>>>>>> Unfortunately I was forced to terminate the Maven deploy step after
>> >>>>>>>>> the upload ran for more than four hours and my build equipment
>> >>>>>>>>> needed to be relocated, with loss of network connectivity. This RC
>> >>>> has
>> >>>>>>>>> been delayed long enough. A temporary Maven repository is not a
>> >>>>>>>>> requirement for a vote. I will retry Maven deploy tomorrow. I can
>> >>>>>>>>> promise the artifacts for this RC will be staged in Apache Nexus
>> >>>> and
>> >>>>>>>>> ready for release well ahead of the earliest possible time this
>> >>>> vote
>> >>>>>>>>> can complete.
>> >>>>>>>>>
>> >>>>>>>>> Artifacts were signed with the apurt...@apache.org key which can
>> >>>> be
>> >>>>>>>>> found
>> >>>>>>>>> in:
>> >>>>>>>>>
>> >>>>>>>>>   https://dist.apache.org/repos/dist/release/hbase/KEYS
>> >>>>>>>>>
>> >>>>>>>>> The API compatibility report for this RC can be found at:
>> >>>>>>>>>
>> >>>>>>>>>
>> >>>>>>>>>
>> >>>>>>>>>
>> >>>>>>
>> >>>> https://dist.apache.org/repos/dist/dev/hbase/2.4.0RC1/api_compare_2.4.0RC1_to_2.3.0.html
>> >>>>>>>>>
>> >>>>>>>>> The changes are mostly added methods, which conform to the
>> >>>>>> compatibility
>> >>>>>>>>> guidelines for a new minor release. There is one change to the
>> >>>> public
>> >>>>>>>>> Region interface that alters the return type of a method. This is
>> >>>>>>>>> equivalent to a removal then addition and can be a binary
>> >>>>>> compatibility
>> >>>>>>>>> problem. However to your RM's eye the change looks intentional and
>> >>>> is
>> >>>>>>>>> part of an API improvement project, and a compatibility method is
>> >>>> not
>> >>>>>>>>> possible here because Java doesn't consider return type when
>> >>>> deciding
>> >>>>>> if
>> >>>>>>>>> one method signature duplicates another.
>> >>>>>>>>>
>> >>>>>>>>> To learn more about Apache HBase, please see
>> >>>>>>>>>
>> >>>>>>>>>   http://hbase.apache.org/
>> >>>>>>>>>
>> >>>>>>>>> Thanks,
>> >>>>>>>>> Your HBase Release Manager
>> >>>>>>>>>
>> >>>>>>>>
>> >>>>>>
>> >>>>>
>> >>>>>
>> >>>>> --
>> >>>>> Best regards,
>> >>>>> Andrew
>> >>>>>
>> >>>>> Words like orphans lost among the crosstalk, meaning torn from truth's
>> >>>>> decrepit hands
>> >>>>>  - A23, Crosstalk
>> >>>>>
>> >>>>
>> >>>>
>> >>>> --
>> >>>> Best regards,
>> >>>> Andrew
>> >>>>
>> >>>> Words like orphans lost among the crosstalk, meaning torn from truth's
>> >>>> decrepit hands
>> >>>>  - A23, Crosstalk
>> >>>>
>> >>>

Reply via email to