Hi Istvan,

I have some progress here!

I was able to overcome the above problem by adding -Dhbase24.compat.version
and  -pl "!phoenix-hbase-compat-2.4.0"

mvn clean install -Dhbase.profile=2.4 -Dhbase.version=2.4.9-SNAPSHOT
-Dhbase24.compat.version=2.4.9-SNAPSHOT -DskipTests -pl
"!phoenix-hbase-compat-2.4.0"

Without explicitly setting -Dhbase24.compat.version it was using 2.4.1 and
I think this was the reason why it was still using HdfsFileStatus
fromHadoop 2.x.

Without -pl "!phoenix-hbase-compat-2.4.0" I've got this compilation error:

ERROR] Failed to execute goal
org.apache.maven.plugins:maven-compiler-plugin:3.8.1:compile
(default-compile) on project phoenix-hbase-compat-2.4.0: Compilation
failure: Compilation failure:
[ERROR]
/home/ubuntu/git/apache/phoenix/phoenix-hbase-compat-2.4.0/src/main/java/org/apache/phoenix/compat/hbase/CompatUtil.java:[72,41]
cannot find symbol
[ERROR]   symbol:   method
getBytesPerChecksum(org.apache.hadoop.conf.Configuration)
[ERROR]   location: class org.apache.hadoop.hbase.regionserver.HStore
[ERROR]
/home/ubuntu/git/apache/phoenix/phoenix-hbase-compat-2.4.0/src/main/java/org/apache/phoenix/compat/hbase/CompatUtil.java:[71,37]
cannot find symbol
[ERROR]   symbol:   method
getChecksumType(org.apache.hadoop.conf.Configuration)
[ERROR]   location: class org.apache.hadoop.hbase.regionserver.HStore
[ERROR]
/home/ubuntu/git/apache/phoenix/phoenix-hbase-compat-2.4.0/src/main/java/org/apache/phoenix/compat/hbase/CompatUtil.java:[136,22]
cannot find symbol
[ERROR]   symbol:   method
getChecksumType(org.apache.hadoop.conf.Configuration)
[ERROR]   location: class org.apache.hadoop.hbase.regionserver.HStore
[ERROR]
/home/ubuntu/git/apache/phoenix/phoenix-hbase-compat-2.4.0/src/main/java/org/apache/phoenix/compat/hbase/CompatUtil.java:[140,22]
cannot find symbol
[ERROR]   symbol:   method
getBytesPerChecksum(org.apache.hadoop.conf.Configuration)
[ERROR]   location: class org.apache.hadoop.hbase.regionserver.HStore
[ERROR] -> [Help 1]

Then I ran
mvn verify -Dhbase.profile=2.4 -Dhbase.version=2.4.9-SNAPSHOT
-Dhbase24.compat.version=2.4.9-SNAPSHOT -pl "!phoenix-hbase-compat-2.4.0"

After 2 hours I've got one test failure:

[ERROR] Errors:
[ERROR]
PermissionNSEnabledIT>BasePermissionsIT.testSystemTablePermissions:1027->BasePermissionsIT.verifyAllowed:894->BasePermissionsIT.verifyAllowed:901
» UndeclaredThrowable
[INFO]
[ERROR] Tests run: 1550, Failures: 0, Errors: 1, Skipped: 60

and one IT test crashed the JVM:

[ERROR] org.apache.maven.surefire.booter.SurefireBooterForkException:
ExecutionException The forked VM terminated without properly saying
goodbye. VM crash or System.exit called?
[ERROR] Command was /bin/sh -c cd
/home/ubuntu/git/apache/phoenix/phoenix-core &&
/usr/lib/jvm/java-8-openjdk-arm64/jre/bin/java
-javaagent:/home/ubuntu/.m2/repository/org/jacoco/org.jacoco.agent/0.8.7/org.jacoco.agent-0.8.7-runtime.jar=destfile=/home/ubuntu/git/apache/phoenix/phoenix-core/target/jacoco.exec
-Xmx3000m -Djava.security.egd=file:/dev/./urandom
'-Djava.library.path=${hadoop.library.path}:/usr/java/packages/lib/aarch64:/usr/lib/aarch64-linux-gnu/jni:/lib/aarch64-linux-gnu:/usr/lib/aarch64-linux-gnu:/usr/lib/jni:/lib:/usr/lib'
-XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/ -XX:NewRatio=4
-XX:SurvivorRatio=8 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC
-XX:+DisableExplicitGC -XX:+UseCMSInitiatingOccupancyOnly
-XX:+CMSClassUnloadingEnabled -XX:+CMSScavengeBeforeRemark
-XX:CMSInitiatingOccupancyFraction=68 -jar
/home/ubuntu/git/apache/phoenix/phoenix-core/target/surefire/surefirebooter3776996098139745421.jar
/home/ubuntu/git/apache/phoenix/phoenix-core/target/surefire
2021-11-23T09-45-47_283-jvmRun7 surefire8003544437266408760tmp
surefire_135681795925087748715tmp
[ERROR] Error occurred in starting fork, check output in log
[ERROR] Process Exit Code: 137
[ERROR] ExecutionException The forked VM terminated without properly saying
goodbye. VM crash or System.exit called?
[ERROR] Command was /bin/sh -c cd
/home/ubuntu/git/apache/phoenix/phoenix-core &&
/usr/lib/jvm/java-8-openjdk-arm64/jre/bin/java
-javaagent:/home/ubuntu/.m2/repository/org/jacoco/org.jacoco.agent/0.8.7/org.jacoco.agent-0.8.7-runtime.jar=destfile=/home/ubuntu/git/apache/phoenix/phoenix-core/target/jacoco.exec
-Xmx3000m -Djava.security.egd=file:/dev/./urandom
'-Djava.library.path=${hadoop.library.path}:/usr/java/packages/lib/aarch64:/usr/lib/aarch64-linux-gnu/jni:/lib/aarch64-linux-gnu:/usr/lib/aarch64-linux-gnu:/usr/lib/jni:/lib:/usr/lib'
-XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/ -XX:NewRatio=4
-XX:SurvivorRatio=8 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC
-XX:+DisableExplicitGC -XX:+UseCMSInitiatingOccupancyOnly
-XX:+CMSClassUnloadingEnabled -XX:+CMSScavengeBeforeRemark
-XX:CMSInitiatingOccupancyFraction=68 -jar
/home/ubuntu/git/apache/phoenix/phoenix-core/target/surefire/surefirebooter8513242350930097363.jar
/home/ubuntu/git/apache/phoenix/phoenix-core/target/surefire
2021-11-23T09-45-47_283-jvmRun5 surefire3478876209177704794tmp
surefire_83451609179705425541tmp
[ERROR] Error occurred in starting fork, check output in log
[ERROR] Process Exit Code: 137
[ERROR] Crashed tests:
[ERROR] org.apache.phoenix.end2end.index.LocalImmutableNonTxIndexIT
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter.awaitResultsDone(ForkStarter.java:532)
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter.runSuitesForkOnceMultiple(ForkStarter.java:405)
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:321)
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:266)
[ERROR] at
org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1314)
[ERROR] at
org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1159)
[ERROR] at
org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:932)
[ERROR] at
org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)
[ERROR] at
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:210)
[ERROR] at
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:156)
[ERROR] at
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:148)
[ERROR] at
org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)
[ERROR] at
org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:81)
[ERROR] at
org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build(SingleThreadedBuilder.java:56)
[ERROR] at
org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:128)
[ERROR] at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:305)
[ERROR] at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:192)
[ERROR] at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:105)
[ERROR] at org.apache.maven.cli.MavenCli.execute(MavenCli.java:972)
[ERROR] at org.apache.maven.cli.MavenCli.doMain(MavenCli.java:293)
[ERROR] at org.apache.maven.cli.MavenCli.main(MavenCli.java:196)
[ERROR] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
[ERROR] at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
[ERROR] at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[ERROR] at java.lang.reflect.Method.invoke(Method.java:498)
[ERROR] at
org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced(Launcher.java:282)
[ERROR] at
org.codehaus.plexus.classworlds.launcher.Launcher.launch(Launcher.java:225)
[ERROR] at
org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode(Launcher.java:406)
[ERROR] at
org.codehaus.plexus.classworlds.launcher.Launcher.main(Launcher.java:347)
[ERROR] Caused by:
org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM
terminated without properly saying goodbye. VM crash or System.exit called?
[ERROR] Command was /bin/sh -c cd
/home/ubuntu/git/apache/phoenix/phoenix-core &&
/usr/lib/jvm/java-8-openjdk-arm64/jre/bin/java
-javaagent:/home/ubuntu/.m2/repository/org/jacoco/org.jacoco.agent/0.8.7/org.jacoco.agent-0.8.7-runtime.jar=destfile=/home/ubuntu/git/apache/phoenix/phoenix-core/target/jacoco.exec
-Xmx3000m -Djava.security.egd=file:/dev/./urandom
'-Djava.library.path=${hadoop.library.path}:/usr/java/packages/lib/aarch64:/usr/lib/aarch64-linux-gnu/jni:/lib/aarch64-linux-gnu:/usr/lib/aarch64-linux-gnu:/usr/lib/jni:/lib:/usr/lib'
-XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/ -XX:NewRatio=4
-XX:SurvivorRatio=8 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC
-XX:+DisableExplicitGC -XX:+UseCMSInitiatingOccupancyOnly
-XX:+CMSClassUnloadingEnabled -XX:+CMSScavengeBeforeRemark
-XX:CMSInitiatingOccupancyFraction=68 -jar
/home/ubuntu/git/apache/phoenix/phoenix-core/target/surefire/surefirebooter8513242350930097363.jar
/home/ubuntu/git/apache/phoenix/phoenix-core/target/surefire
2021-11-23T09-45-47_283-jvmRun5 surefire3478876209177704794tmp
surefire_83451609179705425541tmp
[ERROR] Error occurred in starting fork, check output in log

HBase project has merged my PR to branch-2 and branch-2.4. That's why I use
-2.4.9-SNAPSHOT now.


On Mon, 22 Nov 2021 at 07:59, Istvan Toth <[email protected]> wrote:

> The commands look good.
>
> However, for some reason the Phoenix build command doesn't pick up the
> rebuilt artifacts.
> You need to run your maven processes with debug enabled (-X), and check
> your local maven repo to see what goes wrong.
>
> On Fri, Nov 19, 2021 at 12:22 PM Mark Jens <[email protected]> wrote:
>
> > Hi Istvan,
> >
> > I've patched HBase 2.4.8 with the same changes as in
> > https://github.com/apache/hbase/pull/3860.
> > hbase-2.4.8$ mvn install -Dhadoop.profile=3.0 -DskipTests  passed
> > successfully!
> >
> > Following BUILDING.md I installed Phoenix with: mvn clean install
> > -Dhbase.profile=2.4 -Dhbase.version=2.4.8 -DskipTests
> > And finally tried to test it with: mvn verify -Dhbase.profile=2.4
> > -Dhbase.version=2.4.8 but again it failed with
> >
> > java.lang.IncompatibleClassChangeError: Found interface
> > org.apache.hadoop.hdfs.protocol.HdfsFileStatus, but class was expected
> > at
> > org.apache.hadoop.hbase.io
> >
> .asyncfs.FanOutOneBlockAsyncDFSOutputHelper.createOutput(FanOutOneBlockAsyncDFSOutputHelper.java:536)
> > at
> > org.apache.hadoop.hbase.io
> >
> .asyncfs.FanOutOneBlockAsyncDFSOutputHelper.access$400(FanOutOneBlockAsyncDFSOutputHelper.java:112)
> > at
> > org.apache.hadoop.hbase.io
> >
> .asyncfs.FanOutOneBlockAsyncDFSOutputHelper$8.doCall(FanOutOneBlockAsyncDFSOutputHelper.java:616)
> > at
> > org.apache.hadoop.hbase.io
> >
> .asyncfs.FanOutOneBlockAsyncDFSOutputHelper$8.doCall(FanOutOneBlockAsyncDFSOutputHelper.java:611)
> > at
> >
> >
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
> > at
> > org.apache.hadoop.hbase.io
> >
> .asyncfs.FanOutOneBlockAsyncDFSOutputHelper.createOutput(FanOutOneBlockAsyncDFSOutputHelper.java:624)
> > at
> > org.apache.hadoop.hbase.io
> > .asyncfs.AsyncFSOutputHelper.createOutput(AsyncFSOutputHelper.java:53)
> > at
> >
> >
> org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.initOutput(AsyncProtobufLogWriter.java:180)
> > at
> >
> >
> org.apache.hadoop.hbase.regionserver.wal.AbstractProtobufLogWriter.init(AbstractProtobufLogWriter.java:166)
> > at
> >
> >
> org.apache.hadoop.hbase.wal.AsyncFSWALProvider.createAsyncWriter(AsyncFSWALProvider.java:113)
> > at
> >
> >
> org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.createWriterInstance(AsyncFSWAL.java:669)
> > at
> >
> >
> org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.createWriterInstance(AsyncFSWAL.java:130)
> > at
> >
> >
> org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.rollWriter(AbstractFSWAL.java:841)
> > at
> >
> >
> org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.rollWriter(AbstractFSWAL.java:548)
> > at
> >
> >
> org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.init(AbstractFSWAL.java:489)
> > at
> >
> >
> org.apache.hadoop.hbase.wal.AbstractFSWALProvider.getWAL(AbstractFSWALProvider.java:160)
> > at
> >
> >
> org.apache.hadoop.hbase.wal.AbstractFSWALProvider.getWAL(AbstractFSWALProvider.java:62)
> > at org.apache.hadoop.hbase.wal.WALFactory.getWAL(WALFactory.java:296)
> > at
> >
> >
> org.apache.phoenix.hbase.index.write.recovery.TestPerRegionIndexWriteCache.setUp(TestPerRegionIndexWriteCache.java:109)
> > ...
> >
> > What I am doing wrong ?
> >
> > Thanks!
> >
> > Mark
> >
> >
> >
> > On Thu, 18 Nov 2021 at 11:16, Mark Jens <[email protected]> wrote:
> >
> > > Hi,
> > >
> > > On Thu, 18 Nov 2021 at 10:38, Istvan Toth <[email protected]>
> > > wrote:
> > >
> > >> Phoenix probably won't work with HBase 3.x, but looking at the linked
> > >> commit, it should be fairly straightforward to apply that to HBase
> 2.4.
> > >>
> > >> I'm not sure why that hasn't been backported to 2.x, perhaps HBase
> > doesn't
> > >> have test infra set up for ARM.
> > >> You may want to discuss backporting that change to the active Hbase
> 2.x
> > >> branches with the HBase project.
> > >>
> > >
> > > I've asked for a backport at
> > >
> >
> https://issues.apache.org/jira/browse/HBASE-23612?focusedCommentId=17445761&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-17445761
> > >
> > > Apologies for creating
> > https://issues.apache.org/jira/browse/PHOENIX-6595.
> > > Initially I thought that Phoenix can apply the patch since it already
> > > requires manual build of HBase.
> > > But you are right - it would be much better if the improvement is
> > > backported to HBase 2.x !
> > >
> > >
> > >>
> > >>
> > >> On Wed, Nov 17, 2021 at 12:35 PM Mark Jens <[email protected]>
> > wrote:
> > >>
> > >> > Hi Istvan,
> > >> >
> > >> > It seems HBase support for ARM64 is available only in 3.x:
> > >> >
> > >> >
> > >>
> >
> https://github.com/apache/hbase/commit/5480493f5f7b01b496f54215334543f2a82c6ba7
> > >> > Would Phoenix work with HBase 3.x ?
> > >> >
> > >> > On Wed, 17 Nov 2021 at 13:28, Mark Jens <[email protected]>
> > wrote:
> > >> >
> > >> > > Thanks for the hint!
> > >> > >
> > >> > > Unfortunately HBase-2.4.8 build fails with:
> > >> > >
> > >> > > INFO] BUILD FAILURE
> > >> > > [INFO]
> > >> > >
> > >>
> ------------------------------------------------------------------------
> > >> > > [INFO] Total time:  01:42 min
> > >> > > [INFO] Finished at: 2021-11-17T11:23:19Z
> > >> > > [INFO]
> > >> > >
> > >>
> ------------------------------------------------------------------------
> > >> > > [ERROR] Failed to execute goal
> > >> > > org.xolstice.maven.plugins:protobuf-maven-plugin:0.6.1:compile
> > >> > > (compile-protoc) on project hbase-protocol: Unable to resolve
> > >> artifact:
> > >> > > Missing:
> > >> > > [ERROR] ----------
> > >> > > [ERROR] 1) com.google.protobuf:protoc:exe:linux-aarch_64:2.5.0
> > >> > > [ERROR]
> > >> > > [ERROR]   Try downloading the file manually from the project
> > website.
> > >> > > [ERROR]
> > >> > > [ERROR]   Then, install it using the command:
> > >> > > [ERROR]       mvn install:install-file
> -DgroupId=com.google.protobuf
> > >> > > -DartifactId=protoc -Dversion=2.5.0 -Dclassifier=linux-aarch_64
> > >> > > -Dpackaging=exe -Dfile=/path/to/file
> > >> > > [ERROR]
> > >> > > [ERROR]   Alternatively, if you host your own repository you can
> > >> deploy
> > >> > > the file there:
> > >> > > [ERROR]       mvn deploy:deploy-file -DgroupId=com.google.protobuf
> > >> > > -DartifactId=protoc -Dversion=2.5.0 -Dclassifier=linux-aarch_64
> > >> > > -Dpackaging=exe -Dfile=/path/to/file -Durl=[url]
> -DrepositoryId=[id]
> > >> > > [ERROR]
> > >> > > [ERROR]   Path to dependency:
> > >> > > [ERROR]   1) org.apache.hbase:hbase-protocol:jar:2.4.8
> > >> > > [ERROR]   2) com.google.protobuf:protoc:exe:linux-aarch_64:2.5.0
> > >> > >
> > >> > >
> > >> > >
> > >> > >
> > >> > > On Wed, 17 Nov 2021 at 12:27, Istvan Toth
> > <[email protected]
> > >> >
> > >> > > wrote:
> > >> > >
> > >> > >> You need to recompile HBase.
> > >> > >> See BULIDING.md
> > >> > >>
> > >> > >> On Wed, Nov 17, 2021 at 10:33 AM Mark Jens <
> [email protected]>
> > >> > wrote:
> > >> > >>
> > >> > >> > Hello,
> > >> > >> >
> > >> > >> > I am trying to build Phoenix on Ubuntu 20.04.3 ARM64.
> > >> > >> >
> > >> > >> > Phoenix Core module fails with:
> > >> > >> >
> > >> > >> > [ERROR]
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.phoenix.hbase.index.write.recovery.TestPerRegionIndexWriteCache.testMultipleAddsForSingleRegion
> > >> > >> >  Time elapsed: 0.025 s  <<< ERROR!
> > >> > >> > java.lang.IncompatibleClassChangeError: Found interface
> > >> > >> > org.apache.hadoop.hdfs.protocol.HdfsFileStatus, but class was
> > >> expected
> > >> > >> > at
> > >> > >> > org.apache.hadoop.hbase.io
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> .asyncfs.FanOutOneBlockAsyncDFSOutputHelper.createOutput(FanOutOneBlockAsyncDFSOutputHelper.java:536)
> > >> > >> > at
> > >> > >> > org.apache.hadoop.hbase.io
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> .asyncfs.FanOutOneBlockAsyncDFSOutputHelper.access$400(FanOutOneBlockAsyncDFSOutputHelper.java:112)
> > >> > >> > at
> > >> > >> > org.apache.hadoop.hbase.io
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> .asyncfs.FanOutOneBlockAsyncDFSOutputHelper$8.doCall(FanOutOneBlockAsyncDFSOutputHelper.java:616)
> > >> > >> > at
> > >> > >> > org.apache.hadoop.hbase.io
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> .asyncfs.FanOutOneBlockAsyncDFSOutputHelper$8.doCall(FanOutOneBlockAsyncDFSOutputHelper.java:611)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
> > >> > >> > at
> > >> > >> > org.apache.hadoop.hbase.io
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> .asyncfs.FanOutOneBlockAsyncDFSOutputHelper.createOutput(FanOutOneBlockAsyncDFSOutputHelper.java:624)
> > >> > >> > at
> > >> > >> > org.apache.hadoop.hbase.io
> > >> > >> >
> > >> .asyncfs.AsyncFSOutputHelper.createOutput(AsyncFSOutputHelper.java:53)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.initOutput(AsyncProtobufLogWriter.java:180)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.hadoop.hbase.regionserver.wal.AbstractProtobufLogWriter.init(AbstractProtobufLogWriter.java:166)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.hadoop.hbase.wal.AsyncFSWALProvider.createAsyncWriter(AsyncFSWALProvider.java:113)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.createWriterInstance(AsyncFSWAL.java:669)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.createWriterInstance(AsyncFSWAL.java:130)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.rollWriter(AbstractFSWAL.java:841)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.rollWriter(AbstractFSWAL.java:548)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.init(AbstractFSWAL.java:489)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.hadoop.hbase.wal.AbstractFSWALProvider.getWAL(AbstractFSWALProvider.java:160)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.hadoop.hbase.wal.AbstractFSWALProvider.getWAL(AbstractFSWALProvider.java:62)
> > >> > >> > at
> > >> org.apache.hadoop.hbase.wal.WALFactory.getWAL(WALFactory.java:296)
> > >> > >> > at
> > >> > >> >
> > >> > >> >
> > >> > >>
> > >> >
> > >>
> >
> org.apache.phoenix.hbase.index.write.recovery.TestPerRegionIndexWriteCache.setUp(TestPerRegionIndexWriteCache.java:109)
> > >> > >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > >> > >> > ....
> > >> > >> > [INFO]
> > >> > >> > [INFO] Results:
> > >> > >> > [INFO]
> > >> > >> > [ERROR] Errors:
> > >> > >> > [ERROR]   TestPerRegionIndexWriteCache.setUp:109 »
> > >> > >> IncompatibleClassChange
> > >> > >> > Found interfa...
> > >> > >> > [ERROR]   TestPerRegionIndexWriteCache.setUp:109 »
> > >> > >> IncompatibleClassChange
> > >> > >> > Found interfa...
> > >> > >> > [ERROR]   TestPerRegionIndexWriteCache.setUp:109 »
> > >> > >> IncompatibleClassChange
> > >> > >> > Found interfa...
> > >> > >> > [INFO]
> > >> > >> > [ERROR] Tests run: 1909, Failures: 0, Errors: 3, Skipped: 6
> > >> > >> > [INFO]
> > >> > >> > [INFO]
> > >> > >> >
> > >> >
> > ------------------------------------------------------------------------
> > >> > >> > [INFO] Reactor Summary for Apache Phoenix 5.2.0-SNAPSHOT:
> > >> > >> > [INFO]
> > >> > >> > [INFO] Apache Phoenix .....................................
> > >> SUCCESS [
> > >> > >> >  2.034 s]
> > >> > >> > [INFO] Phoenix Hbase 2.4.1 compatibility ..................
> > >> SUCCESS [
> > >> > >> >  5.358 s]
> > >> > >> > [INFO] Phoenix Hbase 2.4.0 compatibility ..................
> > >> SUCCESS [
> > >> > >> >  3.946 s]
> > >> > >> > [INFO] Phoenix Hbase 2.3.0 compatibility ..................
> > >> SUCCESS [
> > >> > >> >  4.437 s]
> > >> > >> > [INFO] Phoenix Hbase 2.2.5 compatibility ..................
> > >> SUCCESS [
> > >> > >> >  4.004 s]
> > >> > >> > [INFO] Phoenix Hbase 2.1.6 compatibility ..................
> > >> SUCCESS [
> > >> > >> >  3.966 s]
> > >> > >> > [INFO] Phoenix Core .......................................
> > FAILURE
> > >> > >> [01:25
> > >> > >> > min]
> > >> > >> > [INFO] Phoenix - Pherf ....................................
> > SKIPPED
> > >> > >> > ...
> > >> > >> >
> > >> > >> > Any idea why this breaks ?
> > >> > >> >
> > >> > >> > It does not look ARM64 specific to me. I will try on x64 too.
> > >> > >> >
> > >> > >> > Thanks!
> > >> > >> >
> > >> > >> > Mark
> > >> > >> >
> > >> > >>
> > >> > >>
> > >> > >> --
> > >> > >> *István Tóth* | Staff Software Engineer
> > >> > >> [email protected] <https://www.cloudera.com>
> > >> > >> [image: Cloudera] <https://www.cloudera.com/>
> > >> > >> [image: Cloudera on Twitter] <https://twitter.com/cloudera>
> > [image:
> > >> > >> Cloudera on Facebook] <https://www.facebook.com/cloudera>
> [image:
> > >> > >> Cloudera
> > >> > >> on LinkedIn] <https://www.linkedin.com/company/cloudera>
> > >> > >> <https://www.cloudera.com/>
> > >> > >> ------------------------------
> > >> > >>
> > >> > >
> > >> >
> > >>
> > >>
> > >> --
> > >> *István Tóth* | Staff Software Engineer
> > >> [email protected] <https://www.cloudera.com>
> > >> [image: Cloudera] <https://www.cloudera.com/>
> > >> [image: Cloudera on Twitter] <https://twitter.com/cloudera> [image:
> > >> Cloudera on Facebook] <https://www.facebook.com/cloudera> [image:
> > >> Cloudera
> > >> on LinkedIn] <https://www.linkedin.com/company/cloudera>
> > >> <https://www.cloudera.com/>
> > >> ------------------------------
> > >>
> > >
> >
>

Reply via email to