See
<https://builds.apache.org/job/Phoenix-master-matrix/HBASE_PROFILE=2.1/13/display/redirect?page=changes>
Changes:
[kadir] PHOENIX-5743 addendum for multi-column family indexes
------------------------------------------
[...truncated 572.03 KB...]
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.008 s
<<< FAILURE! - in org.apache.phoenix.end2end.ViewMetadataIT
[ERROR] org.apache.phoenix.end2end.ViewMetadataIT Time elapsed: 0.007 s <<<
ERROR!
java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to create native
thread: possibly out of memory or process/resource limits reached
at
org.apache.phoenix.end2end.ViewMetadataIT.doSetup(ViewMetadataIT.java:98)
Caused by: java.lang.OutOfMemoryError: unable to create native thread: possibly
out of memory or process/resource limits reached
at
org.apache.phoenix.end2end.ViewMetadataIT.doSetup(ViewMetadataIT.java:98)
[INFO] Running org.apache.phoenix.end2end.AlterTableWithViewsIT
[INFO] Running org.apache.phoenix.end2end.DropIndexedColsIT
[ERROR] Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 0.006 s
<<< FAILURE! - in org.apache.phoenix.end2end.DropIndexedColsIT
[ERROR] org.apache.phoenix.end2end.DropIndexedColsIT Time elapsed: 0.005 s
<<< FAILURE!
java.lang.AssertionError: Multiple regions on
asf927.gq1.ygridcore.net,37235,1582446465756
[INFO] Running org.apache.phoenix.end2end.DropTableWithViewsIT
[INFO] Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.065 s
- in org.apache.phoenix.end2end.DropTableWithViewsIT
[ERROR] Tests run: 9, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 107.331
s <<< FAILURE! - in org.apache.phoenix.end2end.AlterMultiTenantTableWithViewsIT
[ERROR]
testAddPKColumnToBaseTableWhoseViewsHaveIndices(org.apache.phoenix.end2end.AlterMultiTenantTableWithViewsIT)
Time elapsed: 2.562 s <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException:
org.apache.hadoop.hbase.DoNotRetryIOException: SCHEMA1.N000001:
java.lang.OutOfMemoryError: unable to create native thread: possibly out of
memory or process/resource limits reached
at
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:113)
at
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:2126)
at
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17218)
at
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8265)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2444)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2426)
at
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42286)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:133)
at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324)
at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304)
Caused by: java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to
create native thread: possibly out of memory or process/resource limits reached
at
org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithoutRetries(RpcRetryingCallerImpl.java:200)
at
org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:267)
at
org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:435)
at
org.apache.hadoop.hbase.client.ClientScanner.nextWithSyncCache(ClientScanner.java:310)
at
org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:595)
at org.apache.phoenix.util.ViewUtil.findRelatedViews(ViewUtil.java:127)
at org.apache.phoenix.util.ViewUtil.dropChildViews(ViewUtil.java:200)
at
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1767)
... 9 more
Caused by: java.lang.OutOfMemoryError: unable to create native thread: possibly
out of memory or process/resource limits reached
at java.base/java.lang.Thread.start0(Native Method)
at java.base/java.lang.Thread.start(Thread.java:803)
at
java.base/java.util.concurrent.ThreadPoolExecutor.addWorker(ThreadPoolExecutor.java:937)
at
java.base/java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1343)
at
org.apache.hadoop.hbase.client.ResultBoundedCompletionService.submit(ResultBoundedCompletionService.java:171)
at
org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.addCallsForCurrentReplica(ScannerCallableWithReplicas.java:329)
at
org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:191)
at
org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:58)
at
org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithoutRetries(RpcRetryingCallerImpl.java:192)
... 16 more
at
org.apache.phoenix.end2end.AlterMultiTenantTableWithViewsIT.testAddPKColumnToBaseTableWhoseViewsHaveIndices(AlterMultiTenantTableWithViewsIT.java:295)
Caused by: org.apache.hadoop.hbase.DoNotRetryIOException:
org.apache.hadoop.hbase.DoNotRetryIOException: SCHEMA1.N000001:
java.lang.OutOfMemoryError: unable to create native thread: possibly out of
memory or process/resource limits reached
at
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:113)
at
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:2126)
at
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17218)
at
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8265)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2444)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2426)
at
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42286)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:133)
at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324)
at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304)
Caused by: java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to
create native thread: possibly out of memory or process/resource limits reached
at
org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithoutRetries(RpcRetryingCallerImpl.java:200)
at
org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:267)
at
org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:435)
at
org.apache.hadoop.hbase.client.ClientScanner.nextWithSyncCache(ClientScanner.java:310)
at
org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:595)
at org.apache.phoenix.util.ViewUtil.findRelatedViews(ViewUtil.java:127)
at org.apache.phoenix.util.ViewUtil.dropChildViews(ViewUtil.java:200)
at
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1767)
... 9 more
Caused by: java.lang.OutOfMemoryError: unable to create native thread: possibly
out of memory or process/resource limits reached
at java.base/java.lang.Thread.start0(Native Method)
at java.base/java.lang.Thread.start(Thread.java:803)
at
java.base/java.util.concurrent.ThreadPoolExecutor.addWorker(ThreadPoolExecutor.java:937)
at
java.base/java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1343)
at
org.apache.hadoop.hbase.client.ResultBoundedCompletionService.submit(ResultBoundedCompletionService.java:171)
at
org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.addCallsForCurrentReplica(ScannerCallableWithReplicas.java:329)
at
org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:191)
at
org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:58)
at
org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithoutRetries(RpcRetryingCallerImpl.java:192)
... 16 more
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException:
org.apache.hadoop.hbase.DoNotRetryIOException: SCHEMA1.N000001:
java.lang.OutOfMemoryError: unable to create native thread: possibly out of
memory or process/resource limits reached
at
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:113)
at
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:2126)
at
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17218)
at
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8265)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2444)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2426)
at
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42286)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:133)
at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324)
at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304)
Caused by: java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to
create native thread: possibly out of memory or process/resource limits reached
at
org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithoutRetries(RpcRetryingCallerImpl.java:200)
at
org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:267)
at
org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:435)
at
org.apache.hadoop.hbase.client.ClientScanner.nextWithSyncCache(ClientScanner.java:310)
at
org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:595)
at org.apache.phoenix.util.ViewUtil.findRelatedViews(ViewUtil.java:127)
at org.apache.phoenix.util.ViewUtil.dropChildViews(ViewUtil.java:200)
at
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1767)
... 9 more
Caused by: java.lang.OutOfMemoryError: unable to create native thread: possibly
out of memory or process/resource limits reached
at java.base/java.lang.Thread.start0(Native Method)
at java.base/java.lang.Thread.start(Thread.java:803)
at
java.base/java.util.concurrent.ThreadPoolExecutor.addWorker(ThreadPoolExecutor.java:937)
at
java.base/java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1343)
at
org.apache.hadoop.hbase.client.ResultBoundedCompletionService.submit(ResultBoundedCompletionService.java:171)
at
org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.addCallsForCurrentReplica(ScannerCallableWithReplicas.java:329)
at
org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:191)
at
org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:58)
at
org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithoutRetries(RpcRetryingCallerImpl.java:192)
... 16 more
[INFO] Running org.apache.phoenix.end2end.TenantSpecificViewIndexSaltedIT
[INFO] Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.206 s
- in org.apache.phoenix.end2end.TenantSpecificViewIndexSaltedIT
[INFO] Running org.apache.phoenix.end2end.index.ViewIndexIT
[WARNING] Tests run: 24, Failures: 0, Errors: 0, Skipped: 2, Time elapsed:
93.129 s - in org.apache.phoenix.end2end.index.ViewIndexIT
[ERROR] Tests run: 56, Failures: 0, Errors: 4, Skipped: 0, Time elapsed:
5,012.348 s <<< FAILURE! - in org.apache.phoenix.end2end.AlterTableWithViewsIT
[ERROR]
testDroppingIndexedColDropsViewIndex[AlterTableWithViewsIT_columnEncoded=false,
multiTenant=false,
salted=false](org.apache.phoenix.end2end.AlterTableWithViewsIT) Time elapsed:
1,203.425 s <<< ERROR!
org.apache.phoenix.execute.CommitException:
org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1
action: Operation rpcTimeout: 1 time, servers with issues:
asf927.gq1.ygridcore.net,40149,1582446452481
at
org.apache.phoenix.end2end.AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex(AlterTableWithViewsIT.java:1151)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException:
Failed 1 action: Operation rpcTimeout: 1 time, servers with issues:
asf927.gq1.ygridcore.net,40149,1582446452481
at
org.apache.phoenix.end2end.AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex(AlterTableWithViewsIT.java:1151)
[ERROR]
testDroppingIndexedColDropsViewIndex[AlterTableWithViewsIT_columnEncoded=true,
multiTenant=false,
salted=true](org.apache.phoenix.end2end.AlterTableWithViewsIT) Time elapsed:
1,202.922 s <<< ERROR!
org.apache.phoenix.execute.CommitException:
org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1
action: Operation rpcTimeout: 1 time, servers with issues:
asf927.gq1.ygridcore.net,36725,1582446452114
at
org.apache.phoenix.end2end.AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex(AlterTableWithViewsIT.java:1151)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException:
Failed 1 action: Operation rpcTimeout: 1 time, servers with issues:
asf927.gq1.ygridcore.net,36725,1582446452114
at
org.apache.phoenix.end2end.AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex(AlterTableWithViewsIT.java:1151)
[ERROR]
testDroppingIndexedColDropsViewIndex[AlterTableWithViewsIT_columnEncoded=true,
multiTenant=true,
salted=false](org.apache.phoenix.end2end.AlterTableWithViewsIT) Time elapsed:
1,202.843 s <<< ERROR!
org.apache.phoenix.execute.CommitException:
org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1
action: Operation rpcTimeout: 1 time, servers with issues:
asf927.gq1.ygridcore.net,38417,1582446451877
at
org.apache.phoenix.end2end.AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex(AlterTableWithViewsIT.java:1151)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException:
Failed 1 action: Operation rpcTimeout: 1 time, servers with issues:
asf927.gq1.ygridcore.net,38417,1582446451877
at
org.apache.phoenix.end2end.AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex(AlterTableWithViewsIT.java:1151)
[ERROR]
testDroppingIndexedColDropsViewIndex[AlterTableWithViewsIT_columnEncoded=true,
multiTenant=true,
salted=true](org.apache.phoenix.end2end.AlterTableWithViewsIT) Time elapsed:
1,202.921 s <<< ERROR!
org.apache.phoenix.execute.CommitException:
org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1
action: Operation rpcTimeout: 1 time, servers with issues:
asf927.gq1.ygridcore.net,38417,1582446451877
at
org.apache.phoenix.end2end.AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex(AlterTableWithViewsIT.java:1151)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException:
Failed 1 action: Operation rpcTimeout: 1 time, servers with issues:
asf927.gq1.ygridcore.net,38417,1582446451877
at
org.apache.phoenix.end2end.AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex(AlterTableWithViewsIT.java:1151)
[INFO]
[INFO] Results:
[INFO]
[ERROR] Failures:
[ERROR]
DropIndexedColsIT>SplitSystemCatalogIT.doSetup:60->SplitSystemCatalogIT.splitSystemCatalog:77->BaseTest.splitSystemCatalog:1923->BaseTest.splitTable:1894
Multiple regions on asf927.gq1.ygridcore.net,37235,1582446465756
[ERROR] Errors:
[ERROR]
AlterMultiTenantTableWithViewsIT.testAddPKColumnToBaseTableWhoseViewsHaveIndices:295
» PhoenixIO
[ERROR] AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex:1151 »
Commit org.a...
[ERROR] AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex:1151 »
Commit org.a...
[ERROR] AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex:1151 »
Commit org.a...
[ERROR] AlterTableWithViewsIT.testDroppingIndexedColDropsViewIndex:1151 »
Commit org.a...
[ERROR]
TenantSpecificViewIndexIT>SplitSystemCatalogIT.doSetup:57->BaseTest.setUpTestDriver:515->BaseTest.setUpTestDriver:521->BaseTest.initAndRegisterTestDriver:660
» PhoenixIO
[ERROR]
ViewIT.doSetup:142->BaseTest.setUpTestDriver:520->BaseTest.checkClusterInitialized:434->BaseTest.setUpTestCluster:448->BaseTest.initMiniCluster:549
» Runtime
[ERROR]
ViewMetadataIT.doSetup:98->BaseTest.setUpTestDriver:520->BaseTest.checkClusterInitialized:434->BaseTest.setUpTestCluster:448->BaseTest.initMiniCluster:549
» Runtime
[INFO]
[ERROR] Tests run: 101, Failures: 1, Errors: 8, Skipped: 2
[INFO]
[INFO]
[INFO] --- maven-failsafe-plugin:2.22.0:verify (ParallelStatsEnabledTest) @
phoenix-core ---
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary for Apache Phoenix 5.1.0-SNAPSHOT:
[INFO]
[INFO] Phoenix Hbase 2.2.1 compatibility .................. SUCCESS [ 17.821 s]
[INFO] Phoenix Hbase 2.1.6 compatibility .................. SUCCESS [ 7.542 s]
[INFO] Phoenix Hbase 2.0.1 compatibility .................. SUCCESS [ 8.024 s]
[INFO] Apache Phoenix ..................................... SUCCESS [ 1.255 s]
[INFO] Phoenix Core ....................................... FAILURE [ 03:08 h]
[INFO] Phoenix - Pherf .................................... SKIPPED
[INFO] Phoenix Client ..................................... SKIPPED
[INFO] Phoenix Server ..................................... SKIPPED
[INFO] Phoenix Assembly ................................... SKIPPED
[INFO] Phoenix - Tracing Web Application .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 03:08 h
[INFO] Finished at: 2020-02-23T09:53:28Z
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal
org.apache.maven.plugins:maven-failsafe-plugin:2.22.0:verify
(ParallelStatsEnabledTest) on project phoenix-core: There are test failures.
[ERROR]
[ERROR] Please refer to
<https://builds.apache.org/job/Phoenix-master-matrix/HBASE_PROFILE=2.1/ws/phoenix-core/target/failsafe-reports>
for the individual test results.
[ERROR] Please refer to dump files (if any exist) [date]-jvmRun[N].dump,
[date].dumpstream and [date]-jvmRun[N].dumpstream.
[ERROR] org.apache.maven.surefire.booter.SurefireBooterForkException:
ExecutionException Error occurred in starting fork, check output in log
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter.awaitResultsDone(ForkStarter.java:494)
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter.runSuitesForkPerTestSet(ForkStarter.java:441)
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:293)
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:245)
[ERROR] at
org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1194)
[ERROR] at
org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1022)
[ERROR] at
org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:868)
[ERROR] at
org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)
[ERROR] at
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:210)
[ERROR] at
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:156)
[ERROR] at
org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:148)
[ERROR] at
org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)
[ERROR] at
org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:81)
[ERROR] at
org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build(SingleThreadedBuilder.java:56)
[ERROR] at
org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:128)
[ERROR] at
org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:305)
[ERROR] at
org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:192)
[ERROR] at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:105)
[ERROR] at org.apache.maven.cli.MavenCli.execute(MavenCli.java:957)
[ERROR] at org.apache.maven.cli.MavenCli.doMain(MavenCli.java:289)
[ERROR] at org.apache.maven.cli.MavenCli.main(MavenCli.java:193)
[ERROR] at
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
[ERROR] at
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
[ERROR] at
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[ERROR] at java.base/java.lang.reflect.Method.invoke(Method.java:566)
[ERROR] at
org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced(Launcher.java:282)
[ERROR] at
org.codehaus.plexus.classworlds.launcher.Launcher.launch(Launcher.java:225)
[ERROR] at
org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode(Launcher.java:406)
[ERROR] at
org.codehaus.plexus.classworlds.launcher.Launcher.main(Launcher.java:347)
[ERROR] Caused by:
org.apache.maven.surefire.booter.SurefireBooterForkException: Error occurred in
starting fork, check output in log
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:624)
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:533)
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter.access$600(ForkStarter.java:115)
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter$2.call(ForkStarter.java:429)
[ERROR] at
org.apache.maven.plugin.surefire.booterclient.ForkStarter$2.call(ForkStarter.java:406)
[ERROR] at
java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
[ERROR] at
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
[ERROR] at
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
[ERROR] at java.base/java.lang.Thread.run(Thread.java:834)
[ERROR]
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please
read the following articles:
[ERROR] [Help 1]
http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <args> -rf :phoenix-core
Build step 'Invoke top-level Maven targets' marked build as failure
Archiving artifacts
Recording test results