Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Solaris/6/
Java: multiarch/jdk1.7.0 -d32 -server -XX:+UseParallelGC
2 tests failed.
FAILED: junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsNNFailoverTest
Error Message:
Error while running command to get file permissions : java.io.IOException:
Cannot run program "/bin/ls": error=12, Not enough space at
java.lang.ProcessBuilder.start(ProcessBuilder.java:1047) at
org.apache.hadoop.util.Shell.runCommand(Shell.java:485) at
org.apache.hadoop.util.Shell.run(Shell.java:455) at
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715) at
org.apache.hadoop.util.Shell.execCommand(Shell.java:808) at
org.apache.hadoop.util.Shell.execCommand(Shell.java:791) at
org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097) at
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:582)
at
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:557)
at
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:139)
at org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156) at
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
at
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738) at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608) at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98) at
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606) at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at java.lang.Thread.run(Thread.java:745) Caused by: java.io.IOException:
error=12, Not enough space at java.lang.UNIXProcess.forkAndExec(Native Method)
at java.lang.UNIXProcess.<init>(UNIXProcess.java:137) at
java.lang.ProcessImpl.start(ProcessImpl.java:130) at
java.lang.ProcessBuilder.start(ProcessBuilder.java:1028) ... 44 more
Stack Trace:
java.lang.RuntimeException: Error while running command to get file permissions
: java.io.IOException: Cannot run program "/bin/ls": error=12, Not enough space
at java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
at org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
at org.apache.hadoop.util.Shell.run(Shell.java:455)
at
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)
at org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
at org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)
at
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:582)
at
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:557)
at
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:139)
at org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
at
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
at
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
at
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.IOException: error=12, Not enough space
at java.lang.UNIXProcess.forkAndExec(Native Method)
at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
at java.lang.ProcessImpl.start(ProcessImpl.java:130)
at java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
... 44 more
at __randomizedtesting.SeedInfo.seed([5198E59CD4979997]:0)
at
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:620)
at
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:557)
at
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:139)
at org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
at
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
at
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
at
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at java.lang.Thread.run(Thread.java:745)
FAILED:
junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest
Error Message:
Timed out waiting for Mini HDFS Cluster to start
Stack Trace:
java.io.IOException: Timed out waiting for Mini HDFS Cluster to start
at __randomizedtesting.SeedInfo.seed([5198E59CD4979997]:0)
at
org.apache.hadoop.hdfs.MiniDFSCluster.waitClusterUp(MiniDFSCluster.java:1197)
at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:832)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:64)
at
org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest.beforeClass(HdfsRecoverLeaseTest.java:51)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at java.lang.Thread.run(Thread.java:745)
Build Log:
[...truncated 10931 lines...]
[junit4] Suite: org.apache.solr.cloud.hdfs.HdfsNNFailoverTest
[junit4] 2> Creating dataDir:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5198E59CD4979997-001/init-core-data-001
[junit4] 2> 2524562 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /
[junit4] 1> Formatting using clusterid: testClusterID
[junit4] 2> 2524679 WARN
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[junit4] 2> 2524695 WARN
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 2524699 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
jetty-6.1.26
[junit4] 2> 2524723 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Extract
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/hdfs
to ./temp/Jetty_solaris.vm_53593_hdfs____.avc5a5/webapp
[junit4] 2> 2524894 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log NO JSP
Support for /, did not find org.apache.jasper.servlet.JspServlet
[junit4] 2> 2525793 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Started HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:53593
[junit4] 2> 2540413 WARN
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 2540422 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
jetty-6.1.26
[junit4] 2> 2540450 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Extract
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/datanode
to ./temp/Jetty_solaris.vm_43441_datanode____oy6l5t/webapp
[junit4] 2> 2540656 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log NO JSP
Support for /, did not find org.apache.jasper.servlet.JspServlet
[junit4] 2> 2541435 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Started HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:43441
[junit4] 2> 2546731 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
[junit4] 2> 2546866 ERROR (DataNode:
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data1/,
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data2/]]
heartbeating to solaris-vm/127.0.0.1:43438) [ ] o.a.h.h.s.d.DataNode
Initialization failed for Block pool <registering> (Datanode Uuid unassigned)
service to solaris-vm/127.0.0.1:43438. Exiting.
[junit4] 2> java.nio.channels.ClosedByInterruptException
[junit4] 2> at
java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
[junit4] 2> at
sun.nio.ch.FileChannelImpl.position(FileChannelImpl.java:266)
[junit4] 2> at
org.apache.hadoop.hdfs.server.common.Storage.writeProperties(Storage.java:981)
[junit4] 2> at
org.apache.hadoop.hdfs.server.common.Storage.writeProperties(Storage.java:958)
[junit4] 2> at
org.apache.hadoop.hdfs.server.common.Storage.writeProperties(Storage.java:952)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataStorage.writeAll(DataStorage.java:185)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataStorage.addStorageLocations(DataStorage.java:335)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:403)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:422)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1311)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1276)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:314)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:220)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:828)
[junit4] 2> at java.lang.Thread.run(Thread.java:745)
[junit4] 2> 2546870 WARN (DataNode:
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data1/,
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data2/]]
heartbeating to solaris-vm/127.0.0.1:43438) [ ] o.a.h.h.s.d.DataNode
Ending block pool service for: Block pool <registering> (Datanode Uuid
unassigned) service to solaris-vm/127.0.0.1:43438
[junit4] 2> 2546905 WARN
(org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager$Monitor@43f0c)
[ ] o.a.h.h.s.b.DecommissionManager Monitor interrupted:
java.lang.InterruptedException: sleep interrupted
[junit4] 2> 2546950 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
[junit4] 2> 2547070 ERROR
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.m.l.MethodMetric Error invoking method getBlocksTotal
[junit4] 2> java.lang.reflect.InvocationTargetException
[junit4] 2> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
Method)
[junit4] 2> at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
[junit4] 2> at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[junit4] 2> at java.lang.reflect.Method.invoke(Method.java:606)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.MethodMetric$2.snapshot(MethodMetric.java:111)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.MethodMetric.snapshot(MethodMetric.java:144)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.MetricsRegistry.snapshot(MetricsRegistry.java:387)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.MetricsSourceBuilder$1.getMetrics(MetricsSourceBuilder.java:79)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMetrics(MetricsSourceAdapter.java:195)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.updateJmxCache(MetricsSourceAdapter.java:172)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMBeanInfo(MetricsSourceAdapter.java:151)
[junit4] 2> at
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getClassName(DefaultMBeanServerInterceptor.java:1804)
[junit4] 2> at
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.safeGetClassName(DefaultMBeanServerInterceptor.java:1595)
[junit4] 2> at
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.checkMBeanPermission(DefaultMBeanServerInterceptor.java:1813)
[junit4] 2> at
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.exclusiveUnregisterMBean(DefaultMBeanServerInterceptor.java:430)
[junit4] 2> at
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.unregisterMBean(DefaultMBeanServerInterceptor.java:415)
[junit4] 2> at
com.sun.jmx.mbeanserver.JmxMBeanServer.unregisterMBean(JmxMBeanServer.java:546)
[junit4] 2> at
org.apache.hadoop.metrics2.util.MBeans.unregister(MBeans.java:81)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stopMBeans(MetricsSourceAdapter.java:227)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stop(MetricsSourceAdapter.java:212)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stopSources(MetricsSystemImpl.java:461)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stop(MetricsSystemImpl.java:212)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.shutdown(MetricsSystemImpl.java:592)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdownInstance(DefaultMetricsSystem.java:72)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdown(DefaultMetricsSystem.java:68)
[junit4] 2> at
org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics.shutdown(NameNodeMetrics.java:145)
[junit4] 2> at
org.apache.hadoop.hdfs.server.namenode.NameNode.stop(NameNode.java:822)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1720)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1699)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:838)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
[junit4] 2> at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
[junit4] 2> at
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
[junit4] 2> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
Method)
[junit4] 2> at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
[junit4] 2> at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[junit4] 2> at java.lang.reflect.Method.invoke(Method.java:606)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
[junit4] 2> at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
[junit4] 2> at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
[junit4] 2> at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
[junit4] 2> at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
[junit4] 2> at java.lang.Thread.run(Thread.java:745)
[junit4] 2> Caused by: java.lang.NullPointerException
[junit4] 2> at
org.apache.hadoop.hdfs.server.blockmanagement.BlocksMap.size(BlocksMap.java:198)
[junit4] 2> at
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.getTotalBlocks(BlockManager.java:3291)
[junit4] 2> at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlocksTotal(FSNamesystem.java:6223)
[junit4] 2> ... 58 more
[junit4] 2> 2547097 INFO
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.s.SolrTestCaseJ4 ###deleteCore
[junit4] 2> Aug 28, 2015 8:53:35 PM
com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
[junit4] 2> WARNING: Will linger awaiting termination of 1 leaked
thread(s).
[junit4] 2> NOTE: test params are: codec=Asserting(Lucene53): {},
docValues:{}, sim=RandomSimilarityProvider(queryNorm=true,coord=crazy): {},
locale=ga, timezone=America/Whitehorse
[junit4] 2> NOTE: SunOS 5.11 x86/Oracle Corporation 1.7.0_85
(32-bit)/cpus=3,threads=1,free=242103096,total=507248640
[junit4] 2> NOTE: All tests run in this JVM: [TestFoldingMultitermQuery,
SuggesterTSTTest, DistributedFacetPivotLargeTest, HdfsThreadLeakTest, TestJoin,
TestCloudPivotFacet, BinaryUpdateRequestHandlerTest,
TestDefaultSearchFieldResource, SpellPossibilityIteratorTest, TestRecovery,
TestFileDictionaryLookup, DeleteShardTest, CoreMergeIndexesAdminHandlerTest,
TestConfigOverlay, DistributedDebugComponentTest, TestShardHandlerFactory,
TestHighFrequencyDictionaryFactory, TermVectorComponentDistributedTest,
OpenCloseCoreStressTest, IgnoreCommitOptimizeUpdateProcessorFactoryTest,
TestInitParams, TestStressRecovery, TestFreeTextSuggestions,
TestSolrCoreProperties, AddBlockUpdateTest, TestSolrCLIRunExample,
TestJettySolrRunner, CustomCollectionTest, CloudMLTQParserTest,
TestSolrQueryParser, DirectSolrConnectionTest,
SharedFSAutoReplicaFailoverUtilsTest, DistributedVersionInfoTest,
DateMathParserTest, BasicFunctionalityTest, AnalyticsMergeStrategyTest,
CSVRequestHandlerTest, AddSchemaFieldsUpdateProcessorFactoryTest,
CollectionsAPIDistributedZkTest, DistributedSuggestComponentTest,
SparseHLLTest, FileBasedSpellCheckerTest, RollingRestartTest,
TestDynamicFieldResource, DistribDocExpirationUpdateProcessorTest,
TestCollationField, HighlighterMaxOffsetTest, HttpPartitionTest,
LukeRequestHandlerTest, ResponseHeaderTest, RemoteQueryErrorTest, TestTrie,
NotRequiredUniqueKeyTest, TestRequestStatusCollectionAPI,
TestSha256AuthenticationProvider, ConnectionManagerTest,
PKIAuthenticationIntegrationTest, TestCoreDiscovery,
TestBlendedInfixSuggestions, OpenExchangeRatesOrgProviderTest,
SolrCoreCheckLockOnStartupTest, TestSolrXml, PreAnalyzedFieldTest,
StatsComponentTest, TestRandomFaceting, TestManagedResourceStorage,
PrimUtilsTest, TestScoreJoinQPScore, TestRangeQuery, ShardSplitTest,
SliceStateTest, ConcurrentDeleteAndCreateCollectionTest,
SpellingQueryConverterTest, TestScoreJoinQPNoScore, VersionInfoTest,
CollectionTooManyReplicasTest, EnumFieldTest,
VMParamsZkACLAndCredentialsProvidersTest, TestConfigSetsAPIZkFailure,
TestClusterStateMutator, DistributedExpandComponentTest, TestRecoveryHdfs,
SuggesterWFSTTest, DistributedFacetPivotLongTailTest, TestCustomSort,
TestFuzzyAnalyzedSuggestions, TestBadConfig, TestManagedResource,
TestSolrDynamicMBean, CoreAdminCreateDiscoverTest, ConvertedLegacyTest,
TestManagedSchemaFieldTypeResource, WordBreakSolrSpellCheckerTest,
TestPerFieldSimilarity, NoCacheHeaderTest, QueryParsingTest, BitVectorTest,
DocValuesMissingTest, TestOmitPositions, TestJsonFacets, TestUtils,
UniqFieldsUpdateProcessorFactoryTest, ClusterStateTest, SpellCheckCollatorTest,
TestBlobHandler, DistributedQueueTest, EchoParamsTest,
LeaderElectionIntegrationTest, TestOrdValues, TestShortCircuitedRequests,
SortByFunctionTest, TestWriterPerf, OverseerStatusTest,
TestSolrConfigHandlerCloud, TestSolrDeletionPolicy2, AutoCommitTest,
HdfsUnloadDistributedZkTest, RAMDirectoryFactoryTest, TestConfigSetsAPI,
SimpleMLTQParserTest, DistributedFacetPivotWhiteBoxTest,
CurrencyFieldOpenExchangeTest, TestReqParamsAPI, MergeStrategyTest,
SolrCmdDistributorTest, NumericFieldsTest, TestZkChroot, TestExpandComponent,
TestReversedWildcardFilterFactory, CollectionStateFormat2Test,
TestAuthenticationFramework, BasicZkTest, FullSolrCloudDistribCmdsTest,
ZkSolrClientTest, TestRandomDVFaceting, ZkCLITest, ShardRoutingCustomTest,
TestDistributedSearch, AnalysisAfterCoreReloadTest, SuggesterFSTTest,
CoreAdminHandlerTest, SuggesterTest, TestCSVLoader, TestPseudoReturnFields,
TestAtomicUpdateErrorCases, DirectUpdateHandlerOptimizeTest,
DefaultValueUpdateProcessorTest, DistanceFunctionTest, DebugComponentTest,
IndexBasedSpellCheckerTest, DirectSolrSpellCheckerTest, TestQueryTypes,
PrimitiveFieldTypeTest, RequestHandlersTest, TestValueSourceCache,
TestIndexingPerformance, RequiredFieldsTest, IndexSchemaRuntimeFieldTest,
SolrPluginUtilsTest, ReturnFieldsTest, TestCSVResponseWriter,
TestAnalyzedSuggestions, PingRequestHandlerTest, SearchHandlerTest,
TestLFUCache, TestLMJelinekMercerSimilarityFactory,
TestDefaultSimilarityFactory, TestJmxMonitoredMap, TestFastOutputStream,
URLClassifyProcessorTest, TestSuggestSpellingConverter, DOMUtilTest,
TestLRUCache, UUIDFieldTest, CursorPagingTest, TestSimpleTrackingShardHandler,
AssignTest, AsyncCallRequestStatusResponseTest,
CloudExitableDirectoryReaderTest, CollectionsAPIAsyncDistributedZkTest,
DeleteReplicaTest, DistribCursorPagingTest,
LeaderInitiatedRecoveryOnCommitTest, MultiThreadedOCPTest,
OverseerTaskQueueTest, SSLMigrationTest, SharedFSAutoReplicaFailoverTest,
SolrXmlInZkTest, TestCloudInspectUtil, TestCollectionAPI,
TestDistribDocBasedVersion, TestRebalanceLeaders,
HdfsCollectionsAPIDistributedZkTest, HdfsNNFailoverTest]
[junit4] 2> NOTE: reproduce with: ant test -Dtestcase=HdfsNNFailoverTest
-Dtests.seed=5198E59CD4979997 -Dtests.slow=true -Dtests.locale=ga
-Dtests.timezone=America/Whitehorse -Dtests.asserts=true
-Dtests.file.encoding=UTF-8
[junit4] ERROR 0.00s J1 | HdfsNNFailoverTest (suite) <<<
[junit4] > Throwable #1: java.lang.RuntimeException: Error while running
command to get file permissions : java.io.IOException: Cannot run program
"/bin/ls": error=12, Not enough space
[junit4] > at
java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
[junit4] > at
org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
[junit4] > at org.apache.hadoop.util.Shell.run(Shell.java:455)
[junit4] > at
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)
[junit4] > at
org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
[junit4] > at
org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
[junit4] > at
org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)
[junit4] > at
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:582)
[junit4] > at
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:557)
[junit4] > at
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:139)
[junit4] > at
org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
[junit4] > at
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
[junit4] > at
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
[junit4] > at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
[junit4] > at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
[junit4] > at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
[junit4] > at
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
[junit4] > at java.lang.Thread.run(Thread.java:745)
[junit4] > Caused by: java.io.IOException: error=12, Not enough space
[junit4] > at java.lang.UNIXProcess.forkAndExec(Native Method)
[junit4] > at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
[junit4] > at java.lang.ProcessImpl.start(ProcessImpl.java:130)
[junit4] > at
java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
[junit4] > ... 44 more
[junit4] > at
__randomizedtesting.SeedInfo.seed([5198E59CD4979997]:0)
[junit4] > at
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:620)
[junit4] > at
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:557)
[junit4] > at
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:139)
[junit4] > at
org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
[junit4] > at
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
[junit4] > at
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
[junit4] > at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
[junit4] > at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
[junit4] > at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
[junit4] > at
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
[junit4] > at java.lang.Thread.run(Thread.java:745)
[junit4] Completed [441/536] on J1 in 27.10s, 0 tests, 1 error <<< FAILURES!
[...truncated 11 lines...]
[junit4] Suite: org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest
[junit4] 2> Creating dataDir:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/init-core-data-001
[junit4] 2> 2525133 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false)
[junit4] 1> Formatting using clusterid: testClusterID
[junit4] 2> 2525255 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[junit4] 2> 2525271 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 2525274 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
jetty-6.1.26
[junit4] 2> 2525302 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Extract
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/hdfs
to ./temp/Jetty_solaris.vm_53451_hdfs____klmm10/webapp
[junit4] 2> 2525454 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log NO
JSP Support for /, did not find org.apache.jasper.servlet.JspServlet
[junit4] 2> 2526187 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Started HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:53451
[junit4] 2> 2539120 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.s.d.DataNode Invalid dfs.datanode.data.dir
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data2
:
[junit4] 2> java.io.IOException: Cannot run program "chmod": error=12, Not
enough space
[junit4] 2> at
java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
[junit4] 2> at
org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
[junit4] 2> at org.apache.hadoop.util.Shell.run(Shell.java:455)
[junit4] 2> at
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)
[junit4] 2> at
org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
[junit4] 2> at
org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
[junit4] 2> at
org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:656)
[junit4] 2> at
org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:490)
[junit4] 2> at
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:140)
[junit4] 2> at
org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
[junit4] 2> at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
[junit4] 2> at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:64)
[junit4] 2> at
org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest.beforeClass(HdfsRecoverLeaseTest.java:51)
[junit4] 2> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
Method)
[junit4] 2> at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
[junit4] 2> at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[junit4] 2> at java.lang.reflect.Method.invoke(Method.java:606)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
[junit4] 2> at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
[junit4] 2> at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
[junit4] 2> at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
[junit4] 2> at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
[junit4] 2> at java.lang.Thread.run(Thread.java:745)
[junit4] 2> Caused by: java.io.IOException: error=12, Not enough space
[junit4] 2> at java.lang.UNIXProcess.forkAndExec(Native Method)
[junit4] 2> at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
[junit4] 2> at java.lang.ProcessImpl.start(ProcessImpl.java:130)
[junit4] 2> at
java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
[junit4] 2> ... 44 more
[junit4] 2> 2539184 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 2539189 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
jetty-6.1.26
[junit4] 2> 2539350 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Extract
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/datanode
to ./temp/Jetty_solaris.vm_44696_datanode____axllu/webapp
[junit4] 2> 2539948 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log NO
JSP Support for /, did not find org.apache.jasper.servlet.JspServlet
[junit4] 2> 2541061 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Started HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:44696
[junit4] 2> 2546415 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.s.d.DataNode Invalid dfs.datanode.data.dir
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data3
:
[junit4] 2> java.io.IOException: Cannot run program "chmod": error=12, Not
enough space
[junit4] 2> at
java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
[junit4] 2> at
org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
[junit4] 2> at org.apache.hadoop.util.Shell.run(Shell.java:455)
[junit4] 2> at
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)
[junit4] 2> at
org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
[junit4] 2> at
org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
[junit4] 2> at
org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:656)
[junit4] 2> at
org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:490)
[junit4] 2> at
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:140)
[junit4] 2> at
org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
[junit4] 2> at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
[junit4] 2> at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:64)
[junit4] 2> at
org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest.beforeClass(HdfsRecoverLeaseTest.java:51)
[junit4] 2> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
Method)
[junit4] 2> at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
[junit4] 2> at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[junit4] 2> at java.lang.reflect.Method.invoke(Method.java:606)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
[junit4] 2> at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
[junit4] 2> at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
[junit4] 2> at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
[junit4] 2> at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
[junit4] 2> at java.lang.Thread.run(Thread.java:745)
[junit4] 2> Caused by: java.io.IOException: error=12, Not enough space
[junit4] 2> at java.lang.UNIXProcess.forkAndExec(Native Method)
[junit4] 2> at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
[junit4] 2> at java.lang.ProcessImpl.start(ProcessImpl.java:130)
[junit4] 2> at
java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
[junit4] 2> ... 44 more
[junit4] 2> 2547259 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 2547263 ERROR (DataNode:
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data1/,
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data2/]]
heartbeating to solaris-vm/127.0.0.1:38475) [ ] o.a.h.h.s.d.DataNode
Initialization failed for Block pool <registering> (Datanode Uuid unassigned)
service to solaris-vm/127.0.0.1:38475. Exiting.
[junit4] 2> org.apache.hadoop.util.DiskChecker$DiskErrorException: Too
many failed volumes - current valid volumes: 1, volumes configured: 2, volumes
failed: 1, volume failures tolerated: 0
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.<init>(FsDatasetImpl.java:261)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetFactory.newInstance(FsDatasetFactory.java:34)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetFactory.newInstance(FsDatasetFactory.java:30)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1324)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1276)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:314)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:220)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:828)
[junit4] 2> at java.lang.Thread.run(Thread.java:745)
[junit4] 2> 2547266 WARN (DataNode:
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data1/,
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data2/]]
heartbeating to solaris-vm/127.0.0.1:38475) [ ] o.a.h.h.s.d.DataNode
Ending block pool service for: Block pool <registering> (Datanode Uuid
unassigned) service to solaris-vm/127.0.0.1:38475
[junit4] 2> 2547266 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
jetty-6.1.26
[junit4] 2> 2547285 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Extract
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/datanode
to ./temp/Jetty_solaris.vm_57508_datanode____.x2ad1y/webapp
[junit4] 2> 2547437 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log NO
JSP Support for /, did not find org.apache.jasper.servlet.JspServlet
[junit4] 2> 2549849 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Started HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:57508
[junit4] 2> 2549884 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster BPOfferService in datanode DataNode{data=null,
localName='127.0.0.1:46134',
datanodeUuid='6e5cc4a9-9d93-4225-a80d-c0f90256babc', xmitsInProgress=0} failed
to connect to namenode at solaris-vm/127.0.0.1:38475
[junit4] 2> 2549946 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
[junit4] 2> 2550332 ERROR (DataNode:
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data3/,
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data4/]]
heartbeating to solaris-vm/127.0.0.1:38475) [ ] o.a.h.h.s.d.DataNode
Initialization failed for Block pool <registering> (Datanode Uuid unassigned)
service to solaris-vm/127.0.0.1:38475. Exiting.
[junit4] 2> org.apache.hadoop.util.DiskChecker$DiskErrorException: Too
many failed volumes - current valid volumes: 1, volumes configured: 2, volumes
failed: 1, volume failures tolerated: 0
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.<init>(FsDatasetImpl.java:261)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetFactory.newInstance(FsDatasetFactory.java:34)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetFactory.newInstance(FsDatasetFactory.java:30)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1324)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1276)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:314)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:220)
[junit4] 2> at
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:828)
[junit4] 2> at java.lang.Thread.run(Thread.java:745)
[junit4] 2> 2550339 WARN (DataNode:
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data3/,
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data4/]]
heartbeating to solaris-vm/127.0.0.1:38475) [ ] o.a.h.h.s.d.DataNode
Ending block pool service for: Block pool <registering> (Datanode Uuid
unassigned) service to solaris-vm/127.0.0.1:38475
[junit4] 2> 2550948 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
[junit4] 2> 2551949 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
[junit4] 2> 2552950 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
[junit4] 2> 2553951 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
[junit4] 2> 2554952 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
[junit4] 2> 2555953 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
[junit4] 2> 2556957 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
[junit4] 2> 2557957 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
[junit4] 2> 2558958 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
[junit4] 2> 2559959 WARN
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
[junit4] 2> 2560984 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
[junit4] 2> 2561156 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
[junit4] 2> 2561276 WARN
(org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager$Monitor@14563e3)
[ ] o.a.h.h.s.b.DecommissionManager Monitor interrupted:
java.lang.InterruptedException: sleep interrupted
[junit4] 2> 2561318 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ] o.m.log
Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
[junit4] 2> 2561421 ERROR
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.h.m.l.MethodMetric Error invoking method getBlocksTotal
[junit4] 2> java.lang.reflect.InvocationTargetException
[junit4] 2> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
Method)
[junit4] 2> at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
[junit4] 2> at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[junit4] 2> at java.lang.reflect.Method.invoke(Method.java:606)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.MethodMetric$2.snapshot(MethodMetric.java:111)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.MethodMetric.snapshot(MethodMetric.java:144)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.MetricsRegistry.snapshot(MetricsRegistry.java:387)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.MetricsSourceBuilder$1.getMetrics(MetricsSourceBuilder.java:79)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMetrics(MetricsSourceAdapter.java:195)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.updateJmxCache(MetricsSourceAdapter.java:172)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMBeanInfo(MetricsSourceAdapter.java:151)
[junit4] 2> at
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getClassName(DefaultMBeanServerInterceptor.java:1804)
[junit4] 2> at
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.safeGetClassName(DefaultMBeanServerInterceptor.java:1595)
[junit4] 2> at
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.checkMBeanPermission(DefaultMBeanServerInterceptor.java:1813)
[junit4] 2> at
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.exclusiveUnregisterMBean(DefaultMBeanServerInterceptor.java:430)
[junit4] 2> at
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.unregisterMBean(DefaultMBeanServerInterceptor.java:415)
[junit4] 2> at
com.sun.jmx.mbeanserver.JmxMBeanServer.unregisterMBean(JmxMBeanServer.java:546)
[junit4] 2> at
org.apache.hadoop.metrics2.util.MBeans.unregister(MBeans.java:81)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stopMBeans(MetricsSourceAdapter.java:227)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stop(MetricsSourceAdapter.java:212)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stopSources(MetricsSystemImpl.java:461)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stop(MetricsSystemImpl.java:212)
[junit4] 2> at
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.shutdown(MetricsSystemImpl.java:592)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdownInstance(DefaultMetricsSystem.java:72)
[junit4] 2> at
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdown(DefaultMetricsSystem.java:68)
[junit4] 2> at
org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics.shutdown(NameNodeMetrics.java:145)
[junit4] 2> at
org.apache.hadoop.hdfs.server.namenode.NameNode.stop(NameNode.java:822)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1720)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1699)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:838)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
[junit4] 2> at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
[junit4] 2> at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
[junit4] 2> at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:64)
[junit4] 2> at
org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest.beforeClass(HdfsRecoverLeaseTest.java:51)
[junit4] 2> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
Method)
[junit4] 2> at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
[junit4] 2> at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[junit4] 2> at java.lang.reflect.Method.invoke(Method.java:606)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
[junit4] 2> at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
[junit4] 2> at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
[junit4] 2> at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
[junit4] 2> at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
[junit4] 2> at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
[junit4] 2> at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
[junit4] 2> at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
[junit4] 2> at java.lang.Thread.run(Thread.java:745)
[junit4] 2> Caused by: java.lang.NullPointerException
[junit4] 2> at
org.apache.hadoop.hdfs.server.blockmanagement.BlocksMap.size(BlocksMap.java:198)
[junit4] 2> at
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.getTotalBlocks(BlockManager.java:3291)
[junit4] 2> at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlocksTotal(FSNamesystem.java:6223)
[junit4] 2> ... 59 more
[junit4] 2> 2561431 INFO
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [ ]
o.a.s.SolrTestCaseJ4 ###deleteCore
[junit4] 2> NOTE: test params are: codec=Lucene53,
sim=RandomSimilarityProvider(queryNorm=true,coord=no): {}, locale=es,
timezone=Etc/GMT-10
[junit4] 2> NOTE: SunOS 5.11 x86/Oracle Corporation 1.7.0_85
(32-bit)/cpus=3,threads=1,free=325210976,total=495452160
[junit4] 2> NOTE: All tests run in this JVM: [RecoveryZkTest,
ZkStateReaderTest, HdfsBasicDistributedZk2Test, HighlighterTest,
AnalyticsQueryTest, TestExceedMaxTermLength, BufferStoreTest,
SuggestComponentTest, TestCloudManagedSchema, TestConfigSetImmutable,
TestStressLucene, IndexSchemaTest, TestConfigReload, HdfsLockFactoryTest,
SoftAutoCommitTest, ZkControllerTest, SaslZkACLProviderTest,
TestRemoveLastDynamicCopyField, TestUpdate, TestMaxScoreQueryParser,
TestRemoteStreaming, TestLazyCores, SimpleCollectionCreateDeleteTest,
TestManagedSynonymFilterFactory, TestDynamicLoading, TestSolrJ,
RegexBoostProcessorTest, SpatialRPTFieldTypeTest, TestMissingGroups,
TestRestManager, TestBulkSchemaConcurrent, DistributedMLTComponentTest,
TestCloudSchemaless, HdfsWriteToMultipleCollectionsTest,
StandardRequestHandlerTest, StatelessScriptUpdateProcessorFactoryTest,
TestCollationFieldDocValues, DistributedQueryElevationComponentTest,
TestNamedUpdateProcessors, QueryResultKeyTest, TestArbitraryIndexDir,
SolrCloudExampleTest, ReplicationFactorTest, TestTrackingShardHandlerFactory,
PathHierarchyTokenizerFactoryTest, DocExpirationUpdateProcessorFactoryTest,
TestGroupingSearch, RecoveryAfterSoftCommitTest,
TestLMDirichletSimilarityFactory, TestRestoreCore, TestInitQParser,
BJQParserTest, TestStressReorder, XmlUpdateRequestHandlerTest,
TestWordDelimiterFilterFactory, DistributedQueryComponentOptimizationTest,
SecurityConfHandlerTest, TestFieldCollectionResource, OverseerTest,
ChaosMonkeyNothingIsSafeTest, TestManagedSchemaDynamicFieldResource,
TestFunctionQuery, HdfsDirectoryTest, TestReloadAndDeleteDocs,
TestSolrDeletionPolicy1, TestFastLRUCache, XsltUpdateRequestHandlerTest,
TestDocBasedVersionConstraints, TestFaceting, TestRandomRequestDistribution,
DirectUpdateHandlerTest, TestManagedSchemaFieldResource, UpdateParamsTest,
TestImplicitCoreProperties, TestSolrConfigHandlerConcurrent,
TestDocumentBuilder, PeerSyncTest, ClusterStateUpdateTest,
TestLuceneMatchVersion, SimplePostToolTest, InfoHandlerTest,
TestSurroundQueryParser, DateFieldTest, AlternateDirectoryTest, TestConfig,
TestSchemaNameResource, TestStandardQParsers, TestExactStatsCache,
JSONWriterTest, PolyFieldTest, TestStressVersions, TestMiniSolrCloudCluster,
BlockCacheTest, SyncSliceTest, TestPostingsSolrHighlighter, TestDistribIDF,
TestReplicaProperties, DistributedIntervalFacetingTest, TestSolrConfigHandler,
TestNRTOpen, TestReloadDeadlock, TestSimpleQParserPlugin, LeaderElectionTest,
HdfsChaosMonkeySafeLeaderTest, DistribJoinFromCollectionTest, SolrCoreTest,
TestFastWriter, SystemInfoHandlerTest, TestElisionMultitermQuery, TestSort,
TestEmbeddedSolrServerConstructors, ShardRoutingTest, ResourceLoaderTest,
TestSortingResponseWriter, TestFieldResource, AliasIntegrationTest,
FastVectorHighlighterTest, TestSweetSpotSimilarityFactory,
TestOverriddenPrefixQueryForCustomFieldType, HdfsBasicDistributedZkTest,
TestRTimerTree, SampleTest, TestMiniSolrCloudClusterSSL, HardAutoCommitTest,
FileUtilsTest, JsonLoaderTest, HLLSerializationTest, TestSchemaResource,
ParsingFieldUpdateProcessorsTest, PluginInfoTest, AsyncMigrateRouteKeyTest,
DeleteLastCustomShardedReplicaTest, BasicDistributedZkTest,
ChaosMonkeySafeLeaderTest, BasicDistributedZk2Test, UnloadDistributedZkTest,
TestReplicationHandler, TestDistributedGrouping, TestHashPartitioner,
DistributedSpellCheckComponentTest, TestRealTimeGet, TestReload,
DistributedTermsComponentTest, TestCoreContainer, SimpleFacetsTest,
TestSolr4Spatial, SpellCheckComponentTest, QueryElevationComponentTest,
BadIndexSchemaTest, TestFiltering, TestIndexSearcher,
ShowFileRequestHandlerTest, CurrencyFieldXmlFileTest, SolrIndexSplitterTest,
SignatureUpdateProcessorFactoryTest, TestExtendedDismaxParser,
SolrRequestParserTest, DocValuesTest, SpatialFilterTest,
SchemaVersionSpecificBehaviorTest, FieldMutatingUpdateProcessorTest,
QueryEqualityTest, DocValuesMultiTest, SolrInfoMBeanTest, CacheHeaderTest,
DisMaxRequestHandlerTest, TestQueryUtils, DocumentAnalysisRequestHandlerTest,
TermsComponentTest, DocumentBuilderTest, TermVectorComponentTest,
MoreLikeThisHandlerTest, FieldAnalysisRequestHandlerTest, LoggingHandlerTest,
TestJmxIntegration, UpdateRequestProcessorFactoryTest, MBeansHandlerTest,
TestPartialUpdateDeduplication, TestPHPSerializedResponseWriter,
TestComponentsName, TestBinaryResponseWriter, HighlighterConfigTest,
SOLR749Test, TestQuerySenderListener, TestSolrIndexConfig,
TestQuerySenderNoQuery, CopyFieldTest, ResponseLogComponentTest,
SolrIndexConfigTest, BadComponentTest, TestMergePolicyConfig, MultiTermTest,
TestDocSet, TestBinaryField, TestSearchPerf, MinimalSchemaTest,
OutputWriterTest, ExternalFileFieldSortTest, TestPhraseSuggestions,
TestCharFilters, TestCodecSupport, SynonymTokenizerTest, TestXIncludeConfig,
TestDFRSimilarityFactory, TestIBSimilarityFactory, TestBM25SimilarityFactory,
TimeZoneUtilsTest, ScriptEngineTest, TestSystemIdResolver, ZkNodePropsTest,
CircularListTest, TestRTGBase, SolrTestCaseJ4Test, TestCrossCoreJoin,
TestCursorMarkWithoutUniqueKey, TestDistributedMissingSort,
TestHighlightDedupGrouping, TestTolerantSearch, ActionThrottleTest,
CleanupOldIndexTest, CollectionReloadTest, DeleteInactiveReplicaTest,
LeaderFailoverAfterPartitionTest, MigrateRouteKeyTest,
OutOfBoxZkACLAndCredentialsProvidersTest,
OverriddenZkACLAndCredentialsProvidersTest,
OverseerCollectionConfigSetProcessorTest, OverseerRolesTest,
TestConfigSetsAPIExclusivity, TestCryptoKeys, TestDownShardTolerantSearch,
TestLeaderElectionZkExpiry, TestSolrCloudWithKerberosAlt,
TriLevelCompositeIdRoutingTest, HdfsRecoverLeaseTest]
[junit4] 2> NOTE: reproduce with: ant test
-Dtestcase=HdfsRecoverLeaseTest -Dtests.seed=5198E59CD4979997 -Dtests.slow=true
-Dtests.locale=es -Dtests.timezone=Etc/GMT-10 -Dtests.asserts=true
-Dtests.file.encoding=UTF-8
[junit4] ERROR 0.00s J0 | HdfsRecoverLeaseTest (suite) <<<
[junit4] > Throwable #1: java.io.IOException: Timed out waiting for Mini
HDFS Cluster to start
[junit4] > at
__randomizedtesting.SeedInfo.seed([5198E59CD4979997]:0)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.waitClusterUp(MiniDFSCluster.java:1197)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:832)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
[junit4] > at
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
[junit4] > at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
[junit4] > at
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:64)
[junit4] > at
org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest.beforeClass(HdfsRecoverLeaseTest.java:51)
[junit4] > at java.lang.Thread.run(Thread.java:745)
[junit4] Completed [444/536] on J0 in 36.36s, 0 tests, 1 error <<< FAILURES!
[...truncated 292 lines...]
BUILD FAILED
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:785: The
following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:729: The
following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:59: The
following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build.xml:233: The
following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/common-build.xml:524:
The following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/lucene/common-build.xml:1452:
The following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/lucene/common-build.xml:1006:
There were test failures: 536 suites, 2125 tests, 2 suite-level errors, 46
ignored (34 assumptions)
Total time: 73 minutes 29 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
[WARNINGS] Skipping publisher since build result is FAILURE
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]