Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Solaris/6/
Java: multiarch/jdk1.7.0 -d32 -server -XX:+UseParallelGC

2 tests failed.
FAILED:  junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsNNFailoverTest

Error Message:
Error while running command to get file permissions : java.io.IOException: 
Cannot run program "/bin/ls": error=12, Not enough space  at 
java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)  at 
org.apache.hadoop.util.Shell.runCommand(Shell.java:485)  at 
org.apache.hadoop.util.Shell.run(Shell.java:455)  at 
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)  at 
org.apache.hadoop.util.Shell.execCommand(Shell.java:808)  at 
org.apache.hadoop.util.Shell.execCommand(Shell.java:791)  at 
org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)  at 
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:582)
  at 
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:557)
  at 
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:139)
  at org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)  at 
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
  at 
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
  at 
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
  at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
  at 
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)  
at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
  at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)  at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)  at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)  at 
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)  at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)  
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  at java.lang.reflect.Method.invoke(Method.java:606)  at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
  at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
  at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
  at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
  at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
  at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
  at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
  at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
  at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
  at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
  at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
  at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
  at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
  at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
  at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
  at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
  at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
  at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
  at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
  at java.lang.Thread.run(Thread.java:745) Caused by: java.io.IOException: 
error=12, Not enough space  at java.lang.UNIXProcess.forkAndExec(Native Method) 
 at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)  at 
java.lang.ProcessImpl.start(ProcessImpl.java:130)  at 
java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)  ... 44 more 

Stack Trace:
java.lang.RuntimeException: Error while running command to get file permissions 
: java.io.IOException: Cannot run program "/bin/ls": error=12, Not enough space
        at java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
        at org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
        at org.apache.hadoop.util.Shell.run(Shell.java:455)
        at 
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)
        at org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
        at org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
        at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:582)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:557)
        at 
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:139)
        at org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
        at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
        at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
        at 
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
        at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.IOException: error=12, Not enough space
        at java.lang.UNIXProcess.forkAndExec(Native Method)
        at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
        at java.lang.ProcessImpl.start(ProcessImpl.java:130)
        at java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
        ... 44 more

        at __randomizedtesting.SeedInfo.seed([5198E59CD4979997]:0)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:620)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:557)
        at 
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:139)
        at org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
        at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
        at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
        at 
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
        at java.lang.Thread.run(Thread.java:745)


FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest

Error Message:
Timed out waiting for Mini HDFS Cluster to start

Stack Trace:
java.io.IOException: Timed out waiting for Mini HDFS Cluster to start
        at __randomizedtesting.SeedInfo.seed([5198E59CD4979997]:0)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.waitClusterUp(MiniDFSCluster.java:1197)
        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:832)
        at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
        at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:64)
        at 
org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest.beforeClass(HdfsRecoverLeaseTest.java:51)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
        at java.lang.Thread.run(Thread.java:745)




Build Log:
[...truncated 10931 lines...]
   [junit4] Suite: org.apache.solr.cloud.hdfs.HdfsNNFailoverTest
   [junit4]   2> Creating dataDir: 
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5198E59CD4979997-001/init-core-data-001
   [junit4]   2> 2524562 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 2524679 WARN  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried 
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 2524695 WARN  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 2524699 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
jetty-6.1.26
   [junit4]   2> 2524723 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Extract 
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/hdfs
 to ./temp/Jetty_solaris.vm_53593_hdfs____.avc5a5/webapp
   [junit4]   2> 2524894 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log NO JSP 
Support for /, did not find org.apache.jasper.servlet.JspServlet
   [junit4]   2> 2525793 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Started HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:53593
   [junit4]   2> 2540413 WARN  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 2540422 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
jetty-6.1.26
   [junit4]   2> 2540450 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Extract 
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/datanode
 to ./temp/Jetty_solaris.vm_43441_datanode____oy6l5t/webapp
   [junit4]   2> 2540656 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log NO JSP 
Support for /, did not find org.apache.jasper.servlet.JspServlet
   [junit4]   2> 2541435 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Started HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:43441
   [junit4]   2> 2546731 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
   [junit4]   2> 2546866 ERROR (DataNode: 
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data1/,
 
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data2/]]
  heartbeating to solaris-vm/127.0.0.1:43438) [    ] o.a.h.h.s.d.DataNode 
Initialization failed for Block pool <registering> (Datanode Uuid unassigned) 
service to solaris-vm/127.0.0.1:43438. Exiting. 
   [junit4]   2> java.nio.channels.ClosedByInterruptException
   [junit4]   2>        at 
java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
   [junit4]   2>        at 
sun.nio.ch.FileChannelImpl.position(FileChannelImpl.java:266)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.common.Storage.writeProperties(Storage.java:981)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.common.Storage.writeProperties(Storage.java:958)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.common.Storage.writeProperties(Storage.java:952)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.writeAll(DataStorage.java:185)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.addStorageLocations(DataStorage.java:335)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:403)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:422)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1311)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1276)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:314)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:220)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:828)
   [junit4]   2>        at java.lang.Thread.run(Thread.java:745)
   [junit4]   2> 2546870 WARN  (DataNode: 
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data1/,
 
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data2/]]
  heartbeating to solaris-vm/127.0.0.1:43438) [    ] o.a.h.h.s.d.DataNode 
Ending block pool service for: Block pool <registering> (Datanode Uuid 
unassigned) service to solaris-vm/127.0.0.1:43438
   [junit4]   2> 2546905 WARN  
(org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager$Monitor@43f0c)
 [    ] o.a.h.h.s.b.DecommissionManager Monitor interrupted: 
java.lang.InterruptedException: sleep interrupted
   [junit4]   2> 2546950 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
   [junit4]   2> 2547070 ERROR 
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.m.l.MethodMetric Error invoking method getBlocksTotal
   [junit4]   2> java.lang.reflect.InvocationTargetException
   [junit4]   2>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)
   [junit4]   2>        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
   [junit4]   2>        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   [junit4]   2>        at java.lang.reflect.Method.invoke(Method.java:606)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.MethodMetric$2.snapshot(MethodMetric.java:111)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.MethodMetric.snapshot(MethodMetric.java:144)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.MetricsRegistry.snapshot(MetricsRegistry.java:387)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.MetricsSourceBuilder$1.getMetrics(MetricsSourceBuilder.java:79)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMetrics(MetricsSourceAdapter.java:195)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.updateJmxCache(MetricsSourceAdapter.java:172)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMBeanInfo(MetricsSourceAdapter.java:151)
   [junit4]   2>        at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getClassName(DefaultMBeanServerInterceptor.java:1804)
   [junit4]   2>        at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.safeGetClassName(DefaultMBeanServerInterceptor.java:1595)
   [junit4]   2>        at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.checkMBeanPermission(DefaultMBeanServerInterceptor.java:1813)
   [junit4]   2>        at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.exclusiveUnregisterMBean(DefaultMBeanServerInterceptor.java:430)
   [junit4]   2>        at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.unregisterMBean(DefaultMBeanServerInterceptor.java:415)
   [junit4]   2>        at 
com.sun.jmx.mbeanserver.JmxMBeanServer.unregisterMBean(JmxMBeanServer.java:546)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.util.MBeans.unregister(MBeans.java:81)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stopMBeans(MetricsSourceAdapter.java:227)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stop(MetricsSourceAdapter.java:212)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stopSources(MetricsSystemImpl.java:461)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stop(MetricsSystemImpl.java:212)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.shutdown(MetricsSystemImpl.java:592)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdownInstance(DefaultMetricsSystem.java:72)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdown(DefaultMetricsSystem.java:68)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics.shutdown(NameNodeMetrics.java:145)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.stop(NameNode.java:822)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1720)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1699)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:838)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
   [junit4]   2>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)
   [junit4]   2>        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
   [junit4]   2>        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   [junit4]   2>        at java.lang.reflect.Method.invoke(Method.java:606)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
   [junit4]   2>        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
   [junit4]   2>        at java.lang.Thread.run(Thread.java:745)
   [junit4]   2> Caused by: java.lang.NullPointerException
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.blockmanagement.BlocksMap.size(BlocksMap.java:198)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.getTotalBlocks(BlockManager.java:3291)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlocksTotal(FSNamesystem.java:6223)
   [junit4]   2>        ... 58 more
   [junit4]   2> 2547097 INFO  
(SUITE-HdfsNNFailoverTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.s.SolrTestCaseJ4 ###deleteCore
   [junit4]   2> Aug 28, 2015 8:53:35 PM 
com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
   [junit4]   2> WARNING: Will linger awaiting termination of 1 leaked 
thread(s).
   [junit4]   2> NOTE: test params are: codec=Asserting(Lucene53): {}, 
docValues:{}, sim=RandomSimilarityProvider(queryNorm=true,coord=crazy): {}, 
locale=ga, timezone=America/Whitehorse
   [junit4]   2> NOTE: SunOS 5.11 x86/Oracle Corporation 1.7.0_85 
(32-bit)/cpus=3,threads=1,free=242103096,total=507248640
   [junit4]   2> NOTE: All tests run in this JVM: [TestFoldingMultitermQuery, 
SuggesterTSTTest, DistributedFacetPivotLargeTest, HdfsThreadLeakTest, TestJoin, 
TestCloudPivotFacet, BinaryUpdateRequestHandlerTest, 
TestDefaultSearchFieldResource, SpellPossibilityIteratorTest, TestRecovery, 
TestFileDictionaryLookup, DeleteShardTest, CoreMergeIndexesAdminHandlerTest, 
TestConfigOverlay, DistributedDebugComponentTest, TestShardHandlerFactory, 
TestHighFrequencyDictionaryFactory, TermVectorComponentDistributedTest, 
OpenCloseCoreStressTest, IgnoreCommitOptimizeUpdateProcessorFactoryTest, 
TestInitParams, TestStressRecovery, TestFreeTextSuggestions, 
TestSolrCoreProperties, AddBlockUpdateTest, TestSolrCLIRunExample, 
TestJettySolrRunner, CustomCollectionTest, CloudMLTQParserTest, 
TestSolrQueryParser, DirectSolrConnectionTest, 
SharedFSAutoReplicaFailoverUtilsTest, DistributedVersionInfoTest, 
DateMathParserTest, BasicFunctionalityTest, AnalyticsMergeStrategyTest, 
CSVRequestHandlerTest, AddSchemaFieldsUpdateProcessorFactoryTest, 
CollectionsAPIDistributedZkTest, DistributedSuggestComponentTest, 
SparseHLLTest, FileBasedSpellCheckerTest, RollingRestartTest, 
TestDynamicFieldResource, DistribDocExpirationUpdateProcessorTest, 
TestCollationField, HighlighterMaxOffsetTest, HttpPartitionTest, 
LukeRequestHandlerTest, ResponseHeaderTest, RemoteQueryErrorTest, TestTrie, 
NotRequiredUniqueKeyTest, TestRequestStatusCollectionAPI, 
TestSha256AuthenticationProvider, ConnectionManagerTest, 
PKIAuthenticationIntegrationTest, TestCoreDiscovery, 
TestBlendedInfixSuggestions, OpenExchangeRatesOrgProviderTest, 
SolrCoreCheckLockOnStartupTest, TestSolrXml, PreAnalyzedFieldTest, 
StatsComponentTest, TestRandomFaceting, TestManagedResourceStorage, 
PrimUtilsTest, TestScoreJoinQPScore, TestRangeQuery, ShardSplitTest, 
SliceStateTest, ConcurrentDeleteAndCreateCollectionTest, 
SpellingQueryConverterTest, TestScoreJoinQPNoScore, VersionInfoTest, 
CollectionTooManyReplicasTest, EnumFieldTest, 
VMParamsZkACLAndCredentialsProvidersTest, TestConfigSetsAPIZkFailure, 
TestClusterStateMutator, DistributedExpandComponentTest, TestRecoveryHdfs, 
SuggesterWFSTTest, DistributedFacetPivotLongTailTest, TestCustomSort, 
TestFuzzyAnalyzedSuggestions, TestBadConfig, TestManagedResource, 
TestSolrDynamicMBean, CoreAdminCreateDiscoverTest, ConvertedLegacyTest, 
TestManagedSchemaFieldTypeResource, WordBreakSolrSpellCheckerTest, 
TestPerFieldSimilarity, NoCacheHeaderTest, QueryParsingTest, BitVectorTest, 
DocValuesMissingTest, TestOmitPositions, TestJsonFacets, TestUtils, 
UniqFieldsUpdateProcessorFactoryTest, ClusterStateTest, SpellCheckCollatorTest, 
TestBlobHandler, DistributedQueueTest, EchoParamsTest, 
LeaderElectionIntegrationTest, TestOrdValues, TestShortCircuitedRequests, 
SortByFunctionTest, TestWriterPerf, OverseerStatusTest, 
TestSolrConfigHandlerCloud, TestSolrDeletionPolicy2, AutoCommitTest, 
HdfsUnloadDistributedZkTest, RAMDirectoryFactoryTest, TestConfigSetsAPI, 
SimpleMLTQParserTest, DistributedFacetPivotWhiteBoxTest, 
CurrencyFieldOpenExchangeTest, TestReqParamsAPI, MergeStrategyTest, 
SolrCmdDistributorTest, NumericFieldsTest, TestZkChroot, TestExpandComponent, 
TestReversedWildcardFilterFactory, CollectionStateFormat2Test, 
TestAuthenticationFramework, BasicZkTest, FullSolrCloudDistribCmdsTest, 
ZkSolrClientTest, TestRandomDVFaceting, ZkCLITest, ShardRoutingCustomTest, 
TestDistributedSearch, AnalysisAfterCoreReloadTest, SuggesterFSTTest, 
CoreAdminHandlerTest, SuggesterTest, TestCSVLoader, TestPseudoReturnFields, 
TestAtomicUpdateErrorCases, DirectUpdateHandlerOptimizeTest, 
DefaultValueUpdateProcessorTest, DistanceFunctionTest, DebugComponentTest, 
IndexBasedSpellCheckerTest, DirectSolrSpellCheckerTest, TestQueryTypes, 
PrimitiveFieldTypeTest, RequestHandlersTest, TestValueSourceCache, 
TestIndexingPerformance, RequiredFieldsTest, IndexSchemaRuntimeFieldTest, 
SolrPluginUtilsTest, ReturnFieldsTest, TestCSVResponseWriter, 
TestAnalyzedSuggestions, PingRequestHandlerTest, SearchHandlerTest, 
TestLFUCache, TestLMJelinekMercerSimilarityFactory, 
TestDefaultSimilarityFactory, TestJmxMonitoredMap, TestFastOutputStream, 
URLClassifyProcessorTest, TestSuggestSpellingConverter, DOMUtilTest, 
TestLRUCache, UUIDFieldTest, CursorPagingTest, TestSimpleTrackingShardHandler, 
AssignTest, AsyncCallRequestStatusResponseTest, 
CloudExitableDirectoryReaderTest, CollectionsAPIAsyncDistributedZkTest, 
DeleteReplicaTest, DistribCursorPagingTest, 
LeaderInitiatedRecoveryOnCommitTest, MultiThreadedOCPTest, 
OverseerTaskQueueTest, SSLMigrationTest, SharedFSAutoReplicaFailoverTest, 
SolrXmlInZkTest, TestCloudInspectUtil, TestCollectionAPI, 
TestDistribDocBasedVersion, TestRebalanceLeaders, 
HdfsCollectionsAPIDistributedZkTest, HdfsNNFailoverTest]
   [junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=HdfsNNFailoverTest 
-Dtests.seed=5198E59CD4979997 -Dtests.slow=true -Dtests.locale=ga 
-Dtests.timezone=America/Whitehorse -Dtests.asserts=true 
-Dtests.file.encoding=UTF-8
   [junit4] ERROR   0.00s J1 | HdfsNNFailoverTest (suite) <<<
   [junit4]    > Throwable #1: java.lang.RuntimeException: Error while running 
command to get file permissions : java.io.IOException: Cannot run program 
"/bin/ls": error=12, Not enough space
   [junit4]    >        at 
java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
   [junit4]    >        at 
org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
   [junit4]    >        at org.apache.hadoop.util.Shell.run(Shell.java:455)
   [junit4]    >        at 
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)
   [junit4]    >        at 
org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
   [junit4]    >        at 
org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
   [junit4]    >        at 
org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)
   [junit4]    >        at 
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:582)
   [junit4]    >        at 
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:557)
   [junit4]    >        at 
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:139)
   [junit4]    >        at 
org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
   [junit4]    >        at 
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
   [junit4]    >        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
   [junit4]    >        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
   [junit4]    >        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
   [junit4]    >        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
   [junit4]    >        at 
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
   [junit4]    >        at java.lang.Thread.run(Thread.java:745)
   [junit4]    > Caused by: java.io.IOException: error=12, Not enough space
   [junit4]    >        at java.lang.UNIXProcess.forkAndExec(Native Method)
   [junit4]    >        at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
   [junit4]    >        at java.lang.ProcessImpl.start(ProcessImpl.java:130)
   [junit4]    >        at 
java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
   [junit4]    >        ... 44 more
   [junit4]    >        at 
__randomizedtesting.SeedInfo.seed([5198E59CD4979997]:0)
   [junit4]    >        at 
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.loadPermissionInfo(RawLocalFileSystem.java:620)
   [junit4]    >        at 
org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.getPermission(RawLocalFileSystem.java:557)
   [junit4]    >        at 
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:139)
   [junit4]    >        at 
org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
   [junit4]    >        at 
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
   [junit4]    >        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
   [junit4]    >        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
   [junit4]    >        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
   [junit4]    >        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
   [junit4]    >        at 
org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverTest.java:44)
   [junit4]    >        at java.lang.Thread.run(Thread.java:745)
   [junit4] Completed [441/536] on J1 in 27.10s, 0 tests, 1 error <<< FAILURES!

[...truncated 11 lines...]
   [junit4] Suite: org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest
   [junit4]   2> Creating dataDir: 
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/init-core-data-001
   [junit4]   2> 2525133 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false)
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 2525255 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried 
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 2525271 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 2525274 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
jetty-6.1.26
   [junit4]   2> 2525302 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Extract 
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/hdfs
 to ./temp/Jetty_solaris.vm_53451_hdfs____klmm10/webapp
   [junit4]   2> 2525454 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log NO 
JSP Support for /, did not find org.apache.jasper.servlet.JspServlet
   [junit4]   2> 2526187 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Started HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:53451
   [junit4]   2> 2539120 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.s.d.DataNode Invalid dfs.datanode.data.dir 
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data2
 : 
   [junit4]   2> java.io.IOException: Cannot run program "chmod": error=12, Not 
enough space
   [junit4]   2>        at 
java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
   [junit4]   2>        at 
org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
   [junit4]   2>        at org.apache.hadoop.util.Shell.run(Shell.java:455)
   [junit4]   2>        at 
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)
   [junit4]   2>        at 
org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
   [junit4]   2>        at 
org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
   [junit4]   2>        at 
org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:656)
   [junit4]   2>        at 
org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:490)
   [junit4]   2>        at 
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:140)
   [junit4]   2>        at 
org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:64)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest.beforeClass(HdfsRecoverLeaseTest.java:51)
   [junit4]   2>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)
   [junit4]   2>        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
   [junit4]   2>        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   [junit4]   2>        at java.lang.reflect.Method.invoke(Method.java:606)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
   [junit4]   2>        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
   [junit4]   2>        at java.lang.Thread.run(Thread.java:745)
   [junit4]   2> Caused by: java.io.IOException: error=12, Not enough space
   [junit4]   2>        at java.lang.UNIXProcess.forkAndExec(Native Method)
   [junit4]   2>        at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
   [junit4]   2>        at java.lang.ProcessImpl.start(ProcessImpl.java:130)
   [junit4]   2>        at 
java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
   [junit4]   2>        ... 44 more
   [junit4]   2> 2539184 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 2539189 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
jetty-6.1.26
   [junit4]   2> 2539350 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Extract 
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/datanode
 to ./temp/Jetty_solaris.vm_44696_datanode____axllu/webapp
   [junit4]   2> 2539948 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log NO 
JSP Support for /, did not find org.apache.jasper.servlet.JspServlet
   [junit4]   2> 2541061 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Started HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:44696
   [junit4]   2> 2546415 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.s.d.DataNode Invalid dfs.datanode.data.dir 
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data3
 : 
   [junit4]   2> java.io.IOException: Cannot run program "chmod": error=12, Not 
enough space
   [junit4]   2>        at 
java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
   [junit4]   2>        at 
org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
   [junit4]   2>        at org.apache.hadoop.util.Shell.run(Shell.java:455)
   [junit4]   2>        at 
org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)
   [junit4]   2>        at 
org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
   [junit4]   2>        at 
org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
   [junit4]   2>        at 
org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:656)
   [junit4]   2>        at 
org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:490)
   [junit4]   2>        at 
org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:140)
   [junit4]   2>        at 
org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2239)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2281)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2263)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2155)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1443)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:828)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:64)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest.beforeClass(HdfsRecoverLeaseTest.java:51)
   [junit4]   2>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)
   [junit4]   2>        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
   [junit4]   2>        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   [junit4]   2>        at java.lang.reflect.Method.invoke(Method.java:606)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
   [junit4]   2>        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
   [junit4]   2>        at java.lang.Thread.run(Thread.java:745)
   [junit4]   2> Caused by: java.io.IOException: error=12, Not enough space
   [junit4]   2>        at java.lang.UNIXProcess.forkAndExec(Native Method)
   [junit4]   2>        at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
   [junit4]   2>        at java.lang.ProcessImpl.start(ProcessImpl.java:130)
   [junit4]   2>        at 
java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
   [junit4]   2>        ... 44 more
   [junit4]   2> 2547259 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 2547263 ERROR (DataNode: 
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data1/,
 
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data2/]]
  heartbeating to solaris-vm/127.0.0.1:38475) [    ] o.a.h.h.s.d.DataNode 
Initialization failed for Block pool <registering> (Datanode Uuid unassigned) 
service to solaris-vm/127.0.0.1:38475. Exiting. 
   [junit4]   2> org.apache.hadoop.util.DiskChecker$DiskErrorException: Too 
many failed volumes - current valid volumes: 1, volumes configured: 2, volumes 
failed: 1, volume failures tolerated: 0
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.<init>(FsDatasetImpl.java:261)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetFactory.newInstance(FsDatasetFactory.java:34)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetFactory.newInstance(FsDatasetFactory.java:30)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1324)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1276)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:314)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:220)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:828)
   [junit4]   2>        at java.lang.Thread.run(Thread.java:745)
   [junit4]   2> 2547266 WARN  (DataNode: 
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data1/,
 
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data2/]]
  heartbeating to solaris-vm/127.0.0.1:38475) [    ] o.a.h.h.s.d.DataNode 
Ending block pool service for: Block pool <registering> (Datanode Uuid 
unassigned) service to solaris-vm/127.0.0.1:38475
   [junit4]   2> 2547266 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
jetty-6.1.26
   [junit4]   2> 2547285 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Extract 
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/datanode
 to ./temp/Jetty_solaris.vm_57508_datanode____.x2ad1y/webapp
   [junit4]   2> 2547437 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log NO 
JSP Support for /, did not find org.apache.jasper.servlet.JspServlet
   [junit4]   2> 2549849 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Started HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:57508
   [junit4]   2> 2549884 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster BPOfferService in datanode DataNode{data=null, 
localName='127.0.0.1:46134', 
datanodeUuid='6e5cc4a9-9d93-4225-a80d-c0f90256babc', xmitsInProgress=0} failed 
to connect to namenode at solaris-vm/127.0.0.1:38475
   [junit4]   2> 2549946 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
   [junit4]   2> 2550332 ERROR (DataNode: 
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data3/,
 
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data4/]]
  heartbeating to solaris-vm/127.0.0.1:38475) [    ] o.a.h.h.s.d.DataNode 
Initialization failed for Block pool <registering> (Datanode Uuid unassigned) 
service to solaris-vm/127.0.0.1:38475. Exiting. 
   [junit4]   2> org.apache.hadoop.util.DiskChecker$DiskErrorException: Too 
many failed volumes - current valid volumes: 1, volumes configured: 2, volumes 
failed: 1, volume failures tolerated: 0
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.<init>(FsDatasetImpl.java:261)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetFactory.newInstance(FsDatasetFactory.java:34)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetFactory.newInstance(FsDatasetFactory.java:30)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1324)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1276)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:314)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:220)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:828)
   [junit4]   2>        at java.lang.Thread.run(Thread.java:745)
   [junit4]   2> 2550339 WARN  (DataNode: 
[[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data3/,
 
[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.HdfsRecoverLeaseTest_5198E59CD4979997-001/tempDir-001/hdfsBaseDir/data/data4/]]
  heartbeating to solaris-vm/127.0.0.1:38475) [    ] o.a.h.h.s.d.DataNode 
Ending block pool service for: Block pool <registering> (Datanode Uuid 
unassigned) service to solaris-vm/127.0.0.1:38475
   [junit4]   2> 2550948 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
   [junit4]   2> 2551949 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
   [junit4]   2> 2552950 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
   [junit4]   2> 2553951 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
   [junit4]   2> 2554952 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
   [junit4]   2> 2555953 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
   [junit4]   2> 2556957 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
   [junit4]   2> 2557957 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
   [junit4]   2> 2558958 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
   [junit4]   2> 2559959 WARN  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.h.MiniDFSCluster Waiting for the Mini HDFS Cluster to start...
   [junit4]   2> 2560984 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
   [junit4]   2> 2561156 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
   [junit4]   2> 2561276 WARN  
(org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager$Monitor@14563e3)
 [    ] o.a.h.h.s.b.DecommissionManager Monitor interrupted: 
java.lang.InterruptedException: sleep interrupted
   [junit4]   2> 2561318 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] o.m.log 
Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
   [junit4]   2> 2561421 ERROR 
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.h.m.l.MethodMetric Error invoking method getBlocksTotal
   [junit4]   2> java.lang.reflect.InvocationTargetException
   [junit4]   2>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)
   [junit4]   2>        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
   [junit4]   2>        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   [junit4]   2>        at java.lang.reflect.Method.invoke(Method.java:606)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.MethodMetric$2.snapshot(MethodMetric.java:111)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.MethodMetric.snapshot(MethodMetric.java:144)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.MetricsRegistry.snapshot(MetricsRegistry.java:387)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.MetricsSourceBuilder$1.getMetrics(MetricsSourceBuilder.java:79)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMetrics(MetricsSourceAdapter.java:195)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.updateJmxCache(MetricsSourceAdapter.java:172)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMBeanInfo(MetricsSourceAdapter.java:151)
   [junit4]   2>        at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getClassName(DefaultMBeanServerInterceptor.java:1804)
   [junit4]   2>        at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.safeGetClassName(DefaultMBeanServerInterceptor.java:1595)
   [junit4]   2>        at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.checkMBeanPermission(DefaultMBeanServerInterceptor.java:1813)
   [junit4]   2>        at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.exclusiveUnregisterMBean(DefaultMBeanServerInterceptor.java:430)
   [junit4]   2>        at 
com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.unregisterMBean(DefaultMBeanServerInterceptor.java:415)
   [junit4]   2>        at 
com.sun.jmx.mbeanserver.JmxMBeanServer.unregisterMBean(JmxMBeanServer.java:546)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.util.MBeans.unregister(MBeans.java:81)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stopMBeans(MetricsSourceAdapter.java:227)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stop(MetricsSourceAdapter.java:212)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stopSources(MetricsSystemImpl.java:461)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stop(MetricsSystemImpl.java:212)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.shutdown(MetricsSystemImpl.java:592)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdownInstance(DefaultMetricsSystem.java:72)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdown(DefaultMetricsSystem.java:68)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics.shutdown(NameNodeMetrics.java:145)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.stop(NameNode.java:822)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1720)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1699)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:838)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:64)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest.beforeClass(HdfsRecoverLeaseTest.java:51)
   [junit4]   2>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)
   [junit4]   2>        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
   [junit4]   2>        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   [junit4]   2>        at java.lang.reflect.Method.invoke(Method.java:606)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:776)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
   [junit4]   2>        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
   [junit4]   2>        at java.lang.Thread.run(Thread.java:745)
   [junit4]   2> Caused by: java.lang.NullPointerException
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.blockmanagement.BlocksMap.size(BlocksMap.java:198)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.getTotalBlocks(BlockManager.java:3291)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlocksTotal(FSNamesystem.java:6223)
   [junit4]   2>        ... 59 more
   [junit4]   2> 2561431 INFO  
(SUITE-HdfsRecoverLeaseTest-seed#[5198E59CD4979997]-worker) [    ] 
o.a.s.SolrTestCaseJ4 ###deleteCore
   [junit4]   2> NOTE: test params are: codec=Lucene53, 
sim=RandomSimilarityProvider(queryNorm=true,coord=no): {}, locale=es, 
timezone=Etc/GMT-10
   [junit4]   2> NOTE: SunOS 5.11 x86/Oracle Corporation 1.7.0_85 
(32-bit)/cpus=3,threads=1,free=325210976,total=495452160
   [junit4]   2> NOTE: All tests run in this JVM: [RecoveryZkTest, 
ZkStateReaderTest, HdfsBasicDistributedZk2Test, HighlighterTest, 
AnalyticsQueryTest, TestExceedMaxTermLength, BufferStoreTest, 
SuggestComponentTest, TestCloudManagedSchema, TestConfigSetImmutable, 
TestStressLucene, IndexSchemaTest, TestConfigReload, HdfsLockFactoryTest, 
SoftAutoCommitTest, ZkControllerTest, SaslZkACLProviderTest, 
TestRemoveLastDynamicCopyField, TestUpdate, TestMaxScoreQueryParser, 
TestRemoteStreaming, TestLazyCores, SimpleCollectionCreateDeleteTest, 
TestManagedSynonymFilterFactory, TestDynamicLoading, TestSolrJ, 
RegexBoostProcessorTest, SpatialRPTFieldTypeTest, TestMissingGroups, 
TestRestManager, TestBulkSchemaConcurrent, DistributedMLTComponentTest, 
TestCloudSchemaless, HdfsWriteToMultipleCollectionsTest, 
StandardRequestHandlerTest, StatelessScriptUpdateProcessorFactoryTest, 
TestCollationFieldDocValues, DistributedQueryElevationComponentTest, 
TestNamedUpdateProcessors, QueryResultKeyTest, TestArbitraryIndexDir, 
SolrCloudExampleTest, ReplicationFactorTest, TestTrackingShardHandlerFactory, 
PathHierarchyTokenizerFactoryTest, DocExpirationUpdateProcessorFactoryTest, 
TestGroupingSearch, RecoveryAfterSoftCommitTest, 
TestLMDirichletSimilarityFactory, TestRestoreCore, TestInitQParser, 
BJQParserTest, TestStressReorder, XmlUpdateRequestHandlerTest, 
TestWordDelimiterFilterFactory, DistributedQueryComponentOptimizationTest, 
SecurityConfHandlerTest, TestFieldCollectionResource, OverseerTest, 
ChaosMonkeyNothingIsSafeTest, TestManagedSchemaDynamicFieldResource, 
TestFunctionQuery, HdfsDirectoryTest, TestReloadAndDeleteDocs, 
TestSolrDeletionPolicy1, TestFastLRUCache, XsltUpdateRequestHandlerTest, 
TestDocBasedVersionConstraints, TestFaceting, TestRandomRequestDistribution, 
DirectUpdateHandlerTest, TestManagedSchemaFieldResource, UpdateParamsTest, 
TestImplicitCoreProperties, TestSolrConfigHandlerConcurrent, 
TestDocumentBuilder, PeerSyncTest, ClusterStateUpdateTest, 
TestLuceneMatchVersion, SimplePostToolTest, InfoHandlerTest, 
TestSurroundQueryParser, DateFieldTest, AlternateDirectoryTest, TestConfig, 
TestSchemaNameResource, TestStandardQParsers, TestExactStatsCache, 
JSONWriterTest, PolyFieldTest, TestStressVersions, TestMiniSolrCloudCluster, 
BlockCacheTest, SyncSliceTest, TestPostingsSolrHighlighter, TestDistribIDF, 
TestReplicaProperties, DistributedIntervalFacetingTest, TestSolrConfigHandler, 
TestNRTOpen, TestReloadDeadlock, TestSimpleQParserPlugin, LeaderElectionTest, 
HdfsChaosMonkeySafeLeaderTest, DistribJoinFromCollectionTest, SolrCoreTest, 
TestFastWriter, SystemInfoHandlerTest, TestElisionMultitermQuery, TestSort, 
TestEmbeddedSolrServerConstructors, ShardRoutingTest, ResourceLoaderTest, 
TestSortingResponseWriter, TestFieldResource, AliasIntegrationTest, 
FastVectorHighlighterTest, TestSweetSpotSimilarityFactory, 
TestOverriddenPrefixQueryForCustomFieldType, HdfsBasicDistributedZkTest, 
TestRTimerTree, SampleTest, TestMiniSolrCloudClusterSSL, HardAutoCommitTest, 
FileUtilsTest, JsonLoaderTest, HLLSerializationTest, TestSchemaResource, 
ParsingFieldUpdateProcessorsTest, PluginInfoTest, AsyncMigrateRouteKeyTest, 
DeleteLastCustomShardedReplicaTest, BasicDistributedZkTest, 
ChaosMonkeySafeLeaderTest, BasicDistributedZk2Test, UnloadDistributedZkTest, 
TestReplicationHandler, TestDistributedGrouping, TestHashPartitioner, 
DistributedSpellCheckComponentTest, TestRealTimeGet, TestReload, 
DistributedTermsComponentTest, TestCoreContainer, SimpleFacetsTest, 
TestSolr4Spatial, SpellCheckComponentTest, QueryElevationComponentTest, 
BadIndexSchemaTest, TestFiltering, TestIndexSearcher, 
ShowFileRequestHandlerTest, CurrencyFieldXmlFileTest, SolrIndexSplitterTest, 
SignatureUpdateProcessorFactoryTest, TestExtendedDismaxParser, 
SolrRequestParserTest, DocValuesTest, SpatialFilterTest, 
SchemaVersionSpecificBehaviorTest, FieldMutatingUpdateProcessorTest, 
QueryEqualityTest, DocValuesMultiTest, SolrInfoMBeanTest, CacheHeaderTest, 
DisMaxRequestHandlerTest, TestQueryUtils, DocumentAnalysisRequestHandlerTest, 
TermsComponentTest, DocumentBuilderTest, TermVectorComponentTest, 
MoreLikeThisHandlerTest, FieldAnalysisRequestHandlerTest, LoggingHandlerTest, 
TestJmxIntegration, UpdateRequestProcessorFactoryTest, MBeansHandlerTest, 
TestPartialUpdateDeduplication, TestPHPSerializedResponseWriter, 
TestComponentsName, TestBinaryResponseWriter, HighlighterConfigTest, 
SOLR749Test, TestQuerySenderListener, TestSolrIndexConfig, 
TestQuerySenderNoQuery, CopyFieldTest, ResponseLogComponentTest, 
SolrIndexConfigTest, BadComponentTest, TestMergePolicyConfig, MultiTermTest, 
TestDocSet, TestBinaryField, TestSearchPerf, MinimalSchemaTest, 
OutputWriterTest, ExternalFileFieldSortTest, TestPhraseSuggestions, 
TestCharFilters, TestCodecSupport, SynonymTokenizerTest, TestXIncludeConfig, 
TestDFRSimilarityFactory, TestIBSimilarityFactory, TestBM25SimilarityFactory, 
TimeZoneUtilsTest, ScriptEngineTest, TestSystemIdResolver, ZkNodePropsTest, 
CircularListTest, TestRTGBase, SolrTestCaseJ4Test, TestCrossCoreJoin, 
TestCursorMarkWithoutUniqueKey, TestDistributedMissingSort, 
TestHighlightDedupGrouping, TestTolerantSearch, ActionThrottleTest, 
CleanupOldIndexTest, CollectionReloadTest, DeleteInactiveReplicaTest, 
LeaderFailoverAfterPartitionTest, MigrateRouteKeyTest, 
OutOfBoxZkACLAndCredentialsProvidersTest, 
OverriddenZkACLAndCredentialsProvidersTest, 
OverseerCollectionConfigSetProcessorTest, OverseerRolesTest, 
TestConfigSetsAPIExclusivity, TestCryptoKeys, TestDownShardTolerantSearch, 
TestLeaderElectionZkExpiry, TestSolrCloudWithKerberosAlt, 
TriLevelCompositeIdRoutingTest, HdfsRecoverLeaseTest]
   [junit4]   2> NOTE: reproduce with: ant test  
-Dtestcase=HdfsRecoverLeaseTest -Dtests.seed=5198E59CD4979997 -Dtests.slow=true 
-Dtests.locale=es -Dtests.timezone=Etc/GMT-10 -Dtests.asserts=true 
-Dtests.file.encoding=UTF-8
   [junit4] ERROR   0.00s J0 | HdfsRecoverLeaseTest (suite) <<<
   [junit4]    > Throwable #1: java.io.IOException: Timed out waiting for Mini 
HDFS Cluster to start
   [junit4]    >        at 
__randomizedtesting.SeedInfo.seed([5198E59CD4979997]:0)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.waitClusterUp(MiniDFSCluster.java:1197)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:832)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
   [junit4]    >        at 
org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
   [junit4]    >        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
   [junit4]    >        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:64)
   [junit4]    >        at 
org.apache.solr.cloud.hdfs.HdfsRecoverLeaseTest.beforeClass(HdfsRecoverLeaseTest.java:51)
   [junit4]    >        at java.lang.Thread.run(Thread.java:745)
   [junit4] Completed [444/536] on J0 in 36.36s, 0 tests, 1 error <<< FAILURES!

[...truncated 292 lines...]
BUILD FAILED
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:785: The 
following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:729: The 
following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:59: The 
following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build.xml:233: The 
following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/common-build.xml:524:
 The following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/lucene/common-build.xml:1452:
 The following error occurred while executing this line:
/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/lucene/common-build.xml:1006:
 There were test failures: 536 suites, 2125 tests, 2 suite-level errors, 46 
ignored (34 assumptions)

Total time: 73 minutes 29 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
[WARNINGS] Skipping publisher since build result is FAILURE
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to