I am still digging... On Solaris there seems to be a general forking problem on 
32 bit processes.

Uwe

-----
Uwe Schindler
H.-H.-Meier-Allee 63, D-28213 Bremen
http://www.thetaphi.de
eMail: [email protected]

> -----Original Message-----
> From: Policeman Jenkins Server [mailto:[email protected]]
> Sent: Saturday, August 29, 2015 11:02 PM
> To: [email protected]; [email protected]; [email protected]
> Subject: [JENKINS] Lucene-Solr-5.x-Solaris (multiarch/jdk1.7.0) - Build # 9 -
> Still Failing!
> 
> Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Solaris/9/
> Java: multiarch/jdk1.7.0 -d32 -server -XX:+UseConcMarkSweepGC
> 
> 4 tests failed.
> FAILED:
> junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsNNFailoverTest
> 
> Error Message:
> Error while running command to get file permissions : java.io.IOException:
> Cannot run program "/bin/ls": error=12, Not enough space  at
> java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)  at
> org.apache.hadoop.util.Shell.runCommand(Shell.java:485)  at
> org.apache.hadoop.util.Shell.run(Shell.java:455)  at
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715
> )  at org.apache.hadoop.util.Shell.execCommand(Shell.java:808)  at
> org.apache.hadoop.util.Shell.execCommand(Shell.java:791)  at
> org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)  at
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo
> adPermissionInfo(RawLocalFileSystem.java:582)  at
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g
> etPermission(RawLocalFileSystem.java:557)  at
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> DiskChecker.java:139)  at
> org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)  at
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> checkDir(DataNode.java:2239)  at
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> DataNode.java:2281)  at
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> de.java:2263)  at
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> ataNode.java:2155)  at
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> a:1443)  at
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> ava:828)  at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)  at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)  at
> org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)  at
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> est.java:44)  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)  at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> ava:57)  at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> sorImpl.java:43)  at java.lang.reflect.Method.invoke(Method.java:606)  at
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> dRunner.java:1627)  at
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> mizedRunner.java:776)  at
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> mizedRunner.java:792)  at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)  at
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> evaluate(SystemPropertiesRestoreRule.java:57)  at
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> fterRule.java:46)  at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)  at
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> assName.java:42)  at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)  at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)  at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)  at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)  at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)  at
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> ertionsRequired.java:54)  at
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> .java:48)  at
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> IgnoreAfterMaxFailures.java:65)  at
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> TestSuites.java:55)  at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)  at
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> run(ThreadLeakControl.java:365)  at java.lang.Thread.run(Thread.java:745)
> Caused by: java.io.IOException: error=12, Not enough space  at
> java.lang.UNIXProcess.forkAndExec(Native Method)  at
> java.lang.UNIXProcess.<init>(UNIXProcess.java:137)  at
> java.lang.ProcessImpl.start(ProcessImpl.java:130)  at
> java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)  ... 44 more
> 
> Stack Trace:
> java.lang.RuntimeException: Error while running command to get file
> permissions : java.io.IOException: Cannot run program "/bin/ls": error=12,
> Not enough space
>       at java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
>       at org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
>       at org.apache.hadoop.util.Shell.run(Shell.java:455)
>       at
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715
> )
>       at org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
>       at org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
>       at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)
>       at
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo
> adPermissionInfo(RawLocalFileSystem.java:582)
>       at
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g
> etPermission(RawLocalFileSystem.java:557)
>       at
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> DiskChecker.java:139)
>       at
> org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
>       at
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> checkDir(DataNode.java:2239)
>       at
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> DataNode.java:2281)
>       at
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> de.java:2263)
>       at
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> ataNode.java:2155)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> a:1443)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> ava:828)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
>       at
> org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
>       at
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> est.java:44)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> ava:57)
>       at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> sorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:606)
>       at
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> dRunner.java:1627)
>       at
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> mizedRunner.java:776)
>       at
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> mizedRunner.java:792)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> evaluate(SystemPropertiesRestoreRule.java:57)
>       at
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> fterRule.java:46)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> assName.java:42)
>       at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
>       at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> ertionsRequired.java:54)
>       at
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> .java:48)
>       at
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> IgnoreAfterMaxFailures.java:65)
>       at
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> TestSuites.java:55)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> run(ThreadLeakControl.java:365)
>       at java.lang.Thread.run(Thread.java:745)
> Caused by: java.io.IOException: error=12, Not enough space
>       at java.lang.UNIXProcess.forkAndExec(Native Method)
>       at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
>       at java.lang.ProcessImpl.start(ProcessImpl.java:130)
>       at java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
>       ... 44 more
> 
>       at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
>       at
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo
> adPermissionInfo(RawLocalFileSystem.java:620)
>       at
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g
> etPermission(RawLocalFileSystem.java:557)
>       at
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> DiskChecker.java:139)
>       at
> org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
>       at
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> checkDir(DataNode.java:2239)
>       at
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> DataNode.java:2281)
>       at
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> de.java:2263)
>       at
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> ataNode.java:2155)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> a:1443)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> ava:828)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
>       at
> org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
>       at
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> est.java:44)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> ava:57)
>       at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> sorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:606)
>       at
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> dRunner.java:1627)
>       at
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> mizedRunner.java:776)
>       at
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> mizedRunner.java:792)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> evaluate(SystemPropertiesRestoreRule.java:57)
>       at
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> fterRule.java:46)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> assName.java:42)
>       at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
>       at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> ertionsRequired.java:54)
>       at
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> .java:48)
>       at
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> IgnoreAfterMaxFailures.java:65)
>       at
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> TestSuites.java:55)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> run(ThreadLeakControl.java:365)
>       at java.lang.Thread.run(Thread.java:745)
> 
> 
> FAILED:
> junit.framework.TestSuite.org.apache.solr.store.hdfs.HdfsDirectoryTest
> 
> Error Message:
> access denied ("java.io.FilePermission"
> "/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-
> core/test/J1" "write")
> 
> Stack Trace:
> java.security.AccessControlException: access denied
> ("java.io.FilePermission" "/export/home/jenkins/workspace/Lucene-Solr-
> 5.x-Solaris/solr/build/solr-core/test/J1" "write")
>       at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
>       at
> java.security.AccessControlContext.checkPermission(AccessControlContext.j
> ava:395)
>       at
> java.security.AccessController.checkPermission(AccessController.java:559)
>       at
> java.lang.SecurityManager.checkPermission(SecurityManager.java:549)
>       at java.lang.SecurityManager.checkWrite(SecurityManager.java:979)
>       at java.io.File.canWrite(File.java:785)
>       at org.apache.hadoop.fs.FileUtil.canWrite(FileUtil.java:1002)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.createPermissionsDiagnosisString(
> MiniDFSCluster.java:856)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> ava:812)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
>       at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
>       at
> org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
>       at
> org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:60)
>       at
> org.apache.solr.store.hdfs.HdfsDirectoryTest.beforeClass(HdfsDirectoryTest.
> java:62)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> ava:57)
>       at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> sorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:606)
>       at
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> dRunner.java:1627)
>       at
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> mizedRunner.java:776)
>       at
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> mizedRunner.java:792)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> evaluate(SystemPropertiesRestoreRule.java:57)
>       at
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> fterRule.java:46)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> assName.java:42)
>       at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
>       at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> ertionsRequired.java:54)
>       at
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> .java:48)
>       at
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> IgnoreAfterMaxFailures.java:65)
>       at
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> TestSuites.java:55)
>       at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>       at
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> run(ThreadLeakControl.java:365)
>       at java.lang.Thread.run(Thread.java:745)
> 
> 
> FAILED:
> junit.framework.TestSuite.org.apache.solr.store.hdfs.HdfsDirectoryTest
> 
> Error Message:
> 1 thread leaked from SUITE scope at
> org.apache.solr.store.hdfs.HdfsDirectoryTest:     1) Thread[id=20389,
> name=IPC Server idle connection scanner for port 41610, state=WAITING,
> group=TGRP-HdfsDirectoryTest]         at java.lang.Object.wait(Native
> Method)         at java.lang.Object.wait(Object.java:503)         at
> java.util.TimerThread.mainLoop(Timer.java:526)         at
> java.util.TimerThread.run(Timer.java:505)
> 
> Stack Trace:
> com.carrotsearch.randomizedtesting.ThreadLeakError: 1 thread leaked from
> SUITE scope at org.apache.solr.store.hdfs.HdfsDirectoryTest:
>    1) Thread[id=20389, name=IPC Server idle connection scanner for port
> 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
>         at java.lang.Object.wait(Native Method)
>         at java.lang.Object.wait(Object.java:503)
>         at java.util.TimerThread.mainLoop(Timer.java:526)
>         at java.util.TimerThread.run(Timer.java:505)
>       at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
> 
> 
> FAILED:
> junit.framework.TestSuite.org.apache.solr.store.hdfs.HdfsDirectoryTest
> 
> Error Message:
> There are still zombie threads that couldn't be terminated:    1)
> Thread[id=20389, name=IPC Server idle connection scanner for port 41610,
> state=WAITING, group=TGRP-HdfsDirectoryTest]         at
> java.lang.Object.wait(Native Method)         at
> java.lang.Object.wait(Object.java:503)         at
> java.util.TimerThread.mainLoop(Timer.java:526)         at
> java.util.TimerThread.run(Timer.java:505)
> 
> Stack Trace:
> com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie
> threads that couldn't be terminated:
>    1) Thread[id=20389, name=IPC Server idle connection scanner for port
> 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
>         at java.lang.Object.wait(Native Method)
>         at java.lang.Object.wait(Object.java:503)
>         at java.util.TimerThread.mainLoop(Timer.java:526)
>         at java.util.TimerThread.run(Timer.java:505)
>       at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
> 
> 
> 
> 
> Build Log:
> [...truncated 10577 lines...]
>    [junit4] Suite: org.apache.solr.cloud.hdfs.HdfsNNFailoverTest
>    [junit4]   2> Creating dataDir: /export/home/jenkins/workspace/Lucene-
> Solr-5.x-Solaris/solr/build/solr-
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> 001/init-core-data-001
>    [junit4]   2> 2599844 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.s.BaseDistributedSearchTestCase
> Setting hostContext system property: /
>    [junit4]   2> 2616331 WARN  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.h.u.NativeCodeLoader Unable to
> load native-hadoop library for your platform... using builtin-java classes
> where applicable
>    [junit4]   1> Formatting using clusterid: testClusterID
>    [junit4]   2> 2617524 WARN  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.h.m.i.MetricsConfig Cannot
> locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-
> metrics2.properties
>    [junit4]   2> 2617755 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log Logging to
> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
> org.mortbay.log.Slf4jLog
>    [junit4]   2> 2617771 WARN  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.h.h.HttpRequestLog Jetty
> request log can only be enabled using Log4j
>    [junit4]   2> 2617878 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log jetty-6.1.26
>    [junit4]   2> 2617942 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log Extract
> jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-
> hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/hdfs to
> ./temp/Jetty_solaris.vm_35231_hdfs____thayv4/webapp
>    [junit4]   2> 2618129 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log NO JSP Support for /, did not
> find org.apache.jasper.servlet.JspServlet
>    [junit4]   2> 2619464 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log Started
> HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:35231
>    [junit4]   2> 2637264 WARN  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.h.h.s.d.DataNode Invalid
> dfs.datanode.data.dir /export/home/jenkins/workspace/Lucene-Solr-5.x-
> Solaris/solr/build/solr-
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> 001/tempDir-001/hdfsBaseDir/data/data2 :
>    [junit4]   2> java.io.IOException: Cannot run program "chmod": error=12,
> Not enough space
>    [junit4]   2>      at 
> java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
>    [junit4]   2>      at 
> org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
>    [junit4]   2>      at org.apache.hadoop.util.Shell.run(Shell.java:455)
>    [junit4]   2>      at
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715
> )
>    [junit4]   2>      at
> org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
>    [junit4]   2>      at
> org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
>    [junit4]   2>      at
> org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSyste
> m.java:656)
>    [junit4]   2>      at
> org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:4
> 90)
>    [junit4]   2>      at
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> DiskChecker.java:140)
>    [junit4]   2>      at
> org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> checkDir(DataNode.java:2239)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> DataNode.java:2281)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> de.java:2263)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> ataNode.java:2155)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> a:1443)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> ava:828)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
>    [junit4]   2>      at
> org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
>    [junit4]   2>      at
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> est.java:44)
>    [junit4]   2>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
>    [junit4]   2>      at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> ava:57)
>    [junit4]   2>      at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> sorImpl.java:43)
>    [junit4]   2>      at java.lang.reflect.Method.invoke(Method.java:606)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> dRunner.java:1627)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> mizedRunner.java:776)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> mizedRunner.java:792)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> evaluate(SystemPropertiesRestoreRule.java:57)
>    [junit4]   2>      at
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> fterRule.java:46)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> assName.java:42)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> ertionsRequired.java:54)
>    [junit4]   2>      at
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> .java:48)
>    [junit4]   2>      at
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> IgnoreAfterMaxFailures.java:65)
>    [junit4]   2>      at
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> TestSuites.java:55)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> run(ThreadLeakControl.java:365)
>    [junit4]   2>      at java.lang.Thread.run(Thread.java:745)
>    [junit4]   2> Caused by: java.io.IOException: error=12, Not enough space
>    [junit4]   2>      at java.lang.UNIXProcess.forkAndExec(Native Method)
>    [junit4]   2>      at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
>    [junit4]   2>      at java.lang.ProcessImpl.start(ProcessImpl.java:130)
>    [junit4]   2>      at 
> java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
>    [junit4]   2>      ... 43 more
>    [junit4]   2> 2637287 WARN
> (org.apache.hadoop.util.JvmPauseMonitor$Monitor@be51b7) [    ]
> o.a.h.u.JvmPauseMonitor Detected pause in JVM or host machine (eg GC):
> pause of approximately 15969ms
>    [junit4]   2> No GCs detected
>    [junit4]   2> 2637368 WARN  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.h.h.HttpRequestLog Jetty
> request log can only be enabled using Log4j
>    [junit4]   2> 2637384 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log jetty-6.1.26
>    [junit4]   2> 2637422 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log Extract
> jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-
> hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/datanode to
> ./temp/Jetty_solaris.vm_49465_datanode____96t731/webapp
>    [junit4]   2> 2637655 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log NO JSP Support for /, did not
> find org.apache.jasper.servlet.JspServlet
>    [junit4]   2> 2638756 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log Started
> HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:49465
>    [junit4]   2> 2645079 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log Stopped
> HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
>    [junit4]   2> 2645234 ERROR (DataNode:
> [[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-
> Solaris/solr/build/solr-
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> 001/tempDir-001/hdfsBaseDir/data/data1/,
> [DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-
> Solaris/solr/build/solr-
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> 001/tempDir-001/hdfsBaseDir/data/data2/]]  heartbeating to solaris-
> vm/127.0.0.1:61051) [    ] o.a.h.h.s.d.DataNode Initialization failed for 
> Block
> pool <registering> (Datanode Uuid unassigned) service to solaris-
> vm/127.0.0.1:61051. Exiting.
>    [junit4]   2> java.io.IOException: DN shut down before block pool
> connected
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.datanode.BPServiceActor.retrieveNamespac
> eInfo(BPServiceActor.java:185)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAnd
> Handshake(BPServiceActor.java:215)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceAct
> or.java:828)
>    [junit4]   2>      at java.lang.Thread.run(Thread.java:745)
>    [junit4]   2> 2645236 WARN  (DataNode:
> [[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-
> Solaris/solr/build/solr-
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> 001/tempDir-001/hdfsBaseDir/data/data1/,
> [DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-
> Solaris/solr/build/solr-
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> 001/tempDir-001/hdfsBaseDir/data/data2/]]  heartbeating to solaris-
> vm/127.0.0.1:61051) [    ] o.a.h.h.s.d.DataNode Ending block pool service for:
> Block pool <registering> (Datanode Uuid unassigned) service to solaris-
> vm/127.0.0.1:61051
>    [junit4]   2> 2645259 WARN
> (org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager
> $Monitor@7b7964) [    ] o.a.h.h.s.b.DecommissionManager Monitor
> interrupted: java.lang.InterruptedException: sleep interrupted
>    [junit4]   2> 2645314 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log Stopped
> HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
>    [junit4]   2> 2645418 ERROR (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.h.m.l.MethodMetric Error
> invoking method getBlocksTotal
>    [junit4]   2> java.lang.reflect.InvocationTargetException
>    [junit4]   2>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
>    [junit4]   2>      at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> ava:57)
>    [junit4]   2>      at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> sorImpl.java:43)
>    [junit4]   2>      at java.lang.reflect.Method.invoke(Method.java:606)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.lib.MethodMetric$2.snapshot(MethodMetric.j
> ava:111)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.lib.MethodMetric.snapshot(MethodMetric.jav
> a:144)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.lib.MetricsRegistry.snapshot(MetricsRegistry.ja
> va:387)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.lib.MetricsSourceBuilder$1.getMetrics(MetricsS
> ourceBuilder.java:79)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMetrics(Metrics
> SourceAdapter.java:195)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.updateJmxCache(
> MetricsSourceAdapter.java:172)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMBeanInfo(Me
> tricsSourceAdapter.java:151)
>    [junit4]   2>      at
> com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getClassName(De
> faultMBeanServerInterceptor.java:1804)
>    [junit4]   2>      at
> com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.safeGetClassNam
> e(DefaultMBeanServerInterceptor.java:1595)
>    [junit4]   2>      at
> com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.checkMBeanPer
> mission(DefaultMBeanServerInterceptor.java:1813)
>    [junit4]   2>      at
> com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.exclusiveUnregist
> erMBean(DefaultMBeanServerInterceptor.java:430)
>    [junit4]   2>      at
> com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.unregisterMBean
> (DefaultMBeanServerInterceptor.java:415)
>    [junit4]   2>      at
> com.sun.jmx.mbeanserver.JmxMBeanServer.unregisterMBean(JmxMBeanS
> erver.java:546)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.util.MBeans.unregister(MBeans.java:81)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stopMBeans(Metri
> csSourceAdapter.java:227)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stop(MetricsSourc
> eAdapter.java:212)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stopSources(MetricsS
> ystemImpl.java:461)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stop(MetricsSystemIm
> pl.java:212)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl.shutdown(MetricsSyst
> emImpl.java:592)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdownInstance(D
> efaultMetricsSystem.java:72)
>    [junit4]   2>      at
> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdown(DefaultMe
> tricsSystem.java:68)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics.shut
> down(NameNodeMetrics.java:145)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.namenode.NameNode.stop(NameNode.jav
> a:822)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:172
> 0)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:169
> 9)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> ava:838)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
>    [junit4]   2>      at
> org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
>    [junit4]   2>      at
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> est.java:44)
>    [junit4]   2>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
>    [junit4]   2>      at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> ava:57)
>    [junit4]   2>      at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> sorImpl.java:43)
>    [junit4]   2>      at java.lang.reflect.Method.invoke(Method.java:606)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> dRunner.java:1627)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> mizedRunner.java:776)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> mizedRunner.java:792)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> evaluate(SystemPropertiesRestoreRule.java:57)
>    [junit4]   2>      at
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> fterRule.java:46)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> assName.java:42)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> ertionsRequired.java:54)
>    [junit4]   2>      at
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> .java:48)
>    [junit4]   2>      at
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> IgnoreAfterMaxFailures.java:65)
>    [junit4]   2>      at
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> TestSuites.java:55)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> ementAdapter.java:36)
>    [junit4]   2>      at
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> run(ThreadLeakControl.java:365)
>    [junit4]   2>      at java.lang.Thread.run(Thread.java:745)
>    [junit4]   2> Caused by: java.lang.NullPointerException
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.blockmanagement.BlocksMap.size(BlocksM
> ap.java:198)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.getTotalBl
> ocks(BlockManager.java:3291)
>    [junit4]   2>      at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlocksTotal(F
> SNamesystem.java:6223)
>    [junit4]   2>      ... 58 more
>    [junit4]   2> 2645432 INFO  (SUITE-HdfsNNFailoverTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.s.SolrTestCaseJ4 ###deleteCore
>    [junit4]   2> NOTE: test params are: codec=Asserting(Lucene53),
> sim=RandomSimilarityProvider(queryNorm=true,coord=yes): {},
> locale=mk_MK, timezone=Asia/Shanghai
>    [junit4]   2> NOTE: SunOS 5.11 x86/Oracle Corporation 1.7.0_85 (32-
> bit)/cpus=3,threads=1,free=99794816,total=518979584
>    [junit4]   2> NOTE: All tests run in this JVM: [SolrCloudExampleTest,
> TestStressVersions, TestSerializedLuceneMatchVersion, TestSolrJ,
> DistanceUnitsTest, MultiThreadedOCPTest, TestDistribDocBasedVersion,
> BJQParserTest, ZkCLITest, QueryEqualityTest, PrimitiveFieldTypeTest,
> DistributedQueryComponentOptimizationTest, AliasIntegrationTest,
> TestInitQParser, TestAuthorizationFramework, TestLazyCores,
> SolrIndexConfigTest, TestFunctionQuery, TestXIncludeConfig,
> HardAutoCommitTest, DocValuesMultiTest, TestDefaultStatsCache,
> SolrRequestParserTest, RecoveryZkTest, UpdateParamsTest,
> TestSolrDeletionPolicy1, TestDFRSimilarityFactory, TestFastWriter,
> PathHierarchyTokenizerFactoryTest, TestDynamicLoading,
> TestElisionMultitermQuery, PolyFieldTest, UnloadDistributedZkTest,
> TestJsonRequest, TestRuleBasedAuthorizationPlugin,
> TestManagedStopFilterFactory, TestRawResponseWriter, IndexSchemaTest,
> TestEmbeddedSolrServerConstructors, InfoHandlerTest,
> AlternateDirectoryTest, LeaderElectionTest, JsonLoaderTest,
> TestCoreContainer, DirectSolrSpellCheckerTest, RequestLoggingTest,
> ZkNodePropsTest, TermsComponentTest, TestConfig,
> TestFieldTypeCollectionResource, XsltUpdateRequestHandlerTest,
> TestManagedSchemaFieldResource, TestSchemaResource,
> DataDrivenBlockJoinTest, TestExactStatsCache, TestConfigSetProperties,
> DeleteLastCustomShardedReplicaTest, TestAnalyzedSuggestions,
> DirectUpdateHandlerTest, ExternalFileFieldSortTest, TestIBSimilarityFactory,
> TestMissingGroups, ClusterStateUpdateTest, ActionThrottleTest,
> QueryElevationComponentTest, DocValuesTest, QueryResultKeyTest,
> TestLRUCache, TestPhraseSuggestions, SimplePostToolTest,
> TriLevelCompositeIdRoutingTest, DistributedMLTComponentTest,
> CloudExitableDirectoryReaderTest, TestSolrCloudWithKerberosAlt,
> TestCodecSupport, TestConfigSets, PeerSyncTest,
> XmlUpdateRequestHandlerTest, SpatialHeatmapFacetsTest,
> SoftAutoCommitTest, TestSchemaNameResource,
> PreAnalyzedUpdateProcessorTest, TestJmxMonitoredMap,
> TestDistributedStatsComponentCardinality,
> TestManagedSynonymFilterFactory, JSONWriterTest, TestNRTOpen,
> ReplicationFactorTest, DOMUtilTest, SolrCoreTest,
> DocExpirationUpdateProcessorFactoryTest, FastVectorHighlighterTest,
> SuggesterFSTTest, TestExtendedDismaxParser, TestSolrConfigHandler,
> DocumentAnalysisRequestHandlerTest,
> DistributedFacetPivotSmallAdvancedTest, BlockDirectoryTest,
> TestQuerySenderNoQuery, TestHashPartitioner, DateFieldTest,
> SegmentsInfoRequestHandlerTest, TestFieldCollectionResource,
> RecoveryAfterSoftCommitTest, TestMergePolicyConfig, TestFieldSortValues,
> SecurityConfHandlerTest, TestStressReorder, BufferStoreTest,
> TestRandomRequestDistribution, HdfsBasicDistributedZkTest,
> TestCloudManagedSchemaConcurrent, TestReplicaProperties,
> DisMaxRequestHandlerTest, TestMacros, TestStressLucene,
> TestReloadAndDeleteDocs, BasicAuthIntegrationTest, TestDocSet,
> BasicDistributedZkTest, DistributedQueryElevationComponentTest,
> TestGroupingSearch, TestObjectReleaseTracker, MoreLikeThisHandlerTest,
> OverseerTest, TestFaceting, TestUpdate, TestClassNameShortening,
> TestRestManager, SyncSliceTest, ShardRoutingTest, ZkSolrClientTest,
> TestZkChroot, TestRandomDVFaceting, ShardRoutingCustomTest,
> TestDistributedGrouping, DistributedSpellCheckComponentTest,
> ZkControllerTest, TestRealTimeGet, TestReload,
> DistributedTermsComponentTest, TestRangeQuery, SimpleFacetsTest,
> TestSolr4Spatial, StatsComponentTest, SolrCmdDistributorTest, TestSort,
> CurrencyFieldXmlFileTest, AnalysisAfterCoreReloadTest,
> TestFoldingMultitermQuery, SuggesterTSTTest, TestCSVLoader,
> SchemaVersionSpecificBehaviorTest, SolrCoreCheckLockOnStartupTest,
> DirectUpdateHandlerOptimizeTest,
> StatelessScriptUpdateProcessorFactoryTest, DistanceFunctionTest,
> IndexBasedSpellCheckerTest, StandardRequestHandlerTest,
> TestOmitPositions, DocumentBuilderTest, RequiredFieldsTest,
> TestArbitraryIndexDir, LoggingHandlerTest, ReturnFieldsTest,
> MBeansHandlerTest, UniqFieldsUpdateProcessorFactoryTest,
> PingRequestHandlerTest, TestComponentsName, TestLFUCache,
> PreAnalyzedFieldTest, TestSystemIdResolver, SpellingQueryConverterTest,
> TestUtils, TestDocumentBuilder, SliceStateTest, SystemInfoHandlerTest,
> UUIDFieldTest, FileUtilsTest, CircularListTest, TestRTGBase,
> CursorPagingTest, DistributedIntervalFacetingTest,
> TestDistributedMissingSort, TestSimpleTrackingShardHandler,
> AsyncMigrateRouteKeyTest, DeleteInactiveReplicaTest,
> DistribDocExpirationUpdateProcessorTest, LeaderFailoverAfterPartitionTest,
> OverriddenZkACLAndCredentialsProvidersTest,
> OverseerCollectionConfigSetProcessorTest, OverseerRolesTest,
> OverseerTaskQueueTest, SSLMigrationTest, SaslZkACLProviderTest,
> SimpleCollectionCreateDeleteTest, TestAuthenticationFramework,
> TestCloudInspectUtil, TestCollectionAPI, TestMiniSolrCloudClusterSSL,
> TestRebalanceLeaders, TestRequestStatusCollectionAPI,
> HdfsBasicDistributedZk2Test, HdfsChaosMonkeySafeLeaderTest,
> HdfsCollectionsAPIDistributedZkTest, HdfsNNFailoverTest]
>    [junit4]   2> NOTE: reproduce with: ant test  -
> Dtestcase=HdfsNNFailoverTest -Dtests.seed=5D8F351977870E3F -
> Dtests.slow=true -Dtests.locale=mk_MK -Dtests.timezone=Asia/Shanghai -
> Dtests.asserts=true -Dtests.file.encoding=UTF-8
>    [junit4] ERROR   0.00s J0 | HdfsNNFailoverTest (suite) <<<
>    [junit4]    > Throwable #1: java.lang.RuntimeException: Error while running
> command to get file permissions : java.io.IOException: Cannot run program
> "/bin/ls": error=12, Not enough space
>    [junit4]    >      at 
> java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
>    [junit4]    >      at 
> org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
>    [junit4]    >      at org.apache.hadoop.util.Shell.run(Shell.java:455)
>    [junit4]    >      at
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715
> )
>    [junit4]    >      at
> org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
>    [junit4]    >      at
> org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
>    [junit4]    >      at
> org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)
>    [junit4]    >      at
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo
> adPermissionInfo(RawLocalFileSystem.java:582)
>    [junit4]    >      at
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g
> etPermission(RawLocalFileSystem.java:557)
>    [junit4]    >      at
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> DiskChecker.java:139)
>    [junit4]    >      at
> org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> checkDir(DataNode.java:2239)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> DataNode.java:2281)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> de.java:2263)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> ataNode.java:2155)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> a:1443)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> ava:828)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
>    [junit4]    >      at
> org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
>    [junit4]    >      at
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> est.java:44)
>    [junit4]    >      at java.lang.Thread.run(Thread.java:745)
>    [junit4]    > Caused by: java.io.IOException: error=12, Not enough space
>    [junit4]    >      at java.lang.UNIXProcess.forkAndExec(Native Method)
>    [junit4]    >      at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
>    [junit4]    >      at java.lang.ProcessImpl.start(ProcessImpl.java:130)
>    [junit4]    >      at 
> java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
>    [junit4]    >      ... 44 more
>    [junit4]    >      at
> __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
>    [junit4]    >      at
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo
> adPermissionInfo(RawLocalFileSystem.java:620)
>    [junit4]    >      at
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g
> etPermission(RawLocalFileSystem.java:557)
>    [junit4]    >      at
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> DiskChecker.java:139)
>    [junit4]    >      at
> org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> checkDir(DataNode.java:2239)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> DataNode.java:2281)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> de.java:2263)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> ataNode.java:2155)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> a:1443)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> ava:828)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
>    [junit4]    >      at
> org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
>    [junit4]    >      at
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> est.java:44)
>    [junit4]    >      at java.lang.Thread.run(Thread.java:745)
>    [junit4] Completed [426/536] on J0 in 45.66s, 0 tests, 1 error <<< 
> FAILURES!
> 
> [...truncated 300 lines...]
>    [junit4] Suite: org.apache.solr.store.hdfs.HdfsDirectoryTest
>    [junit4]   2> Creating dataDir: /export/home/jenkins/workspace/Lucene-
> Solr-5.x-Solaris/solr/build/solr-
> core/test/J1/temp/solr.store.hdfs.HdfsDirectoryTest_5D8F351977870E3F-
> 001/init-core-data-001
>    [junit4]   2> 3147821 INFO  (SUITE-HdfsDirectoryTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized ssl
> (false) and clientAuth (false)
>    [junit4]   1> Formatting using clusterid: testClusterID
>    [junit4]   2> 3147964 WARN  (SUITE-HdfsDirectoryTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.h.m.i.MetricsConfig Cannot
> locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-
> metrics2.properties
>    [junit4]   2> 3147974 WARN  (SUITE-HdfsDirectoryTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.h.h.HttpRequestLog Jetty
> request log can only be enabled using Log4j
>    [junit4]   2> 3147976 INFO  (SUITE-HdfsDirectoryTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log jetty-6.1.26
>    [junit4]   2> 3147994 INFO  (SUITE-HdfsDirectoryTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log Extract
> jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-
> hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/hdfs to
> ./temp/Jetty_solaris.vm_46547_hdfs____.vwfmpk/webapp
>    [junit4]   2> 3148170 INFO  (SUITE-HdfsDirectoryTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log NO JSP Support for /, did not
> find org.apache.jasper.servlet.JspServlet
>    [junit4]   2> 3148982 INFO  (SUITE-HdfsDirectoryTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log Started
> HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:46547
>    [junit4]   2> 3157264 INFO  (SUITE-HdfsDirectoryTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.m.log Stopped
> HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
>    [junit4]   2> 3157403 INFO  (SUITE-HdfsDirectoryTest-
> seed#[5D8F351977870E3F]-worker) [    ] o.a.s.SolrTestCaseJ4 ###deleteCore
>    [junit4]   2> Aug 29, 2015 9:01:07 PM
> com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
>    [junit4]   2> WARNING: Will linger awaiting termination of 1 leaked
> thread(s).
>    [junit4]   2> Aug 29, 2015 9:01:27 PM
> com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
>    [junit4]   2> SEVERE: 1 thread leaked from SUITE scope at
> org.apache.solr.store.hdfs.HdfsDirectoryTest:
>    [junit4]   2>    1) Thread[id=20389, name=IPC Server idle connection 
> scanner
> for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
>    [junit4]   2>         at java.lang.Object.wait(Native Method)
>    [junit4]   2>         at java.lang.Object.wait(Object.java:503)
>    [junit4]   2>         at java.util.TimerThread.mainLoop(Timer.java:526)
>    [junit4]   2>         at java.util.TimerThread.run(Timer.java:505)
>    [junit4]   2> Aug 29, 2015 9:01:27 PM
> com.carrotsearch.randomizedtesting.ThreadLeakControl tryToInterruptAll
>    [junit4]   2> INFO: Starting to interrupt leaked threads:
>    [junit4]   2>    1) Thread[id=20389, name=IPC Server idle connection 
> scanner
> for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
>    [junit4]   2> Aug 29, 2015 9:01:30 PM
> com.carrotsearch.randomizedtesting.ThreadLeakControl tryToInterruptAll
>    [junit4]   2> SEVERE: There are still zombie threads that couldn't be
> terminated:
>    [junit4]   2>    1) Thread[id=20389, name=IPC Server idle connection 
> scanner
> for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
>    [junit4]   2>         at java.lang.Object.wait(Native Method)
>    [junit4]   2>         at java.lang.Object.wait(Object.java:503)
>    [junit4]   2>         at java.util.TimerThread.mainLoop(Timer.java:526)
>    [junit4]   2>         at java.util.TimerThread.run(Timer.java:505)
>    [junit4]   2> NOTE: test params are: codec=Asserting(Lucene53): {},
> docValues:{}, sim=DefaultSimilarity, locale=es_BO,
> timezone=Antarctica/South_Pole
>    [junit4]   2> NOTE: SunOS 5.11 x86/Oracle Corporation 1.7.0_85 (32-
> bit)/cpus=3,threads=2,free=136627544,total=518979584
>    [junit4]   2> NOTE: All tests run in this JVM: [TestIndexingPerformance,
> TestCSVResponseWriter, DistributedQueryComponentCustomSortTest,
> DirectSolrConnectionTest, FullSolrCloudDistribCmdsTest,
> TestShardHandlerFactory, CacheHeaderTest, BasicZkTest, TestTrie,
> FieldAnalysisRequestHandlerTest, PKIAuthenticationIntegrationTest,
> OpenCloseCoreStressTest, TestSuggestSpellingConverter, StressHdfsTest,
> CleanupOldIndexTest, DistributedExpandComponentTest,
> TestHdfsUpdateLog, TestSolrXml, TestAddFieldRealTimeGet, TestJsonFacets,
> DistributedSuggestComponentTest,
> OutOfBoxZkACLAndCredentialsProvidersTest, AnalyticsMergeStrategyTest,
> HLLUtilTest, ResponseHeaderTest, SearchHandlerTest,
> BinaryUpdateRequestHandlerTest, DistributedFacetPivotWhiteBoxTest,
> ConnectionManagerTest, SpellCheckComponentTest,
> TestScoreJoinQPNoScore, SolrTestCaseJ4Test, SolrIndexSplitterTest,
> TestConfigSetsAPI, TestDefaultSearchFieldResource, TestCryptoKeys,
> TestNonDefinedSimilarityFactory, TestCoreDiscovery, RollingRestartTest,
> SolrInfoMBeanTest, CustomCollectionTest, DistributedVersionInfoTest,
> ClusterStateTest, TestReversedWildcardFilterFactory, SolrXmlInZkTest,
> DistributedFacetPivotLongTailTest, URLClassifyProcessorTest,
> TestLMJelinekMercerSimilarityFactory, RequestHandlersTest,
> RemoteQueryErrorTest, LeaderElectionIntegrationTest,
> SharedFSAutoReplicaFailoverTest, TestBadConfig,
> SignatureUpdateProcessorFactoryTest, TestCursorMarkWithoutUniqueKey,
> TestCrossCoreJoin, SparseHLLTest, DistributedQueueTest,
> BigEndianAscendingWordSerializerTest, TestBM25SimilarityFactory,
> AutoCommitTest, DateMathParserTest, BasicFunctionalityTest,
> SuggesterWFSTTest, TestCollapseQParserPlugin, TestManagedResource,
> TestSha256AuthenticationProvider, CollectionTooManyReplicasTest,
> BadCopyFieldTest, TestDownShardTolerantSearch, CloudMLTQParserTest,
> NotRequiredUniqueKeyTest, TestAnalyzeInfixSuggestions,
> ExitableDirectoryReaderTest, TestScoreJoinQPScore, DeleteShardTest,
> RankQueryTest, TestSchemaManager, UpdateRequestProcessorFactoryTest,
> CursorMarkTest, DistributedDebugComponentTest, DeleteReplicaTest,
> RAMDirectoryFactoryTest, ConcurrentDeleteAndCreateCollectionTest,
> TestQueryTypes, OutputWriterTest, TestSchemaSimilarityResource,
> HighlighterMaxOffsetTest, ResponseLogComponentTest,
> TestCloudPivotFacet, DocValuesMissingTest,
> FieldMutatingUpdateProcessorTest, HttpPartitionTest, TestCollationField,
> ZkStateWriterTest, TestQuerySenderListener, AtomicUpdatesTest,
> TestStressRecovery, TestRandomFaceting,
> SharedFSAutoReplicaFailoverUtilsTest, CoreAdminHandlerTest,
> HighlighterConfigTest, TestCustomSort, MultiTermTest,
> VMParamsZkACLAndCredentialsProvidersTest,
> IgnoreCommitOptimizeUpdateProcessorFactoryTest, CollectionReloadTest,
> PrimUtilsTest, TestRecovery, TestWriterPerf,
> AddSchemaFieldsUpdateProcessorFactoryTest, TimeZoneUtilsTest,
> CurrencyFieldOpenExchangeTest, TestSolrCLIRunExample,
> TestPHPSerializedResponseWriter, ChaosMonkeySafeLeaderTest,
> TestIndexSearcher, EnumFieldTest, TestSolrIndexConfig,
> TermVectorComponentDistributedTest, TestJoin, TestExpandComponent,
> TestManagedResourceStorage, SortByFunctionTest,
> TestDefaultSimilarityFactory, SuggesterTest, TestValueSourceCache,
> SolrPluginUtilsTest, TermVectorComponentTest, TestFiltering,
> TestQueryUtils, FileBasedSpellCheckerTest, BasicDistributedZk2Test,
> CollectionsAPIDistributedZkTest, TestReplicationHandler,
> TestDistributedSearch, BadIndexSchemaTest, ConvertedLegacyTest,
> HighlighterTest, ShowFileRequestHandlerTest, SpellCheckCollatorTest,
> SpatialFilterTest, NoCacheHeaderTest, WordBreakSolrSpellCheckerTest,
> TestPseudoReturnFields, TestAtomicUpdateErrorCases,
> TestWordDelimiterFilterFactory, DefaultValueUpdateProcessorTest,
> TestRemoteStreaming, DebugComponentTest, TestSurroundQueryParser,
> LukeRequestHandlerTest, TestSolrQueryParser,
> IndexSchemaRuntimeFieldTest, RegexBoostProcessorTest,
> TestJmxIntegration, QueryParsingTest, TestPartialUpdateDeduplication,
> CSVRequestHandlerTest, TestBinaryResponseWriter, SOLR749Test,
> CopyFieldTest, BadComponentTest, TestSolrDeletionPolicy2, SampleTest,
> TestBinaryField, TestSearchPerf, NumericFieldsTest, MinimalSchemaTest,
> TestFuzzyAnalyzedSuggestions, TestSolrCoreProperties,
> TestPostingsSolrHighlighter, TestLuceneMatchVersion,
> SpellPossibilityIteratorTest, TestCharFilters, SynonymTokenizerTest,
> EchoParamsTest, TestSweetSpotSimilarityFactory, TestPerFieldSimilarity,
> TestLMDirichletSimilarityFactory, ResourceLoaderTest,
> TestFastOutputStream, ScriptEngineTest,
> OpenExchangeRatesOrgProviderTest, PluginInfoTest, TestFastLRUCache,
> ChaosMonkeyNothingIsSafeTest, TestHighlightDedupGrouping,
> TestTolerantSearch, TestJettySolrRunner, AssignTest,
> AsyncCallRequestStatusResponseTest, CollectionStateFormat2Test,
> CollectionsAPIAsyncDistributedZkTest, DistribCursorPagingTest,
> DistribJoinFromCollectionTest, LeaderInitiatedRecoveryOnCommitTest,
> MigrateRouteKeyTest, OverseerStatusTest, ShardSplitTest,
> TestConfigSetsAPIExclusivity, TestConfigSetsAPIZkFailure,
> TestLeaderElectionZkExpiry, TestMiniSolrCloudCluster,
> TestShortCircuitedRequests, HdfsRecoverLeaseTest,
> CachingDirectoryFactoryTest, HdfsDirectoryFactoryTest, TestConfigOverlay,
> TestConfigSetImmutable, TestImplicitCoreProperties,
> TestInfoStreamLogging, TestInitParams, TestSolrDynamicMBean,
> TestBlobHandler, TestConfigReload, TestReplicationHandlerBackup,
> TestSolrConfigHandlerConcurrent, CoreAdminCreateDiscoverTest,
> CoreAdminRequestStatusTest, CoreMergeIndexesAdminHandlerTest,
> DistributedFacetPivotLargeTest, DistributedFacetPivotSmallTest,
> FacetPivotSmallTest, SuggestComponentTest, JavabinLoaderTest,
> SmileWriterTest, TestIntervalFaceting, TestChildDocTransformer,
> TestCustomDocTransformer, TestSortingResponseWriter,
> TestBulkSchemaAPI, TestFieldResource,
> TestManagedSchemaDynamicFieldResource, TestBulkSchemaConcurrent,
> TestCloudSchemaless, TestReloadDeadlock, TestSearcherReuse,
> TestSimpleQParserPlugin, TestSmileRequest, TestSolr4Spatial2,
> TestStandardQParsers, TestStressUserVersions, TestTrieFacet,
> TestMinMaxOnMultiValuedField, TestOrdValues,
> TestSortByMinMaxFunction, SimpleMLTQParserTest, TestDistribIDF,
> TestExactSharedStatsCache, TestPKIAuthenticationPlugin,
> TestBlendedInfixSuggestions, TestFileDictionaryLookup,
> TestFreeTextSuggestions, TestHighFrequencyDictionaryFactory,
> BlockCacheTest, HdfsDirectoryTest]
>    [junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=HdfsDirectoryTest 
> -
> Dtests.seed=5D8F351977870E3F -Dtests.slow=true -Dtests.locale=es_BO -
> Dtests.timezone=Antarctica/South_Pole -Dtests.asserts=true -
> Dtests.file.encoding=UTF-8
>    [junit4] ERROR   0.00s J1 | HdfsDirectoryTest (suite) <<<
>    [junit4]    > Throwable #1: java.security.AccessControlException: access
> denied ("java.io.FilePermission" "/export/home/jenkins/workspace/Lucene-
> Solr-5.x-Solaris/solr/build/solr-core/test/J1" "write")
>    [junit4]    >      at
> __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
>    [junit4]    >      at
> java.security.AccessControlContext.checkPermission(AccessControlContext.j
> ava:395)
>    [junit4]    >      at
> java.security.AccessController.checkPermission(AccessController.java:559)
>    [junit4]    >      at
> java.lang.SecurityManager.checkPermission(SecurityManager.java:549)
>    [junit4]    >      at
> java.lang.SecurityManager.checkWrite(SecurityManager.java:979)
>    [junit4]    >      at java.io.File.canWrite(File.java:785)
>    [junit4]    >      at 
> org.apache.hadoop.fs.FileUtil.canWrite(FileUtil.java:1002)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.createPermissionsDiagnosisString(
> MiniDFSCluster.java:856)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> ava:812)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
>    [junit4]    >      at
> org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
>    [junit4]    >      at
> org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
>    [junit4]    >      at
> org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:60)
>    [junit4]    >      at
> org.apache.solr.store.hdfs.HdfsDirectoryTest.beforeClass(HdfsDirectoryTest.
> java:62)
>    [junit4]    >      at java.lang.Thread.run(Thread.java:745)Throwable #2:
> com.carrotsearch.randomizedtesting.ThreadLeakError: 1 thread leaked from
> SUITE scope at org.apache.solr.store.hdfs.HdfsDirectoryTest:
>    [junit4]    >    1) Thread[id=20389, name=IPC Server idle connection 
> scanner
> for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
>    [junit4]    >         at java.lang.Object.wait(Native Method)
>    [junit4]    >         at java.lang.Object.wait(Object.java:503)
>    [junit4]    >         at java.util.TimerThread.mainLoop(Timer.java:526)
>    [junit4]    >         at java.util.TimerThread.run(Timer.java:505)
>    [junit4]    >      at
> __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)Throwable #3:
> com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie
> threads that couldn't be terminated:
>    [junit4]    >    1) Thread[id=20389, name=IPC Server idle connection 
> scanner
> for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
>    [junit4]    >         at java.lang.Object.wait(Native Method)
>    [junit4]    >         at java.lang.Object.wait(Object.java:503)
>    [junit4]    >         at java.util.TimerThread.mainLoop(Timer.java:526)
>    [junit4]    >         at java.util.TimerThread.run(Timer.java:505)
>    [junit4]    >      at
> __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
>    [junit4] Completed [521/536] on J1 in 33.05s, 0 tests, 3 errors <<< 
> FAILURES!
> 
> [...truncated 64 lines...]
> BUILD FAILED
> /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:785:
> The following error occurred while executing this line:
> /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:729:
> The following error occurred while executing this line:
> /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:59: The
> following error occurred while executing this line:
> /export/home/jenkins/workspace/Lucene-Solr-5.x-
> Solaris/solr/build.xml:233: The following error occurred while executing this
> line:
> /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/common-
> build.xml:524: The following error occurred while executing this line:
> /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/lucene/common-
> build.xml:1452: The following error occurred while executing this line:
> /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/lucene/common-
> build.xml:1006: There were test failures: 536 suites, 2123 tests, 4 
> suite-level
> errors, 108 ignored (34 assumptions)
> 
> Total time: 77 minutes 51 seconds
> Build step 'Invoke Ant' marked build as failure
> Archiving artifacts
> [WARNINGS] Skipping publisher since build result is FAILURE
> Recording test results
> Email was triggered for: Failure - Any
> Sending email for trigger: Failure - Any
> 



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to