Hi,

This is a problem of Java 7 on Solaris. It still uses fork to spawn processes 
in Java 7; Java 8 has fixed this. Unfortunately the whole fork stuff is badly 
implemented in Solaris and really allocates the same size of memory again and 
does not use it - due to the large heaps this needs a lot of memory. Workaround 
is to allocate enough Swap (which is never used):

https://developer.opencloud.com/forum/posts/list/620.page

For now I raised swap space (which is really simple to do with ZFS... Way cool):

root@solaris-vm:~# zfs set volsize=6g rpool/swap

Super cool.

In Java 8, it uses the new posix_spawn launch mechanism... (on Linux vfork). 
This was (by the way) the buggy code that was fixed in U40, which caused the 
Turkish Locale to fail :-)

Uwe

-----
Uwe Schindler
H.-H.-Meier-Allee 63, D-28213 Bremen
http://www.thetaphi.de
eMail: [email protected]


> -----Original Message-----
> From: Uwe Schindler [mailto:[email protected]]
> Sent: Saturday, August 29, 2015 11:52 PM
> To: [email protected]
> Subject: RE: [JENKINS] Lucene-Solr-5.x-Solaris (multiarch/jdk1.7.0) - Build # 
> 9
> - Still Failing!
> 
> I am still digging... On Solaris there seems to be a general forking problem 
> on
> 32 bit processes.
> 
> Uwe
> 
> -----
> Uwe Schindler
> H.-H.-Meier-Allee 63, D-28213 Bremen
> http://www.thetaphi.de
> eMail: [email protected]
> 
> > -----Original Message-----
> > From: Policeman Jenkins Server [mailto:[email protected]]
> > Sent: Saturday, August 29, 2015 11:02 PM
> > To: [email protected]; [email protected]; [email protected]
> > Subject: [JENKINS] Lucene-Solr-5.x-Solaris (multiarch/jdk1.7.0) - Build # 9 
> > -
> > Still Failing!
> >
> > Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Solaris/9/
> > Java: multiarch/jdk1.7.0 -d32 -server -XX:+UseConcMarkSweepGC
> >
> > 4 tests failed.
> > FAILED:
> > junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsNNFailoverTest
> >
> > Error Message:
> > Error while running command to get file permissions : java.io.IOException:
> > Cannot run program "/bin/ls": error=12, Not enough space  at
> > java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)  at
> > org.apache.hadoop.util.Shell.runCommand(Shell.java:485)  at
> > org.apache.hadoop.util.Shell.run(Shell.java:455)  at
> >
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715
> > )  at org.apache.hadoop.util.Shell.execCommand(Shell.java:808)  at
> > org.apache.hadoop.util.Shell.execCommand(Shell.java:791)  at
> > org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)  at
> >
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo
> > adPermissionInfo(RawLocalFileSystem.java:582)  at
> >
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g
> > etPermission(RawLocalFileSystem.java:557)  at
> >
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> > DiskChecker.java:139)  at
> > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)  at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> > checkDir(DataNode.java:2239)  at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> > DataNode.java:2281)  at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> > de.java:2263)  at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> > ataNode.java:2155)  at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> > a:1443)  at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> > ava:828)  at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
> at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
> at
> > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
> at
> >
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> > est.java:44)  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> > Method)  at
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> > ava:57)  at
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> > sorImpl.java:43)  at java.lang.reflect.Method.invoke(Method.java:606)  at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> > dRunner.java:1627)  at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> > mizedRunner.java:776)  at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> > mizedRunner.java:792)  at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)  at
> >
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> > evaluate(SystemPropertiesRestoreRule.java:57)  at
> >
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> > fterRule.java:46)  at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)  at
> >
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> > assName.java:42)  at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)  at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)  at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)  at
> >
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> > ertionsRequired.java:54)  at
> >
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> > .java:48)  at
> >
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> > IgnoreAfterMaxFailures.java:65)  at
> >
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> > TestSuites.java:55)  at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)  at
> >
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> > run(ThreadLeakControl.java:365)  at java.lang.Thread.run(Thread.java:745)
> > Caused by: java.io.IOException: error=12, Not enough space  at
> > java.lang.UNIXProcess.forkAndExec(Native Method)  at
> > java.lang.UNIXProcess.<init>(UNIXProcess.java:137)  at
> > java.lang.ProcessImpl.start(ProcessImpl.java:130)  at
> > java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)  ... 44 more
> >
> > Stack Trace:
> > java.lang.RuntimeException: Error while running command to get file
> > permissions : java.io.IOException: Cannot run program "/bin/ls": error=12,
> > Not enough space
> >     at java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
> >     at org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
> >     at org.apache.hadoop.util.Shell.run(Shell.java:455)
> >     at
> >
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715
> > )
> >     at org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
> >     at org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
> >     at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)
> >     at
> >
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo
> > adPermissionInfo(RawLocalFileSystem.java:582)
> >     at
> >
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g
> > etPermission(RawLocalFileSystem.java:557)
> >     at
> >
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> > DiskChecker.java:139)
> >     at
> > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
> >     at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> > checkDir(DataNode.java:2239)
> >     at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> > DataNode.java:2281)
> >     at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> > de.java:2263)
> >     at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> > ataNode.java:2155)
> >     at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> > a:1443)
> >     at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> > ava:828)
> >     at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
> >     at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
> >     at
> > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
> >     at
> >
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> > est.java:44)
> >     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >     at
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> > ava:57)
> >     at
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> > sorImpl.java:43)
> >     at java.lang.reflect.Method.invoke(Method.java:606)
> >     at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> > dRunner.java:1627)
> >     at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> > mizedRunner.java:776)
> >     at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> > mizedRunner.java:792)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> > evaluate(SystemPropertiesRestoreRule.java:57)
> >     at
> >
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> > fterRule.java:46)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> > assName.java:42)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> > ertionsRequired.java:54)
> >     at
> >
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> > .java:48)
> >     at
> >
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> > IgnoreAfterMaxFailures.java:65)
> >     at
> >
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> > TestSuites.java:55)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> > run(ThreadLeakControl.java:365)
> >     at java.lang.Thread.run(Thread.java:745)
> > Caused by: java.io.IOException: error=12, Not enough space
> >     at java.lang.UNIXProcess.forkAndExec(Native Method)
> >     at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
> >     at java.lang.ProcessImpl.start(ProcessImpl.java:130)
> >     at java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
> >     ... 44 more
> >
> >     at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
> >     at
> >
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo
> > adPermissionInfo(RawLocalFileSystem.java:620)
> >     at
> >
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g
> > etPermission(RawLocalFileSystem.java:557)
> >     at
> >
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> > DiskChecker.java:139)
> >     at
> > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
> >     at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> > checkDir(DataNode.java:2239)
> >     at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> > DataNode.java:2281)
> >     at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> > de.java:2263)
> >     at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> > ataNode.java:2155)
> >     at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> > a:1443)
> >     at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> > ava:828)
> >     at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
> >     at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
> >     at
> > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
> >     at
> >
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> > est.java:44)
> >     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >     at
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> > ava:57)
> >     at
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> > sorImpl.java:43)
> >     at java.lang.reflect.Method.invoke(Method.java:606)
> >     at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> > dRunner.java:1627)
> >     at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> > mizedRunner.java:776)
> >     at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> > mizedRunner.java:792)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> > evaluate(SystemPropertiesRestoreRule.java:57)
> >     at
> >
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> > fterRule.java:46)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> > assName.java:42)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> > ertionsRequired.java:54)
> >     at
> >
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> > .java:48)
> >     at
> >
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> > IgnoreAfterMaxFailures.java:65)
> >     at
> >
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> > TestSuites.java:55)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> > run(ThreadLeakControl.java:365)
> >     at java.lang.Thread.run(Thread.java:745)
> >
> >
> > FAILED:
> > junit.framework.TestSuite.org.apache.solr.store.hdfs.HdfsDirectoryTest
> >
> > Error Message:
> > access denied ("java.io.FilePermission"
> > "/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr-
> > core/test/J1" "write")
> >
> > Stack Trace:
> > java.security.AccessControlException: access denied
> > ("java.io.FilePermission" "/export/home/jenkins/workspace/Lucene-Solr-
> > 5.x-Solaris/solr/build/solr-core/test/J1" "write")
> >     at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
> >     at
> >
> java.security.AccessControlContext.checkPermission(AccessControlContext.j
> > ava:395)
> >     at
> > java.security.AccessController.checkPermission(AccessController.java:559)
> >     at
> > java.lang.SecurityManager.checkPermission(SecurityManager.java:549)
> >     at java.lang.SecurityManager.checkWrite(SecurityManager.java:979)
> >     at java.io.File.canWrite(File.java:785)
> >     at org.apache.hadoop.fs.FileUtil.canWrite(FileUtil.java:1002)
> >     at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.createPermissionsDiagnosisString(
> > MiniDFSCluster.java:856)
> >     at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> > ava:812)
> >     at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
> >     at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
> >     at
> > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
> >     at
> > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:60)
> >     at
> >
> org.apache.solr.store.hdfs.HdfsDirectoryTest.beforeClass(HdfsDirectoryTest.
> > java:62)
> >     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >     at
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> > ava:57)
> >     at
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> > sorImpl.java:43)
> >     at java.lang.reflect.Method.invoke(Method.java:606)
> >     at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> > dRunner.java:1627)
> >     at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> > mizedRunner.java:776)
> >     at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> > mizedRunner.java:792)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> > evaluate(SystemPropertiesRestoreRule.java:57)
> >     at
> >
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> > fterRule.java:46)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> > assName.java:42)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> > ertionsRequired.java:54)
> >     at
> >
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> > .java:48)
> >     at
> >
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> > IgnoreAfterMaxFailures.java:65)
> >     at
> >
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> > TestSuites.java:55)
> >     at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >     at
> >
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> > run(ThreadLeakControl.java:365)
> >     at java.lang.Thread.run(Thread.java:745)
> >
> >
> > FAILED:
> > junit.framework.TestSuite.org.apache.solr.store.hdfs.HdfsDirectoryTest
> >
> > Error Message:
> > 1 thread leaked from SUITE scope at
> > org.apache.solr.store.hdfs.HdfsDirectoryTest:     1) Thread[id=20389,
> > name=IPC Server idle connection scanner for port 41610, state=WAITING,
> > group=TGRP-HdfsDirectoryTest]         at java.lang.Object.wait(Native
> > Method)         at java.lang.Object.wait(Object.java:503)         at
> > java.util.TimerThread.mainLoop(Timer.java:526)         at
> > java.util.TimerThread.run(Timer.java:505)
> >
> > Stack Trace:
> > com.carrotsearch.randomizedtesting.ThreadLeakError: 1 thread leaked
> from
> > SUITE scope at org.apache.solr.store.hdfs.HdfsDirectoryTest:
> >    1) Thread[id=20389, name=IPC Server idle connection scanner for port
> > 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
> >         at java.lang.Object.wait(Native Method)
> >         at java.lang.Object.wait(Object.java:503)
> >         at java.util.TimerThread.mainLoop(Timer.java:526)
> >         at java.util.TimerThread.run(Timer.java:505)
> >     at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
> >
> >
> > FAILED:
> > junit.framework.TestSuite.org.apache.solr.store.hdfs.HdfsDirectoryTest
> >
> > Error Message:
> > There are still zombie threads that couldn't be terminated:    1)
> > Thread[id=20389, name=IPC Server idle connection scanner for port 41610,
> > state=WAITING, group=TGRP-HdfsDirectoryTest]         at
> > java.lang.Object.wait(Native Method)         at
> > java.lang.Object.wait(Object.java:503)         at
> > java.util.TimerThread.mainLoop(Timer.java:526)         at
> > java.util.TimerThread.run(Timer.java:505)
> >
> > Stack Trace:
> > com.carrotsearch.randomizedtesting.ThreadLeakError: There are still
> zombie
> > threads that couldn't be terminated:
> >    1) Thread[id=20389, name=IPC Server idle connection scanner for port
> > 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
> >         at java.lang.Object.wait(Native Method)
> >         at java.lang.Object.wait(Object.java:503)
> >         at java.util.TimerThread.mainLoop(Timer.java:526)
> >         at java.util.TimerThread.run(Timer.java:505)
> >     at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
> >
> >
> >
> >
> > Build Log:
> > [...truncated 10577 lines...]
> >    [junit4] Suite: org.apache.solr.cloud.hdfs.HdfsNNFailoverTest
> >    [junit4]   2> Creating dataDir: /export/home/jenkins/workspace/Lucene-
> > Solr-5.x-Solaris/solr/build/solr-
> >
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> > 001/init-core-data-001
> >    [junit4]   2> 2599844 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ]
> o.a.s.BaseDistributedSearchTestCase
> > Setting hostContext system property: /
> >    [junit4]   2> 2616331 WARN  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.a.h.u.NativeCodeLoader Unable
> to
> > load native-hadoop library for your platform... using builtin-java classes
> > where applicable
> >    [junit4]   1> Formatting using clusterid: testClusterID
> >    [junit4]   2> 2617524 WARN  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.a.h.m.i.MetricsConfig Cannot
> > locate configuration: tried hadoop-metrics2-
> namenode.properties,hadoop-
> > metrics2.properties
> >    [junit4]   2> 2617755 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log Logging to
> > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
> > org.mortbay.log.Slf4jLog
> >    [junit4]   2> 2617771 WARN  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.a.h.h.HttpRequestLog Jetty
> > request log can only be enabled using Log4j
> >    [junit4]   2> 2617878 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log jetty-6.1.26
> >    [junit4]   2> 2617942 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log Extract
> > jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-
> > hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/hdfs to
> > ./temp/Jetty_solaris.vm_35231_hdfs____thayv4/webapp
> >    [junit4]   2> 2618129 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log NO JSP Support for /, did
> not
> > find org.apache.jasper.servlet.JspServlet
> >    [junit4]   2> 2619464 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log Started
> > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:35231
> >    [junit4]   2> 2637264 WARN  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.a.h.h.s.d.DataNode Invalid
> > dfs.datanode.data.dir /export/home/jenkins/workspace/Lucene-Solr-5.x-
> > Solaris/solr/build/solr-
> >
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> > 001/tempDir-001/hdfsBaseDir/data/data2 :
> >    [junit4]   2> java.io.IOException: Cannot run program "chmod": error=12,
> > Not enough space
> >    [junit4]   2>    at 
> > java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
> >    [junit4]   2>    at 
> > org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
> >    [junit4]   2>    at org.apache.hadoop.util.Shell.run(Shell.java:455)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715
> > )
> >    [junit4]   2>    at
> > org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
> >    [junit4]   2>    at
> > org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSyste
> > m.java:656)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:4
> > 90)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> > DiskChecker.java:140)
> >    [junit4]   2>    at
> > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> > checkDir(DataNode.java:2239)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> > DataNode.java:2281)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> > de.java:2263)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> > ataNode.java:2155)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> > a:1443)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> > ava:828)
> >    [junit4]   2>    at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
> >    [junit4]   2>    at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
> >    [junit4]   2>    at
> > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
> >    [junit4]   2>    at
> >
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> > est.java:44)
> >    [junit4]   2>    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> > Method)
> >    [junit4]   2>    at
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> > ava:57)
> >    [junit4]   2>    at
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> > sorImpl.java:43)
> >    [junit4]   2>    at java.lang.reflect.Method.invoke(Method.java:606)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> > dRunner.java:1627)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> > mizedRunner.java:776)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> > mizedRunner.java:792)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> > evaluate(SystemPropertiesRestoreRule.java:57)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> > fterRule.java:46)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> > assName.java:42)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> > ertionsRequired.java:54)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> > .java:48)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> > IgnoreAfterMaxFailures.java:65)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> > TestSuites.java:55)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> > run(ThreadLeakControl.java:365)
> >    [junit4]   2>    at java.lang.Thread.run(Thread.java:745)
> >    [junit4]   2> Caused by: java.io.IOException: error=12, Not enough space
> >    [junit4]   2>    at java.lang.UNIXProcess.forkAndExec(Native Method)
> >    [junit4]   2>    at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
> >    [junit4]   2>    at java.lang.ProcessImpl.start(ProcessImpl.java:130)
> >    [junit4]   2>    at 
> > java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
> >    [junit4]   2>    ... 43 more
> >    [junit4]   2> 2637287 WARN
> > (org.apache.hadoop.util.JvmPauseMonitor$Monitor@be51b7) [    ]
> > o.a.h.u.JvmPauseMonitor Detected pause in JVM or host machine (eg GC):
> > pause of approximately 15969ms
> >    [junit4]   2> No GCs detected
> >    [junit4]   2> 2637368 WARN  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.a.h.h.HttpRequestLog Jetty
> > request log can only be enabled using Log4j
> >    [junit4]   2> 2637384 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log jetty-6.1.26
> >    [junit4]   2> 2637422 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log Extract
> > jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-
> > hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/datanode to
> > ./temp/Jetty_solaris.vm_49465_datanode____96t731/webapp
> >    [junit4]   2> 2637655 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log NO JSP Support for /, did
> not
> > find org.apache.jasper.servlet.JspServlet
> >    [junit4]   2> 2638756 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log Started
> > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:49465
> >    [junit4]   2> 2645079 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log Stopped
> > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
> >    [junit4]   2> 2645234 ERROR (DataNode:
> > [[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-
> > Solaris/solr/build/solr-
> >
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> > 001/tempDir-001/hdfsBaseDir/data/data1/,
> > [DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-
> > Solaris/solr/build/solr-
> >
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> > 001/tempDir-001/hdfsBaseDir/data/data2/]]  heartbeating to solaris-
> > vm/127.0.0.1:61051) [    ] o.a.h.h.s.d.DataNode Initialization failed for 
> > Block
> > pool <registering> (Datanode Uuid unassigned) service to solaris-
> > vm/127.0.0.1:61051. Exiting.
> >    [junit4]   2> java.io.IOException: DN shut down before block pool
> > connected
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.datanode.BPServiceActor.retrieveNamespac
> > eInfo(BPServiceActor.java:185)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAnd
> > Handshake(BPServiceActor.java:215)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceAct
> > or.java:828)
> >    [junit4]   2>    at java.lang.Thread.run(Thread.java:745)
> >    [junit4]   2> 2645236 WARN  (DataNode:
> > [[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-
> > Solaris/solr/build/solr-
> >
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> > 001/tempDir-001/hdfsBaseDir/data/data1/,
> > [DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x-
> > Solaris/solr/build/solr-
> >
> core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F-
> > 001/tempDir-001/hdfsBaseDir/data/data2/]]  heartbeating to solaris-
> > vm/127.0.0.1:61051) [    ] o.a.h.h.s.d.DataNode Ending block pool service
> for:
> > Block pool <registering> (Datanode Uuid unassigned) service to solaris-
> > vm/127.0.0.1:61051
> >    [junit4]   2> 2645259 WARN
> >
> (org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager
> > $Monitor@7b7964) [    ] o.a.h.h.s.b.DecommissionManager Monitor
> > interrupted: java.lang.InterruptedException: sleep interrupted
> >    [junit4]   2> 2645314 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log Stopped
> > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
> >    [junit4]   2> 2645418 ERROR (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.a.h.m.l.MethodMetric Error
> > invoking method getBlocksTotal
> >    [junit4]   2> java.lang.reflect.InvocationTargetException
> >    [junit4]   2>    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> > Method)
> >    [junit4]   2>    at
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> > ava:57)
> >    [junit4]   2>    at
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> > sorImpl.java:43)
> >    [junit4]   2>    at java.lang.reflect.Method.invoke(Method.java:606)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.lib.MethodMetric$2.snapshot(MethodMetric.j
> > ava:111)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.lib.MethodMetric.snapshot(MethodMetric.jav
> > a:144)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.lib.MetricsRegistry.snapshot(MetricsRegistry.ja
> > va:387)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.lib.MetricsSourceBuilder$1.getMetrics(MetricsS
> > ourceBuilder.java:79)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMetrics(Metrics
> > SourceAdapter.java:195)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.updateJmxCache(
> > MetricsSourceAdapter.java:172)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMBeanInfo(Me
> > tricsSourceAdapter.java:151)
> >    [junit4]   2>    at
> >
> com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getClassName(De
> > faultMBeanServerInterceptor.java:1804)
> >    [junit4]   2>    at
> >
> com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.safeGetClassNam
> > e(DefaultMBeanServerInterceptor.java:1595)
> >    [junit4]   2>    at
> > com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.checkMBeanPer
> > mission(DefaultMBeanServerInterceptor.java:1813)
> >    [junit4]   2>    at
> >
> com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.exclusiveUnregist
> > erMBean(DefaultMBeanServerInterceptor.java:430)
> >    [junit4]   2>    at
> >
> com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.unregisterMBean
> > (DefaultMBeanServerInterceptor.java:415)
> >    [junit4]   2>    at
> >
> com.sun.jmx.mbeanserver.JmxMBeanServer.unregisterMBean(JmxMBeanS
> > erver.java:546)
> >    [junit4]   2>    at
> > org.apache.hadoop.metrics2.util.MBeans.unregister(MBeans.java:81)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stopMBeans(Metri
> > csSourceAdapter.java:227)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stop(MetricsSourc
> > eAdapter.java:212)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stopSources(MetricsS
> > ystemImpl.java:461)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stop(MetricsSystemIm
> > pl.java:212)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.impl.MetricsSystemImpl.shutdown(MetricsSyst
> > emImpl.java:592)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdownInstance(D
> > efaultMetricsSystem.java:72)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdown(DefaultMe
> > tricsSystem.java:68)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics.shut
> > down(NameNodeMetrics.java:145)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.stop(NameNode.jav
> > a:822)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:172
> > 0)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:169
> > 9)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> > ava:838)
> >    [junit4]   2>    at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
> >    [junit4]   2>    at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
> >    [junit4]   2>    at
> > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
> >    [junit4]   2>    at
> >
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> > est.java:44)
> >    [junit4]   2>    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> > Method)
> >    [junit4]   2>    at
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> > ava:57)
> >    [junit4]   2>    at
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
> > sorImpl.java:43)
> >    [junit4]   2>    at java.lang.reflect.Method.invoke(Method.java:606)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize
> > dRunner.java:1627)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando
> > mizedRunner.java:776)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando
> > mizedRunner.java:792)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.
> > evaluate(SystemPropertiesRestoreRule.java:57)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA
> > fterRule.java:46)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl
> > assName.java:42)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet
> > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss
> > ertionsRequired.java:54)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure
> > .java:48)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule
> > IgnoreAfterMaxFailures.java:65)
> >    [junit4]   2>    at
> >
> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore
> > TestSuites.java:55)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat
> > ementAdapter.java:36)
> >    [junit4]   2>    at
> >
> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.
> > run(ThreadLeakControl.java:365)
> >    [junit4]   2>    at java.lang.Thread.run(Thread.java:745)
> >    [junit4]   2> Caused by: java.lang.NullPointerException
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.blockmanagement.BlocksMap.size(BlocksM
> > ap.java:198)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.getTotalBl
> > ocks(BlockManager.java:3291)
> >    [junit4]   2>    at
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlocksTotal(F
> > SNamesystem.java:6223)
> >    [junit4]   2>    ... 58 more
> >    [junit4]   2> 2645432 INFO  (SUITE-HdfsNNFailoverTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.a.s.SolrTestCaseJ4
> ###deleteCore
> >    [junit4]   2> NOTE: test params are: codec=Asserting(Lucene53),
> > sim=RandomSimilarityProvider(queryNorm=true,coord=yes): {},
> > locale=mk_MK, timezone=Asia/Shanghai
> >    [junit4]   2> NOTE: SunOS 5.11 x86/Oracle Corporation 1.7.0_85 (32-
> > bit)/cpus=3,threads=1,free=99794816,total=518979584
> >    [junit4]   2> NOTE: All tests run in this JVM: [SolrCloudExampleTest,
> > TestStressVersions, TestSerializedLuceneMatchVersion, TestSolrJ,
> > DistanceUnitsTest, MultiThreadedOCPTest, TestDistribDocBasedVersion,
> > BJQParserTest, ZkCLITest, QueryEqualityTest, PrimitiveFieldTypeTest,
> > DistributedQueryComponentOptimizationTest, AliasIntegrationTest,
> > TestInitQParser, TestAuthorizationFramework, TestLazyCores,
> > SolrIndexConfigTest, TestFunctionQuery, TestXIncludeConfig,
> > HardAutoCommitTest, DocValuesMultiTest, TestDefaultStatsCache,
> > SolrRequestParserTest, RecoveryZkTest, UpdateParamsTest,
> > TestSolrDeletionPolicy1, TestDFRSimilarityFactory, TestFastWriter,
> > PathHierarchyTokenizerFactoryTest, TestDynamicLoading,
> > TestElisionMultitermQuery, PolyFieldTest, UnloadDistributedZkTest,
> > TestJsonRequest, TestRuleBasedAuthorizationPlugin,
> > TestManagedStopFilterFactory, TestRawResponseWriter,
> IndexSchemaTest,
> > TestEmbeddedSolrServerConstructors, InfoHandlerTest,
> > AlternateDirectoryTest, LeaderElectionTest, JsonLoaderTest,
> > TestCoreContainer, DirectSolrSpellCheckerTest, RequestLoggingTest,
> > ZkNodePropsTest, TermsComponentTest, TestConfig,
> > TestFieldTypeCollectionResource, XsltUpdateRequestHandlerTest,
> > TestManagedSchemaFieldResource, TestSchemaResource,
> > DataDrivenBlockJoinTest, TestExactStatsCache, TestConfigSetProperties,
> > DeleteLastCustomShardedReplicaTest, TestAnalyzedSuggestions,
> > DirectUpdateHandlerTest, ExternalFileFieldSortTest,
> TestIBSimilarityFactory,
> > TestMissingGroups, ClusterStateUpdateTest, ActionThrottleTest,
> > QueryElevationComponentTest, DocValuesTest, QueryResultKeyTest,
> > TestLRUCache, TestPhraseSuggestions, SimplePostToolTest,
> > TriLevelCompositeIdRoutingTest, DistributedMLTComponentTest,
> > CloudExitableDirectoryReaderTest, TestSolrCloudWithKerberosAlt,
> > TestCodecSupport, TestConfigSets, PeerSyncTest,
> > XmlUpdateRequestHandlerTest, SpatialHeatmapFacetsTest,
> > SoftAutoCommitTest, TestSchemaNameResource,
> > PreAnalyzedUpdateProcessorTest, TestJmxMonitoredMap,
> > TestDistributedStatsComponentCardinality,
> > TestManagedSynonymFilterFactory, JSONWriterTest, TestNRTOpen,
> > ReplicationFactorTest, DOMUtilTest, SolrCoreTest,
> > DocExpirationUpdateProcessorFactoryTest, FastVectorHighlighterTest,
> > SuggesterFSTTest, TestExtendedDismaxParser, TestSolrConfigHandler,
> > DocumentAnalysisRequestHandlerTest,
> > DistributedFacetPivotSmallAdvancedTest, BlockDirectoryTest,
> > TestQuerySenderNoQuery, TestHashPartitioner, DateFieldTest,
> > SegmentsInfoRequestHandlerTest, TestFieldCollectionResource,
> > RecoveryAfterSoftCommitTest, TestMergePolicyConfig,
> TestFieldSortValues,
> > SecurityConfHandlerTest, TestStressReorder, BufferStoreTest,
> > TestRandomRequestDistribution, HdfsBasicDistributedZkTest,
> > TestCloudManagedSchemaConcurrent, TestReplicaProperties,
> > DisMaxRequestHandlerTest, TestMacros, TestStressLucene,
> > TestReloadAndDeleteDocs, BasicAuthIntegrationTest, TestDocSet,
> > BasicDistributedZkTest, DistributedQueryElevationComponentTest,
> > TestGroupingSearch, TestObjectReleaseTracker, MoreLikeThisHandlerTest,
> > OverseerTest, TestFaceting, TestUpdate, TestClassNameShortening,
> > TestRestManager, SyncSliceTest, ShardRoutingTest, ZkSolrClientTest,
> > TestZkChroot, TestRandomDVFaceting, ShardRoutingCustomTest,
> > TestDistributedGrouping, DistributedSpellCheckComponentTest,
> > ZkControllerTest, TestRealTimeGet, TestReload,
> > DistributedTermsComponentTest, TestRangeQuery, SimpleFacetsTest,
> > TestSolr4Spatial, StatsComponentTest, SolrCmdDistributorTest, TestSort,
> > CurrencyFieldXmlFileTest, AnalysisAfterCoreReloadTest,
> > TestFoldingMultitermQuery, SuggesterTSTTest, TestCSVLoader,
> > SchemaVersionSpecificBehaviorTest, SolrCoreCheckLockOnStartupTest,
> > DirectUpdateHandlerOptimizeTest,
> > StatelessScriptUpdateProcessorFactoryTest, DistanceFunctionTest,
> > IndexBasedSpellCheckerTest, StandardRequestHandlerTest,
> > TestOmitPositions, DocumentBuilderTest, RequiredFieldsTest,
> > TestArbitraryIndexDir, LoggingHandlerTest, ReturnFieldsTest,
> > MBeansHandlerTest, UniqFieldsUpdateProcessorFactoryTest,
> > PingRequestHandlerTest, TestComponentsName, TestLFUCache,
> > PreAnalyzedFieldTest, TestSystemIdResolver, SpellingQueryConverterTest,
> > TestUtils, TestDocumentBuilder, SliceStateTest, SystemInfoHandlerTest,
> > UUIDFieldTest, FileUtilsTest, CircularListTest, TestRTGBase,
> > CursorPagingTest, DistributedIntervalFacetingTest,
> > TestDistributedMissingSort, TestSimpleTrackingShardHandler,
> > AsyncMigrateRouteKeyTest, DeleteInactiveReplicaTest,
> > DistribDocExpirationUpdateProcessorTest,
> LeaderFailoverAfterPartitionTest,
> > OverriddenZkACLAndCredentialsProvidersTest,
> > OverseerCollectionConfigSetProcessorTest, OverseerRolesTest,
> > OverseerTaskQueueTest, SSLMigrationTest, SaslZkACLProviderTest,
> > SimpleCollectionCreateDeleteTest, TestAuthenticationFramework,
> > TestCloudInspectUtil, TestCollectionAPI, TestMiniSolrCloudClusterSSL,
> > TestRebalanceLeaders, TestRequestStatusCollectionAPI,
> > HdfsBasicDistributedZk2Test, HdfsChaosMonkeySafeLeaderTest,
> > HdfsCollectionsAPIDistributedZkTest, HdfsNNFailoverTest]
> >    [junit4]   2> NOTE: reproduce with: ant test  -
> > Dtestcase=HdfsNNFailoverTest -Dtests.seed=5D8F351977870E3F -
> > Dtests.slow=true -Dtests.locale=mk_MK -Dtests.timezone=Asia/Shanghai -
> > Dtests.asserts=true -Dtests.file.encoding=UTF-8
> >    [junit4] ERROR   0.00s J0 | HdfsNNFailoverTest (suite) <<<
> >    [junit4]    > Throwable #1: java.lang.RuntimeException: Error while
> running
> > command to get file permissions : java.io.IOException: Cannot run program
> > "/bin/ls": error=12, Not enough space
> >    [junit4]    >    at 
> > java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
> >    [junit4]    >    at 
> > org.apache.hadoop.util.Shell.runCommand(Shell.java:485)
> >    [junit4]    >    at org.apache.hadoop.util.Shell.run(Shell.java:455)
> >    [junit4]    >    at
> >
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715
> > )
> >    [junit4]    >    at
> > org.apache.hadoop.util.Shell.execCommand(Shell.java:808)
> >    [junit4]    >    at
> > org.apache.hadoop.util.Shell.execCommand(Shell.java:791)
> >    [junit4]    >    at
> > org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097)
> >    [junit4]    >    at
> >
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo
> > adPermissionInfo(RawLocalFileSystem.java:582)
> >    [junit4]    >    at
> >
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g
> > etPermission(RawLocalFileSystem.java:557)
> >    [junit4]    >    at
> >
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> > DiskChecker.java:139)
> >    [junit4]    >    at
> > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> > checkDir(DataNode.java:2239)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> > DataNode.java:2281)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> > de.java:2263)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> > ataNode.java:2155)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> > a:1443)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> > ava:828)
> >    [junit4]    >    at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
> >    [junit4]    >    at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
> >    [junit4]    >    at
> > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
> >    [junit4]    >    at
> >
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> > est.java:44)
> >    [junit4]    >    at java.lang.Thread.run(Thread.java:745)
> >    [junit4]    > Caused by: java.io.IOException: error=12, Not enough space
> >    [junit4]    >    at java.lang.UNIXProcess.forkAndExec(Native Method)
> >    [junit4]    >    at java.lang.UNIXProcess.<init>(UNIXProcess.java:137)
> >    [junit4]    >    at java.lang.ProcessImpl.start(ProcessImpl.java:130)
> >    [junit4]    >    at 
> > java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
> >    [junit4]    >    ... 44 more
> >    [junit4]    >    at
> > __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
> >    [junit4]    >    at
> >
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo
> > adPermissionInfo(RawLocalFileSystem.java:620)
> >    [junit4]    >    at
> >
> org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g
> > etPermission(RawLocalFileSystem.java:557)
> >    [junit4]    >    at
> >
> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(
> > DiskChecker.java:139)
> >    [junit4]    >    at
> > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.
> > checkDir(DataNode.java:2239)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(
> > DataNode.java:2281)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo
> > de.java:2263)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D
> > ataNode.java:2155)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav
> > a:1443)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> > ava:828)
> >    [junit4]    >    at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
> >    [junit4]    >    at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
> >    [junit4]    >    at
> > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
> >    [junit4]    >    at
> >
> org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT
> > est.java:44)
> >    [junit4]    >    at java.lang.Thread.run(Thread.java:745)
> >    [junit4] Completed [426/536] on J0 in 45.66s, 0 tests, 1 error <<<
> FAILURES!
> >
> > [...truncated 300 lines...]
> >    [junit4] Suite: org.apache.solr.store.hdfs.HdfsDirectoryTest
> >    [junit4]   2> Creating dataDir: /export/home/jenkins/workspace/Lucene-
> > Solr-5.x-Solaris/solr/build/solr-
> > core/test/J1/temp/solr.store.hdfs.HdfsDirectoryTest_5D8F351977870E3F-
> > 001/init-core-data-001
> >    [junit4]   2> 3147821 INFO  (SUITE-HdfsDirectoryTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized
> ssl
> > (false) and clientAuth (false)
> >    [junit4]   1> Formatting using clusterid: testClusterID
> >    [junit4]   2> 3147964 WARN  (SUITE-HdfsDirectoryTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.a.h.m.i.MetricsConfig Cannot
> > locate configuration: tried hadoop-metrics2-
> namenode.properties,hadoop-
> > metrics2.properties
> >    [junit4]   2> 3147974 WARN  (SUITE-HdfsDirectoryTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.a.h.h.HttpRequestLog Jetty
> > request log can only be enabled using Log4j
> >    [junit4]   2> 3147976 INFO  (SUITE-HdfsDirectoryTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log jetty-6.1.26
> >    [junit4]   2> 3147994 INFO  (SUITE-HdfsDirectoryTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log Extract
> > jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-
> > hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/hdfs to
> > ./temp/Jetty_solaris.vm_46547_hdfs____.vwfmpk/webapp
> >    [junit4]   2> 3148170 INFO  (SUITE-HdfsDirectoryTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log NO JSP Support for /, did
> not
> > find org.apache.jasper.servlet.JspServlet
> >    [junit4]   2> 3148982 INFO  (SUITE-HdfsDirectoryTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log Started
> > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:46547
> >    [junit4]   2> 3157264 INFO  (SUITE-HdfsDirectoryTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.m.log Stopped
> > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0
> >    [junit4]   2> 3157403 INFO  (SUITE-HdfsDirectoryTest-
> > seed#[5D8F351977870E3F]-worker) [    ] o.a.s.SolrTestCaseJ4
> ###deleteCore
> >    [junit4]   2> Aug 29, 2015 9:01:07 PM
> > com.carrotsearch.randomizedtesting.ThreadLeakControl
> checkThreadLeaks
> >    [junit4]   2> WARNING: Will linger awaiting termination of 1 leaked
> > thread(s).
> >    [junit4]   2> Aug 29, 2015 9:01:27 PM
> > com.carrotsearch.randomizedtesting.ThreadLeakControl
> checkThreadLeaks
> >    [junit4]   2> SEVERE: 1 thread leaked from SUITE scope at
> > org.apache.solr.store.hdfs.HdfsDirectoryTest:
> >    [junit4]   2>    1) Thread[id=20389, name=IPC Server idle connection
> scanner
> > for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
> >    [junit4]   2>         at java.lang.Object.wait(Native Method)
> >    [junit4]   2>         at java.lang.Object.wait(Object.java:503)
> >    [junit4]   2>         at java.util.TimerThread.mainLoop(Timer.java:526)
> >    [junit4]   2>         at java.util.TimerThread.run(Timer.java:505)
> >    [junit4]   2> Aug 29, 2015 9:01:27 PM
> > com.carrotsearch.randomizedtesting.ThreadLeakControl tryToInterruptAll
> >    [junit4]   2> INFO: Starting to interrupt leaked threads:
> >    [junit4]   2>    1) Thread[id=20389, name=IPC Server idle connection
> scanner
> > for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
> >    [junit4]   2> Aug 29, 2015 9:01:30 PM
> > com.carrotsearch.randomizedtesting.ThreadLeakControl tryToInterruptAll
> >    [junit4]   2> SEVERE: There are still zombie threads that couldn't be
> > terminated:
> >    [junit4]   2>    1) Thread[id=20389, name=IPC Server idle connection
> scanner
> > for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
> >    [junit4]   2>         at java.lang.Object.wait(Native Method)
> >    [junit4]   2>         at java.lang.Object.wait(Object.java:503)
> >    [junit4]   2>         at java.util.TimerThread.mainLoop(Timer.java:526)
> >    [junit4]   2>         at java.util.TimerThread.run(Timer.java:505)
> >    [junit4]   2> NOTE: test params are: codec=Asserting(Lucene53): {},
> > docValues:{}, sim=DefaultSimilarity, locale=es_BO,
> > timezone=Antarctica/South_Pole
> >    [junit4]   2> NOTE: SunOS 5.11 x86/Oracle Corporation 1.7.0_85 (32-
> > bit)/cpus=3,threads=2,free=136627544,total=518979584
> >    [junit4]   2> NOTE: All tests run in this JVM: [TestIndexingPerformance,
> > TestCSVResponseWriter, DistributedQueryComponentCustomSortTest,
> > DirectSolrConnectionTest, FullSolrCloudDistribCmdsTest,
> > TestShardHandlerFactory, CacheHeaderTest, BasicZkTest, TestTrie,
> > FieldAnalysisRequestHandlerTest, PKIAuthenticationIntegrationTest,
> > OpenCloseCoreStressTest, TestSuggestSpellingConverter, StressHdfsTest,
> > CleanupOldIndexTest, DistributedExpandComponentTest,
> > TestHdfsUpdateLog, TestSolrXml, TestAddFieldRealTimeGet,
> TestJsonFacets,
> > DistributedSuggestComponentTest,
> > OutOfBoxZkACLAndCredentialsProvidersTest, AnalyticsMergeStrategyTest,
> > HLLUtilTest, ResponseHeaderTest, SearchHandlerTest,
> > BinaryUpdateRequestHandlerTest, DistributedFacetPivotWhiteBoxTest,
> > ConnectionManagerTest, SpellCheckComponentTest,
> > TestScoreJoinQPNoScore, SolrTestCaseJ4Test, SolrIndexSplitterTest,
> > TestConfigSetsAPI, TestDefaultSearchFieldResource, TestCryptoKeys,
> > TestNonDefinedSimilarityFactory, TestCoreDiscovery, RollingRestartTest,
> > SolrInfoMBeanTest, CustomCollectionTest, DistributedVersionInfoTest,
> > ClusterStateTest, TestReversedWildcardFilterFactory, SolrXmlInZkTest,
> > DistributedFacetPivotLongTailTest, URLClassifyProcessorTest,
> > TestLMJelinekMercerSimilarityFactory, RequestHandlersTest,
> > RemoteQueryErrorTest, LeaderElectionIntegrationTest,
> > SharedFSAutoReplicaFailoverTest, TestBadConfig,
> > SignatureUpdateProcessorFactoryTest,
> TestCursorMarkWithoutUniqueKey,
> > TestCrossCoreJoin, SparseHLLTest, DistributedQueueTest,
> > BigEndianAscendingWordSerializerTest, TestBM25SimilarityFactory,
> > AutoCommitTest, DateMathParserTest, BasicFunctionalityTest,
> > SuggesterWFSTTest, TestCollapseQParserPlugin, TestManagedResource,
> > TestSha256AuthenticationProvider, CollectionTooManyReplicasTest,
> > BadCopyFieldTest, TestDownShardTolerantSearch, CloudMLTQParserTest,
> > NotRequiredUniqueKeyTest, TestAnalyzeInfixSuggestions,
> > ExitableDirectoryReaderTest, TestScoreJoinQPScore, DeleteShardTest,
> > RankQueryTest, TestSchemaManager,
> UpdateRequestProcessorFactoryTest,
> > CursorMarkTest, DistributedDebugComponentTest, DeleteReplicaTest,
> > RAMDirectoryFactoryTest, ConcurrentDeleteAndCreateCollectionTest,
> > TestQueryTypes, OutputWriterTest, TestSchemaSimilarityResource,
> > HighlighterMaxOffsetTest, ResponseLogComponentTest,
> > TestCloudPivotFacet, DocValuesMissingTest,
> > FieldMutatingUpdateProcessorTest, HttpPartitionTest, TestCollationField,
> > ZkStateWriterTest, TestQuerySenderListener, AtomicUpdatesTest,
> > TestStressRecovery, TestRandomFaceting,
> > SharedFSAutoReplicaFailoverUtilsTest, CoreAdminHandlerTest,
> > HighlighterConfigTest, TestCustomSort, MultiTermTest,
> > VMParamsZkACLAndCredentialsProvidersTest,
> > IgnoreCommitOptimizeUpdateProcessorFactoryTest, CollectionReloadTest,
> > PrimUtilsTest, TestRecovery, TestWriterPerf,
> > AddSchemaFieldsUpdateProcessorFactoryTest, TimeZoneUtilsTest,
> > CurrencyFieldOpenExchangeTest, TestSolrCLIRunExample,
> > TestPHPSerializedResponseWriter, ChaosMonkeySafeLeaderTest,
> > TestIndexSearcher, EnumFieldTest, TestSolrIndexConfig,
> > TermVectorComponentDistributedTest, TestJoin, TestExpandComponent,
> > TestManagedResourceStorage, SortByFunctionTest,
> > TestDefaultSimilarityFactory, SuggesterTest, TestValueSourceCache,
> > SolrPluginUtilsTest, TermVectorComponentTest, TestFiltering,
> > TestQueryUtils, FileBasedSpellCheckerTest, BasicDistributedZk2Test,
> > CollectionsAPIDistributedZkTest, TestReplicationHandler,
> > TestDistributedSearch, BadIndexSchemaTest, ConvertedLegacyTest,
> > HighlighterTest, ShowFileRequestHandlerTest, SpellCheckCollatorTest,
> > SpatialFilterTest, NoCacheHeaderTest, WordBreakSolrSpellCheckerTest,
> > TestPseudoReturnFields, TestAtomicUpdateErrorCases,
> > TestWordDelimiterFilterFactory, DefaultValueUpdateProcessorTest,
> > TestRemoteStreaming, DebugComponentTest, TestSurroundQueryParser,
> > LukeRequestHandlerTest, TestSolrQueryParser,
> > IndexSchemaRuntimeFieldTest, RegexBoostProcessorTest,
> > TestJmxIntegration, QueryParsingTest, TestPartialUpdateDeduplication,
> > CSVRequestHandlerTest, TestBinaryResponseWriter, SOLR749Test,
> > CopyFieldTest, BadComponentTest, TestSolrDeletionPolicy2, SampleTest,
> > TestBinaryField, TestSearchPerf, NumericFieldsTest, MinimalSchemaTest,
> > TestFuzzyAnalyzedSuggestions, TestSolrCoreProperties,
> > TestPostingsSolrHighlighter, TestLuceneMatchVersion,
> > SpellPossibilityIteratorTest, TestCharFilters, SynonymTokenizerTest,
> > EchoParamsTest, TestSweetSpotSimilarityFactory, TestPerFieldSimilarity,
> > TestLMDirichletSimilarityFactory, ResourceLoaderTest,
> > TestFastOutputStream, ScriptEngineTest,
> > OpenExchangeRatesOrgProviderTest, PluginInfoTest, TestFastLRUCache,
> > ChaosMonkeyNothingIsSafeTest, TestHighlightDedupGrouping,
> > TestTolerantSearch, TestJettySolrRunner, AssignTest,
> > AsyncCallRequestStatusResponseTest, CollectionStateFormat2Test,
> > CollectionsAPIAsyncDistributedZkTest, DistribCursorPagingTest,
> > DistribJoinFromCollectionTest, LeaderInitiatedRecoveryOnCommitTest,
> > MigrateRouteKeyTest, OverseerStatusTest, ShardSplitTest,
> > TestConfigSetsAPIExclusivity, TestConfigSetsAPIZkFailure,
> > TestLeaderElectionZkExpiry, TestMiniSolrCloudCluster,
> > TestShortCircuitedRequests, HdfsRecoverLeaseTest,
> > CachingDirectoryFactoryTest, HdfsDirectoryFactoryTest, TestConfigOverlay,
> > TestConfigSetImmutable, TestImplicitCoreProperties,
> > TestInfoStreamLogging, TestInitParams, TestSolrDynamicMBean,
> > TestBlobHandler, TestConfigReload, TestReplicationHandlerBackup,
> > TestSolrConfigHandlerConcurrent, CoreAdminCreateDiscoverTest,
> > CoreAdminRequestStatusTest, CoreMergeIndexesAdminHandlerTest,
> > DistributedFacetPivotLargeTest, DistributedFacetPivotSmallTest,
> > FacetPivotSmallTest, SuggestComponentTest, JavabinLoaderTest,
> > SmileWriterTest, TestIntervalFaceting, TestChildDocTransformer,
> > TestCustomDocTransformer, TestSortingResponseWriter,
> > TestBulkSchemaAPI, TestFieldResource,
> > TestManagedSchemaDynamicFieldResource, TestBulkSchemaConcurrent,
> > TestCloudSchemaless, TestReloadDeadlock, TestSearcherReuse,
> > TestSimpleQParserPlugin, TestSmileRequest, TestSolr4Spatial2,
> > TestStandardQParsers, TestStressUserVersions, TestTrieFacet,
> > TestMinMaxOnMultiValuedField, TestOrdValues,
> > TestSortByMinMaxFunction, SimpleMLTQParserTest, TestDistribIDF,
> > TestExactSharedStatsCache, TestPKIAuthenticationPlugin,
> > TestBlendedInfixSuggestions, TestFileDictionaryLookup,
> > TestFreeTextSuggestions, TestHighFrequencyDictionaryFactory,
> > BlockCacheTest, HdfsDirectoryTest]
> >    [junit4]   2> NOTE: reproduce with: ant test  
> > -Dtestcase=HdfsDirectoryTest
> -
> > Dtests.seed=5D8F351977870E3F -Dtests.slow=true -Dtests.locale=es_BO -
> > Dtests.timezone=Antarctica/South_Pole -Dtests.asserts=true -
> > Dtests.file.encoding=UTF-8
> >    [junit4] ERROR   0.00s J1 | HdfsDirectoryTest (suite) <<<
> >    [junit4]    > Throwable #1: java.security.AccessControlException: access
> > denied ("java.io.FilePermission"
> "/export/home/jenkins/workspace/Lucene-
> > Solr-5.x-Solaris/solr/build/solr-core/test/J1" "write")
> >    [junit4]    >    at
> > __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
> >    [junit4]    >    at
> >
> java.security.AccessControlContext.checkPermission(AccessControlContext.j
> > ava:395)
> >    [junit4]    >    at
> > java.security.AccessController.checkPermission(AccessController.java:559)
> >    [junit4]    >    at
> > java.lang.SecurityManager.checkPermission(SecurityManager.java:549)
> >    [junit4]    >    at
> > java.lang.SecurityManager.checkWrite(SecurityManager.java:979)
> >    [junit4]    >    at java.io.File.canWrite(File.java:785)
> >    [junit4]    >    at 
> > org.apache.hadoop.fs.FileUtil.canWrite(FileUtil.java:1002)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.createPermissionsDiagnosisString(
> > MiniDFSCluster.java:856)
> >    [junit4]    >    at
> >
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j
> > ava:812)
> >    [junit4]    >    at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738)
> >    [junit4]    >    at
> > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608)
> >    [junit4]    >    at
> > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98)
> >    [junit4]    >    at
> > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:60)
> >    [junit4]    >    at
> >
> org.apache.solr.store.hdfs.HdfsDirectoryTest.beforeClass(HdfsDirectoryTest.
> > java:62)
> >    [junit4]    >    at java.lang.Thread.run(Thread.java:745)Throwable #2:
> > com.carrotsearch.randomizedtesting.ThreadLeakError: 1 thread leaked
> from
> > SUITE scope at org.apache.solr.store.hdfs.HdfsDirectoryTest:
> >    [junit4]    >    1) Thread[id=20389, name=IPC Server idle connection
> scanner
> > for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
> >    [junit4]    >         at java.lang.Object.wait(Native Method)
> >    [junit4]    >         at java.lang.Object.wait(Object.java:503)
> >    [junit4]    >         at java.util.TimerThread.mainLoop(Timer.java:526)
> >    [junit4]    >         at java.util.TimerThread.run(Timer.java:505)
> >    [junit4]    >    at
> > __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)Throwable #3:
> > com.carrotsearch.randomizedtesting.ThreadLeakError: There are still
> zombie
> > threads that couldn't be terminated:
> >    [junit4]    >    1) Thread[id=20389, name=IPC Server idle connection
> scanner
> > for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest]
> >    [junit4]    >         at java.lang.Object.wait(Native Method)
> >    [junit4]    >         at java.lang.Object.wait(Object.java:503)
> >    [junit4]    >         at java.util.TimerThread.mainLoop(Timer.java:526)
> >    [junit4]    >         at java.util.TimerThread.run(Timer.java:505)
> >    [junit4]    >    at
> > __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)
> >    [junit4] Completed [521/536] on J1 in 33.05s, 0 tests, 3 errors <<<
> FAILURES!
> >
> > [...truncated 64 lines...]
> > BUILD FAILED
> > /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:785:
> > The following error occurred while executing this line:
> > /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:729:
> > The following error occurred while executing this line:
> > /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:59:
> The
> > following error occurred while executing this line:
> > /export/home/jenkins/workspace/Lucene-Solr-5.x-
> > Solaris/solr/build.xml:233: The following error occurred while executing 
> > this
> > line:
> > /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/common-
> > build.xml:524: The following error occurred while executing this line:
> > /export/home/jenkins/workspace/Lucene-Solr-5.x-
> Solaris/lucene/common-
> > build.xml:1452: The following error occurred while executing this line:
> > /export/home/jenkins/workspace/Lucene-Solr-5.x-
> Solaris/lucene/common-
> > build.xml:1006: There were test failures: 536 suites, 2123 tests, 4 
> > suite-level
> > errors, 108 ignored (34 assumptions)
> >
> > Total time: 77 minutes 51 seconds
> > Build step 'Invoke Ant' marked build as failure
> > Archiving artifacts
> > [WARNINGS] Skipping publisher since build result is FAILURE
> > Recording test results
> > Email was triggered for: Failure - Any
> > Sending email for trigger: Failure - Any
> >
> 
> 
> 
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: [email protected]
> For additional commands, e-mail: [email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to