I am still digging... On Solaris there seems to be a general forking problem on 32 bit processes.
Uwe ----- Uwe Schindler H.-H.-Meier-Allee 63, D-28213 Bremen http://www.thetaphi.de eMail: [email protected] > -----Original Message----- > From: Policeman Jenkins Server [mailto:[email protected]] > Sent: Saturday, August 29, 2015 11:02 PM > To: [email protected]; [email protected]; [email protected] > Subject: [JENKINS] Lucene-Solr-5.x-Solaris (multiarch/jdk1.7.0) - Build # 9 - > Still Failing! > > Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Solaris/9/ > Java: multiarch/jdk1.7.0 -d32 -server -XX:+UseConcMarkSweepGC > > 4 tests failed. > FAILED: > junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsNNFailoverTest > > Error Message: > Error while running command to get file permissions : java.io.IOException: > Cannot run program "/bin/ls": error=12, Not enough space at > java.lang.ProcessBuilder.start(ProcessBuilder.java:1047) at > org.apache.hadoop.util.Shell.runCommand(Shell.java:485) at > org.apache.hadoop.util.Shell.run(Shell.java:455) at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715 > ) at org.apache.hadoop.util.Shell.execCommand(Shell.java:808) at > org.apache.hadoop.util.Shell.execCommand(Shell.java:791) at > org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097) at > org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo > adPermissionInfo(RawLocalFileSystem.java:582) at > org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g > etPermission(RawLocalFileSystem.java:557) at > org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck( > DiskChecker.java:139) at > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156) at > org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker. > checkDir(DataNode.java:2239) at > org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations( > DataNode.java:2281) at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo > de.java:2263) at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D > ataNode.java:2155) at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav > a:1443) at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j > ava:828) at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738) at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608) at > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98) at > org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT > est.java:44) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j > ava:57) at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces > sorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at > com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize > dRunner.java:1627) at > com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando > mizedRunner.java:776) at > com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando > mizedRunner.java:792) at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) at > com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1. > evaluate(SystemPropertiesRestoreRule.java:57) at > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA > fterRule.java:46) at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) at > org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl > assName.java:42) at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) at > org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss > ertionsRequired.java:54) at > org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure > .java:48) at > org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule > IgnoreAfterMaxFailures.java:65) at > org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore > TestSuites.java:55) at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) at > com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner. > run(ThreadLeakControl.java:365) at java.lang.Thread.run(Thread.java:745) > Caused by: java.io.IOException: error=12, Not enough space at > java.lang.UNIXProcess.forkAndExec(Native Method) at > java.lang.UNIXProcess.<init>(UNIXProcess.java:137) at > java.lang.ProcessImpl.start(ProcessImpl.java:130) at > java.lang.ProcessBuilder.start(ProcessBuilder.java:1028) ... 44 more > > Stack Trace: > java.lang.RuntimeException: Error while running command to get file > permissions : java.io.IOException: Cannot run program "/bin/ls": error=12, > Not enough space > at java.lang.ProcessBuilder.start(ProcessBuilder.java:1047) > at org.apache.hadoop.util.Shell.runCommand(Shell.java:485) > at org.apache.hadoop.util.Shell.run(Shell.java:455) > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715 > ) > at org.apache.hadoop.util.Shell.execCommand(Shell.java:808) > at org.apache.hadoop.util.Shell.execCommand(Shell.java:791) > at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097) > at > org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo > adPermissionInfo(RawLocalFileSystem.java:582) > at > org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g > etPermission(RawLocalFileSystem.java:557) > at > org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck( > DiskChecker.java:139) > at > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156) > at > org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker. > checkDir(DataNode.java:2239) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations( > DataNode.java:2281) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo > de.java:2263) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D > ataNode.java:2155) > at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav > a:1443) > at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j > ava:828) > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738) > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608) > at > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98) > at > org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT > est.java:44) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j > ava:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces > sorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at > com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize > dRunner.java:1627) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando > mizedRunner.java:776) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando > mizedRunner.java:792) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1. > evaluate(SystemPropertiesRestoreRule.java:57) > at > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA > fterRule.java:46) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl > assName.java:42) > at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss > ertionsRequired.java:54) > at > org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure > .java:48) > at > org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule > IgnoreAfterMaxFailures.java:65) > at > org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore > TestSuites.java:55) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner. > run(ThreadLeakControl.java:365) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.io.IOException: error=12, Not enough space > at java.lang.UNIXProcess.forkAndExec(Native Method) > at java.lang.UNIXProcess.<init>(UNIXProcess.java:137) > at java.lang.ProcessImpl.start(ProcessImpl.java:130) > at java.lang.ProcessBuilder.start(ProcessBuilder.java:1028) > ... 44 more > > at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0) > at > org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo > adPermissionInfo(RawLocalFileSystem.java:620) > at > org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g > etPermission(RawLocalFileSystem.java:557) > at > org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck( > DiskChecker.java:139) > at > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156) > at > org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker. > checkDir(DataNode.java:2239) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations( > DataNode.java:2281) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo > de.java:2263) > at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D > ataNode.java:2155) > at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav > a:1443) > at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j > ava:828) > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738) > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608) > at > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98) > at > org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT > est.java:44) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j > ava:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces > sorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at > com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize > dRunner.java:1627) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando > mizedRunner.java:776) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando > mizedRunner.java:792) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1. > evaluate(SystemPropertiesRestoreRule.java:57) > at > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA > fterRule.java:46) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl > assName.java:42) > at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss > ertionsRequired.java:54) > at > org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure > .java:48) > at > org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule > IgnoreAfterMaxFailures.java:65) > at > org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore > TestSuites.java:55) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner. > run(ThreadLeakControl.java:365) > at java.lang.Thread.run(Thread.java:745) > > > FAILED: > junit.framework.TestSuite.org.apache.solr.store.hdfs.HdfsDirectoryTest > > Error Message: > access denied ("java.io.FilePermission" > "/export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/build/solr- > core/test/J1" "write") > > Stack Trace: > java.security.AccessControlException: access denied > ("java.io.FilePermission" "/export/home/jenkins/workspace/Lucene-Solr- > 5.x-Solaris/solr/build/solr-core/test/J1" "write") > at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0) > at > java.security.AccessControlContext.checkPermission(AccessControlContext.j > ava:395) > at > java.security.AccessController.checkPermission(AccessController.java:559) > at > java.lang.SecurityManager.checkPermission(SecurityManager.java:549) > at java.lang.SecurityManager.checkWrite(SecurityManager.java:979) > at java.io.File.canWrite(File.java:785) > at org.apache.hadoop.fs.FileUtil.canWrite(FileUtil.java:1002) > at > org.apache.hadoop.hdfs.MiniDFSCluster.createPermissionsDiagnosisString( > MiniDFSCluster.java:856) > at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j > ava:812) > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738) > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608) > at > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98) > at > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:60) > at > org.apache.solr.store.hdfs.HdfsDirectoryTest.beforeClass(HdfsDirectoryTest. > java:62) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j > ava:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces > sorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at > com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize > dRunner.java:1627) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando > mizedRunner.java:776) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando > mizedRunner.java:792) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1. > evaluate(SystemPropertiesRestoreRule.java:57) > at > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA > fterRule.java:46) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl > assName.java:42) > at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss > ertionsRequired.java:54) > at > org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure > .java:48) > at > org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule > IgnoreAfterMaxFailures.java:65) > at > org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore > TestSuites.java:55) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner. > run(ThreadLeakControl.java:365) > at java.lang.Thread.run(Thread.java:745) > > > FAILED: > junit.framework.TestSuite.org.apache.solr.store.hdfs.HdfsDirectoryTest > > Error Message: > 1 thread leaked from SUITE scope at > org.apache.solr.store.hdfs.HdfsDirectoryTest: 1) Thread[id=20389, > name=IPC Server idle connection scanner for port 41610, state=WAITING, > group=TGRP-HdfsDirectoryTest] at java.lang.Object.wait(Native > Method) at java.lang.Object.wait(Object.java:503) at > java.util.TimerThread.mainLoop(Timer.java:526) at > java.util.TimerThread.run(Timer.java:505) > > Stack Trace: > com.carrotsearch.randomizedtesting.ThreadLeakError: 1 thread leaked from > SUITE scope at org.apache.solr.store.hdfs.HdfsDirectoryTest: > 1) Thread[id=20389, name=IPC Server idle connection scanner for port > 41610, state=WAITING, group=TGRP-HdfsDirectoryTest] > at java.lang.Object.wait(Native Method) > at java.lang.Object.wait(Object.java:503) > at java.util.TimerThread.mainLoop(Timer.java:526) > at java.util.TimerThread.run(Timer.java:505) > at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0) > > > FAILED: > junit.framework.TestSuite.org.apache.solr.store.hdfs.HdfsDirectoryTest > > Error Message: > There are still zombie threads that couldn't be terminated: 1) > Thread[id=20389, name=IPC Server idle connection scanner for port 41610, > state=WAITING, group=TGRP-HdfsDirectoryTest] at > java.lang.Object.wait(Native Method) at > java.lang.Object.wait(Object.java:503) at > java.util.TimerThread.mainLoop(Timer.java:526) at > java.util.TimerThread.run(Timer.java:505) > > Stack Trace: > com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie > threads that couldn't be terminated: > 1) Thread[id=20389, name=IPC Server idle connection scanner for port > 41610, state=WAITING, group=TGRP-HdfsDirectoryTest] > at java.lang.Object.wait(Native Method) > at java.lang.Object.wait(Object.java:503) > at java.util.TimerThread.mainLoop(Timer.java:526) > at java.util.TimerThread.run(Timer.java:505) > at __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0) > > > > > Build Log: > [...truncated 10577 lines...] > [junit4] Suite: org.apache.solr.cloud.hdfs.HdfsNNFailoverTest > [junit4] 2> Creating dataDir: /export/home/jenkins/workspace/Lucene- > Solr-5.x-Solaris/solr/build/solr- > core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F- > 001/init-core-data-001 > [junit4] 2> 2599844 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.s.BaseDistributedSearchTestCase > Setting hostContext system property: / > [junit4] 2> 2616331 WARN (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.h.u.NativeCodeLoader Unable to > load native-hadoop library for your platform... using builtin-java classes > where applicable > [junit4] 1> Formatting using clusterid: testClusterID > [junit4] 2> 2617524 WARN (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.h.m.i.MetricsConfig Cannot > locate configuration: tried hadoop-metrics2-namenode.properties,hadoop- > metrics2.properties > [junit4] 2> 2617755 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log Logging to > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via > org.mortbay.log.Slf4jLog > [junit4] 2> 2617771 WARN (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.h.h.HttpRequestLog Jetty > request log can only be enabled using Log4j > [junit4] 2> 2617878 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log jetty-6.1.26 > [junit4] 2> 2617942 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log Extract > jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop- > hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/hdfs to > ./temp/Jetty_solaris.vm_35231_hdfs____thayv4/webapp > [junit4] 2> 2618129 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log NO JSP Support for /, did not > find org.apache.jasper.servlet.JspServlet > [junit4] 2> 2619464 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log Started > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:35231 > [junit4] 2> 2637264 WARN (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.h.h.s.d.DataNode Invalid > dfs.datanode.data.dir /export/home/jenkins/workspace/Lucene-Solr-5.x- > Solaris/solr/build/solr- > core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F- > 001/tempDir-001/hdfsBaseDir/data/data2 : > [junit4] 2> java.io.IOException: Cannot run program "chmod": error=12, > Not enough space > [junit4] 2> at > java.lang.ProcessBuilder.start(ProcessBuilder.java:1047) > [junit4] 2> at > org.apache.hadoop.util.Shell.runCommand(Shell.java:485) > [junit4] 2> at org.apache.hadoop.util.Shell.run(Shell.java:455) > [junit4] 2> at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715 > ) > [junit4] 2> at > org.apache.hadoop.util.Shell.execCommand(Shell.java:808) > [junit4] 2> at > org.apache.hadoop.util.Shell.execCommand(Shell.java:791) > [junit4] 2> at > org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSyste > m.java:656) > [junit4] 2> at > org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:4 > 90) > [junit4] 2> at > org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck( > DiskChecker.java:140) > [junit4] 2> at > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156) > [junit4] 2> at > org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker. > checkDir(DataNode.java:2239) > [junit4] 2> at > org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations( > DataNode.java:2281) > [junit4] 2> at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo > de.java:2263) > [junit4] 2> at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D > ataNode.java:2155) > [junit4] 2> at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav > a:1443) > [junit4] 2> at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j > ava:828) > [junit4] 2> at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738) > [junit4] 2> at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608) > [junit4] 2> at > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98) > [junit4] 2> at > org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT > est.java:44) > [junit4] 2> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) > [junit4] 2> at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j > ava:57) > [junit4] 2> at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces > sorImpl.java:43) > [junit4] 2> at java.lang.reflect.Method.invoke(Method.java:606) > [junit4] 2> at > com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize > dRunner.java:1627) > [junit4] 2> at > com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando > mizedRunner.java:776) > [junit4] 2> at > com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando > mizedRunner.java:792) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1. > evaluate(SystemPropertiesRestoreRule.java:57) > [junit4] 2> at > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA > fterRule.java:46) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl > assName.java:42) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss > ertionsRequired.java:54) > [junit4] 2> at > org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure > .java:48) > [junit4] 2> at > org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule > IgnoreAfterMaxFailures.java:65) > [junit4] 2> at > org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore > TestSuites.java:55) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner. > run(ThreadLeakControl.java:365) > [junit4] 2> at java.lang.Thread.run(Thread.java:745) > [junit4] 2> Caused by: java.io.IOException: error=12, Not enough space > [junit4] 2> at java.lang.UNIXProcess.forkAndExec(Native Method) > [junit4] 2> at java.lang.UNIXProcess.<init>(UNIXProcess.java:137) > [junit4] 2> at java.lang.ProcessImpl.start(ProcessImpl.java:130) > [junit4] 2> at > java.lang.ProcessBuilder.start(ProcessBuilder.java:1028) > [junit4] 2> ... 43 more > [junit4] 2> 2637287 WARN > (org.apache.hadoop.util.JvmPauseMonitor$Monitor@be51b7) [ ] > o.a.h.u.JvmPauseMonitor Detected pause in JVM or host machine (eg GC): > pause of approximately 15969ms > [junit4] 2> No GCs detected > [junit4] 2> 2637368 WARN (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.h.h.HttpRequestLog Jetty > request log can only be enabled using Log4j > [junit4] 2> 2637384 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log jetty-6.1.26 > [junit4] 2> 2637422 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log Extract > jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop- > hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/datanode to > ./temp/Jetty_solaris.vm_49465_datanode____96t731/webapp > [junit4] 2> 2637655 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log NO JSP Support for /, did not > find org.apache.jasper.servlet.JspServlet > [junit4] 2> 2638756 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log Started > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:49465 > [junit4] 2> 2645079 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log Stopped > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0 > [junit4] 2> 2645234 ERROR (DataNode: > [[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x- > Solaris/solr/build/solr- > core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F- > 001/tempDir-001/hdfsBaseDir/data/data1/, > [DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x- > Solaris/solr/build/solr- > core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F- > 001/tempDir-001/hdfsBaseDir/data/data2/]] heartbeating to solaris- > vm/127.0.0.1:61051) [ ] o.a.h.h.s.d.DataNode Initialization failed for > Block > pool <registering> (Datanode Uuid unassigned) service to solaris- > vm/127.0.0.1:61051. Exiting. > [junit4] 2> java.io.IOException: DN shut down before block pool > connected > [junit4] 2> at > org.apache.hadoop.hdfs.server.datanode.BPServiceActor.retrieveNamespac > eInfo(BPServiceActor.java:185) > [junit4] 2> at > org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAnd > Handshake(BPServiceActor.java:215) > [junit4] 2> at > org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceAct > or.java:828) > [junit4] 2> at java.lang.Thread.run(Thread.java:745) > [junit4] 2> 2645236 WARN (DataNode: > [[[DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x- > Solaris/solr/build/solr- > core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F- > 001/tempDir-001/hdfsBaseDir/data/data1/, > [DISK]file:/export/home/jenkins/workspace/Lucene-Solr-5.x- > Solaris/solr/build/solr- > core/test/J0/temp/solr.cloud.hdfs.HdfsNNFailoverTest_5D8F351977870E3F- > 001/tempDir-001/hdfsBaseDir/data/data2/]] heartbeating to solaris- > vm/127.0.0.1:61051) [ ] o.a.h.h.s.d.DataNode Ending block pool service for: > Block pool <registering> (Datanode Uuid unassigned) service to solaris- > vm/127.0.0.1:61051 > [junit4] 2> 2645259 WARN > (org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager > $Monitor@7b7964) [ ] o.a.h.h.s.b.DecommissionManager Monitor > interrupted: java.lang.InterruptedException: sleep interrupted > [junit4] 2> 2645314 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log Stopped > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0 > [junit4] 2> 2645418 ERROR (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.h.m.l.MethodMetric Error > invoking method getBlocksTotal > [junit4] 2> java.lang.reflect.InvocationTargetException > [junit4] 2> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) > [junit4] 2> at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j > ava:57) > [junit4] 2> at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces > sorImpl.java:43) > [junit4] 2> at java.lang.reflect.Method.invoke(Method.java:606) > [junit4] 2> at > org.apache.hadoop.metrics2.lib.MethodMetric$2.snapshot(MethodMetric.j > ava:111) > [junit4] 2> at > org.apache.hadoop.metrics2.lib.MethodMetric.snapshot(MethodMetric.jav > a:144) > [junit4] 2> at > org.apache.hadoop.metrics2.lib.MetricsRegistry.snapshot(MetricsRegistry.ja > va:387) > [junit4] 2> at > org.apache.hadoop.metrics2.lib.MetricsSourceBuilder$1.getMetrics(MetricsS > ourceBuilder.java:79) > [junit4] 2> at > org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMetrics(Metrics > SourceAdapter.java:195) > [junit4] 2> at > org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.updateJmxCache( > MetricsSourceAdapter.java:172) > [junit4] 2> at > org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.getMBeanInfo(Me > tricsSourceAdapter.java:151) > [junit4] 2> at > com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getClassName(De > faultMBeanServerInterceptor.java:1804) > [junit4] 2> at > com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.safeGetClassNam > e(DefaultMBeanServerInterceptor.java:1595) > [junit4] 2> at > com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.checkMBeanPer > mission(DefaultMBeanServerInterceptor.java:1813) > [junit4] 2> at > com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.exclusiveUnregist > erMBean(DefaultMBeanServerInterceptor.java:430) > [junit4] 2> at > com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.unregisterMBean > (DefaultMBeanServerInterceptor.java:415) > [junit4] 2> at > com.sun.jmx.mbeanserver.JmxMBeanServer.unregisterMBean(JmxMBeanS > erver.java:546) > [junit4] 2> at > org.apache.hadoop.metrics2.util.MBeans.unregister(MBeans.java:81) > [junit4] 2> at > org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stopMBeans(Metri > csSourceAdapter.java:227) > [junit4] 2> at > org.apache.hadoop.metrics2.impl.MetricsSourceAdapter.stop(MetricsSourc > eAdapter.java:212) > [junit4] 2> at > org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stopSources(MetricsS > ystemImpl.java:461) > [junit4] 2> at > org.apache.hadoop.metrics2.impl.MetricsSystemImpl.stop(MetricsSystemIm > pl.java:212) > [junit4] 2> at > org.apache.hadoop.metrics2.impl.MetricsSystemImpl.shutdown(MetricsSyst > emImpl.java:592) > [junit4] 2> at > org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdownInstance(D > efaultMetricsSystem.java:72) > [junit4] 2> at > org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdown(DefaultMe > tricsSystem.java:68) > [junit4] 2> at > org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics.shut > down(NameNodeMetrics.java:145) > [junit4] 2> at > org.apache.hadoop.hdfs.server.namenode.NameNode.stop(NameNode.jav > a:822) > [junit4] 2> at > org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:172 > 0) > [junit4] 2> at > org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:169 > 9) > [junit4] 2> at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j > ava:838) > [junit4] 2> at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738) > [junit4] 2> at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608) > [junit4] 2> at > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98) > [junit4] 2> at > org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT > est.java:44) > [junit4] 2> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) > [junit4] 2> at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j > ava:57) > [junit4] 2> at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces > sorImpl.java:43) > [junit4] 2> at java.lang.reflect.Method.invoke(Method.java:606) > [junit4] 2> at > com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(Randomize > dRunner.java:1627) > [junit4] 2> at > com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(Rando > mizedRunner.java:776) > [junit4] 2> at > com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(Rando > mizedRunner.java:792) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1. > evaluate(SystemPropertiesRestoreRule.java:57) > [junit4] 2> at > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeA > fterRule.java:46) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreCl > assName.java:42) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMet > hodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAss > ertionsRequired.java:54) > [junit4] 2> at > org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure > .java:48) > [junit4] 2> at > org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRule > IgnoreAfterMaxFailures.java:65) > [junit4] 2> at > org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnore > TestSuites.java:55) > [junit4] 2> at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stat > ementAdapter.java:36) > [junit4] 2> at > com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner. > run(ThreadLeakControl.java:365) > [junit4] 2> at java.lang.Thread.run(Thread.java:745) > [junit4] 2> Caused by: java.lang.NullPointerException > [junit4] 2> at > org.apache.hadoop.hdfs.server.blockmanagement.BlocksMap.size(BlocksM > ap.java:198) > [junit4] 2> at > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.getTotalBl > ocks(BlockManager.java:3291) > [junit4] 2> at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlocksTotal(F > SNamesystem.java:6223) > [junit4] 2> ... 58 more > [junit4] 2> 2645432 INFO (SUITE-HdfsNNFailoverTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.s.SolrTestCaseJ4 ###deleteCore > [junit4] 2> NOTE: test params are: codec=Asserting(Lucene53), > sim=RandomSimilarityProvider(queryNorm=true,coord=yes): {}, > locale=mk_MK, timezone=Asia/Shanghai > [junit4] 2> NOTE: SunOS 5.11 x86/Oracle Corporation 1.7.0_85 (32- > bit)/cpus=3,threads=1,free=99794816,total=518979584 > [junit4] 2> NOTE: All tests run in this JVM: [SolrCloudExampleTest, > TestStressVersions, TestSerializedLuceneMatchVersion, TestSolrJ, > DistanceUnitsTest, MultiThreadedOCPTest, TestDistribDocBasedVersion, > BJQParserTest, ZkCLITest, QueryEqualityTest, PrimitiveFieldTypeTest, > DistributedQueryComponentOptimizationTest, AliasIntegrationTest, > TestInitQParser, TestAuthorizationFramework, TestLazyCores, > SolrIndexConfigTest, TestFunctionQuery, TestXIncludeConfig, > HardAutoCommitTest, DocValuesMultiTest, TestDefaultStatsCache, > SolrRequestParserTest, RecoveryZkTest, UpdateParamsTest, > TestSolrDeletionPolicy1, TestDFRSimilarityFactory, TestFastWriter, > PathHierarchyTokenizerFactoryTest, TestDynamicLoading, > TestElisionMultitermQuery, PolyFieldTest, UnloadDistributedZkTest, > TestJsonRequest, TestRuleBasedAuthorizationPlugin, > TestManagedStopFilterFactory, TestRawResponseWriter, IndexSchemaTest, > TestEmbeddedSolrServerConstructors, InfoHandlerTest, > AlternateDirectoryTest, LeaderElectionTest, JsonLoaderTest, > TestCoreContainer, DirectSolrSpellCheckerTest, RequestLoggingTest, > ZkNodePropsTest, TermsComponentTest, TestConfig, > TestFieldTypeCollectionResource, XsltUpdateRequestHandlerTest, > TestManagedSchemaFieldResource, TestSchemaResource, > DataDrivenBlockJoinTest, TestExactStatsCache, TestConfigSetProperties, > DeleteLastCustomShardedReplicaTest, TestAnalyzedSuggestions, > DirectUpdateHandlerTest, ExternalFileFieldSortTest, TestIBSimilarityFactory, > TestMissingGroups, ClusterStateUpdateTest, ActionThrottleTest, > QueryElevationComponentTest, DocValuesTest, QueryResultKeyTest, > TestLRUCache, TestPhraseSuggestions, SimplePostToolTest, > TriLevelCompositeIdRoutingTest, DistributedMLTComponentTest, > CloudExitableDirectoryReaderTest, TestSolrCloudWithKerberosAlt, > TestCodecSupport, TestConfigSets, PeerSyncTest, > XmlUpdateRequestHandlerTest, SpatialHeatmapFacetsTest, > SoftAutoCommitTest, TestSchemaNameResource, > PreAnalyzedUpdateProcessorTest, TestJmxMonitoredMap, > TestDistributedStatsComponentCardinality, > TestManagedSynonymFilterFactory, JSONWriterTest, TestNRTOpen, > ReplicationFactorTest, DOMUtilTest, SolrCoreTest, > DocExpirationUpdateProcessorFactoryTest, FastVectorHighlighterTest, > SuggesterFSTTest, TestExtendedDismaxParser, TestSolrConfigHandler, > DocumentAnalysisRequestHandlerTest, > DistributedFacetPivotSmallAdvancedTest, BlockDirectoryTest, > TestQuerySenderNoQuery, TestHashPartitioner, DateFieldTest, > SegmentsInfoRequestHandlerTest, TestFieldCollectionResource, > RecoveryAfterSoftCommitTest, TestMergePolicyConfig, TestFieldSortValues, > SecurityConfHandlerTest, TestStressReorder, BufferStoreTest, > TestRandomRequestDistribution, HdfsBasicDistributedZkTest, > TestCloudManagedSchemaConcurrent, TestReplicaProperties, > DisMaxRequestHandlerTest, TestMacros, TestStressLucene, > TestReloadAndDeleteDocs, BasicAuthIntegrationTest, TestDocSet, > BasicDistributedZkTest, DistributedQueryElevationComponentTest, > TestGroupingSearch, TestObjectReleaseTracker, MoreLikeThisHandlerTest, > OverseerTest, TestFaceting, TestUpdate, TestClassNameShortening, > TestRestManager, SyncSliceTest, ShardRoutingTest, ZkSolrClientTest, > TestZkChroot, TestRandomDVFaceting, ShardRoutingCustomTest, > TestDistributedGrouping, DistributedSpellCheckComponentTest, > ZkControllerTest, TestRealTimeGet, TestReload, > DistributedTermsComponentTest, TestRangeQuery, SimpleFacetsTest, > TestSolr4Spatial, StatsComponentTest, SolrCmdDistributorTest, TestSort, > CurrencyFieldXmlFileTest, AnalysisAfterCoreReloadTest, > TestFoldingMultitermQuery, SuggesterTSTTest, TestCSVLoader, > SchemaVersionSpecificBehaviorTest, SolrCoreCheckLockOnStartupTest, > DirectUpdateHandlerOptimizeTest, > StatelessScriptUpdateProcessorFactoryTest, DistanceFunctionTest, > IndexBasedSpellCheckerTest, StandardRequestHandlerTest, > TestOmitPositions, DocumentBuilderTest, RequiredFieldsTest, > TestArbitraryIndexDir, LoggingHandlerTest, ReturnFieldsTest, > MBeansHandlerTest, UniqFieldsUpdateProcessorFactoryTest, > PingRequestHandlerTest, TestComponentsName, TestLFUCache, > PreAnalyzedFieldTest, TestSystemIdResolver, SpellingQueryConverterTest, > TestUtils, TestDocumentBuilder, SliceStateTest, SystemInfoHandlerTest, > UUIDFieldTest, FileUtilsTest, CircularListTest, TestRTGBase, > CursorPagingTest, DistributedIntervalFacetingTest, > TestDistributedMissingSort, TestSimpleTrackingShardHandler, > AsyncMigrateRouteKeyTest, DeleteInactiveReplicaTest, > DistribDocExpirationUpdateProcessorTest, LeaderFailoverAfterPartitionTest, > OverriddenZkACLAndCredentialsProvidersTest, > OverseerCollectionConfigSetProcessorTest, OverseerRolesTest, > OverseerTaskQueueTest, SSLMigrationTest, SaslZkACLProviderTest, > SimpleCollectionCreateDeleteTest, TestAuthenticationFramework, > TestCloudInspectUtil, TestCollectionAPI, TestMiniSolrCloudClusterSSL, > TestRebalanceLeaders, TestRequestStatusCollectionAPI, > HdfsBasicDistributedZk2Test, HdfsChaosMonkeySafeLeaderTest, > HdfsCollectionsAPIDistributedZkTest, HdfsNNFailoverTest] > [junit4] 2> NOTE: reproduce with: ant test - > Dtestcase=HdfsNNFailoverTest -Dtests.seed=5D8F351977870E3F - > Dtests.slow=true -Dtests.locale=mk_MK -Dtests.timezone=Asia/Shanghai - > Dtests.asserts=true -Dtests.file.encoding=UTF-8 > [junit4] ERROR 0.00s J0 | HdfsNNFailoverTest (suite) <<< > [junit4] > Throwable #1: java.lang.RuntimeException: Error while running > command to get file permissions : java.io.IOException: Cannot run program > "/bin/ls": error=12, Not enough space > [junit4] > at > java.lang.ProcessBuilder.start(ProcessBuilder.java:1047) > [junit4] > at > org.apache.hadoop.util.Shell.runCommand(Shell.java:485) > [junit4] > at org.apache.hadoop.util.Shell.run(Shell.java:455) > [junit4] > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715 > ) > [junit4] > at > org.apache.hadoop.util.Shell.execCommand(Shell.java:808) > [junit4] > at > org.apache.hadoop.util.Shell.execCommand(Shell.java:791) > [junit4] > at > org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:1097) > [junit4] > at > org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo > adPermissionInfo(RawLocalFileSystem.java:582) > [junit4] > at > org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g > etPermission(RawLocalFileSystem.java:557) > [junit4] > at > org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck( > DiskChecker.java:139) > [junit4] > at > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156) > [junit4] > at > org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker. > checkDir(DataNode.java:2239) > [junit4] > at > org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations( > DataNode.java:2281) > [junit4] > at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo > de.java:2263) > [junit4] > at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D > ataNode.java:2155) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav > a:1443) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j > ava:828) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608) > [junit4] > at > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98) > [junit4] > at > org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT > est.java:44) > [junit4] > at java.lang.Thread.run(Thread.java:745) > [junit4] > Caused by: java.io.IOException: error=12, Not enough space > [junit4] > at java.lang.UNIXProcess.forkAndExec(Native Method) > [junit4] > at java.lang.UNIXProcess.<init>(UNIXProcess.java:137) > [junit4] > at java.lang.ProcessImpl.start(ProcessImpl.java:130) > [junit4] > at > java.lang.ProcessBuilder.start(ProcessBuilder.java:1028) > [junit4] > ... 44 more > [junit4] > at > __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0) > [junit4] > at > org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.lo > adPermissionInfo(RawLocalFileSystem.java:620) > [junit4] > at > org.apache.hadoop.fs.RawLocalFileSystem$DeprecatedRawLocalFileStatus.g > etPermission(RawLocalFileSystem.java:557) > [junit4] > at > org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck( > DiskChecker.java:139) > [junit4] > at > org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156) > [junit4] > at > org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker. > checkDir(DataNode.java:2239) > [junit4] > at > org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations( > DataNode.java:2281) > [junit4] > at > org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNo > de.java:2263) > [junit4] > at > org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(D > ataNode.java:2155) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.jav > a:1443) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j > ava:828) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608) > [junit4] > at > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98) > [junit4] > at > org.apache.solr.cloud.hdfs.HdfsNNFailoverTest.setupClass(HdfsNNFailoverT > est.java:44) > [junit4] > at java.lang.Thread.run(Thread.java:745) > [junit4] Completed [426/536] on J0 in 45.66s, 0 tests, 1 error <<< > FAILURES! > > [...truncated 300 lines...] > [junit4] Suite: org.apache.solr.store.hdfs.HdfsDirectoryTest > [junit4] 2> Creating dataDir: /export/home/jenkins/workspace/Lucene- > Solr-5.x-Solaris/solr/build/solr- > core/test/J1/temp/solr.store.hdfs.HdfsDirectoryTest_5D8F351977870E3F- > 001/init-core-data-001 > [junit4] 2> 3147821 INFO (SUITE-HdfsDirectoryTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl > (false) and clientAuth (false) > [junit4] 1> Formatting using clusterid: testClusterID > [junit4] 2> 3147964 WARN (SUITE-HdfsDirectoryTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.h.m.i.MetricsConfig Cannot > locate configuration: tried hadoop-metrics2-namenode.properties,hadoop- > metrics2.properties > [junit4] 2> 3147974 WARN (SUITE-HdfsDirectoryTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.h.h.HttpRequestLog Jetty > request log can only be enabled using Log4j > [junit4] 2> 3147976 INFO (SUITE-HdfsDirectoryTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log jetty-6.1.26 > [junit4] 2> 3147994 INFO (SUITE-HdfsDirectoryTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log Extract > jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop- > hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/hdfs to > ./temp/Jetty_solaris.vm_46547_hdfs____.vwfmpk/webapp > [junit4] 2> 3148170 INFO (SUITE-HdfsDirectoryTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log NO JSP Support for /, did not > find org.apache.jasper.servlet.JspServlet > [junit4] 2> 3148982 INFO (SUITE-HdfsDirectoryTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log Started > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:46547 > [junit4] 2> 3157264 INFO (SUITE-HdfsDirectoryTest- > seed#[5D8F351977870E3F]-worker) [ ] o.m.log Stopped > HttpServer2$SelectChannelConnectorWithSafeStartup@solaris-vm:0 > [junit4] 2> 3157403 INFO (SUITE-HdfsDirectoryTest- > seed#[5D8F351977870E3F]-worker) [ ] o.a.s.SolrTestCaseJ4 ###deleteCore > [junit4] 2> Aug 29, 2015 9:01:07 PM > com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks > [junit4] 2> WARNING: Will linger awaiting termination of 1 leaked > thread(s). > [junit4] 2> Aug 29, 2015 9:01:27 PM > com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks > [junit4] 2> SEVERE: 1 thread leaked from SUITE scope at > org.apache.solr.store.hdfs.HdfsDirectoryTest: > [junit4] 2> 1) Thread[id=20389, name=IPC Server idle connection > scanner > for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest] > [junit4] 2> at java.lang.Object.wait(Native Method) > [junit4] 2> at java.lang.Object.wait(Object.java:503) > [junit4] 2> at java.util.TimerThread.mainLoop(Timer.java:526) > [junit4] 2> at java.util.TimerThread.run(Timer.java:505) > [junit4] 2> Aug 29, 2015 9:01:27 PM > com.carrotsearch.randomizedtesting.ThreadLeakControl tryToInterruptAll > [junit4] 2> INFO: Starting to interrupt leaked threads: > [junit4] 2> 1) Thread[id=20389, name=IPC Server idle connection > scanner > for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest] > [junit4] 2> Aug 29, 2015 9:01:30 PM > com.carrotsearch.randomizedtesting.ThreadLeakControl tryToInterruptAll > [junit4] 2> SEVERE: There are still zombie threads that couldn't be > terminated: > [junit4] 2> 1) Thread[id=20389, name=IPC Server idle connection > scanner > for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest] > [junit4] 2> at java.lang.Object.wait(Native Method) > [junit4] 2> at java.lang.Object.wait(Object.java:503) > [junit4] 2> at java.util.TimerThread.mainLoop(Timer.java:526) > [junit4] 2> at java.util.TimerThread.run(Timer.java:505) > [junit4] 2> NOTE: test params are: codec=Asserting(Lucene53): {}, > docValues:{}, sim=DefaultSimilarity, locale=es_BO, > timezone=Antarctica/South_Pole > [junit4] 2> NOTE: SunOS 5.11 x86/Oracle Corporation 1.7.0_85 (32- > bit)/cpus=3,threads=2,free=136627544,total=518979584 > [junit4] 2> NOTE: All tests run in this JVM: [TestIndexingPerformance, > TestCSVResponseWriter, DistributedQueryComponentCustomSortTest, > DirectSolrConnectionTest, FullSolrCloudDistribCmdsTest, > TestShardHandlerFactory, CacheHeaderTest, BasicZkTest, TestTrie, > FieldAnalysisRequestHandlerTest, PKIAuthenticationIntegrationTest, > OpenCloseCoreStressTest, TestSuggestSpellingConverter, StressHdfsTest, > CleanupOldIndexTest, DistributedExpandComponentTest, > TestHdfsUpdateLog, TestSolrXml, TestAddFieldRealTimeGet, TestJsonFacets, > DistributedSuggestComponentTest, > OutOfBoxZkACLAndCredentialsProvidersTest, AnalyticsMergeStrategyTest, > HLLUtilTest, ResponseHeaderTest, SearchHandlerTest, > BinaryUpdateRequestHandlerTest, DistributedFacetPivotWhiteBoxTest, > ConnectionManagerTest, SpellCheckComponentTest, > TestScoreJoinQPNoScore, SolrTestCaseJ4Test, SolrIndexSplitterTest, > TestConfigSetsAPI, TestDefaultSearchFieldResource, TestCryptoKeys, > TestNonDefinedSimilarityFactory, TestCoreDiscovery, RollingRestartTest, > SolrInfoMBeanTest, CustomCollectionTest, DistributedVersionInfoTest, > ClusterStateTest, TestReversedWildcardFilterFactory, SolrXmlInZkTest, > DistributedFacetPivotLongTailTest, URLClassifyProcessorTest, > TestLMJelinekMercerSimilarityFactory, RequestHandlersTest, > RemoteQueryErrorTest, LeaderElectionIntegrationTest, > SharedFSAutoReplicaFailoverTest, TestBadConfig, > SignatureUpdateProcessorFactoryTest, TestCursorMarkWithoutUniqueKey, > TestCrossCoreJoin, SparseHLLTest, DistributedQueueTest, > BigEndianAscendingWordSerializerTest, TestBM25SimilarityFactory, > AutoCommitTest, DateMathParserTest, BasicFunctionalityTest, > SuggesterWFSTTest, TestCollapseQParserPlugin, TestManagedResource, > TestSha256AuthenticationProvider, CollectionTooManyReplicasTest, > BadCopyFieldTest, TestDownShardTolerantSearch, CloudMLTQParserTest, > NotRequiredUniqueKeyTest, TestAnalyzeInfixSuggestions, > ExitableDirectoryReaderTest, TestScoreJoinQPScore, DeleteShardTest, > RankQueryTest, TestSchemaManager, UpdateRequestProcessorFactoryTest, > CursorMarkTest, DistributedDebugComponentTest, DeleteReplicaTest, > RAMDirectoryFactoryTest, ConcurrentDeleteAndCreateCollectionTest, > TestQueryTypes, OutputWriterTest, TestSchemaSimilarityResource, > HighlighterMaxOffsetTest, ResponseLogComponentTest, > TestCloudPivotFacet, DocValuesMissingTest, > FieldMutatingUpdateProcessorTest, HttpPartitionTest, TestCollationField, > ZkStateWriterTest, TestQuerySenderListener, AtomicUpdatesTest, > TestStressRecovery, TestRandomFaceting, > SharedFSAutoReplicaFailoverUtilsTest, CoreAdminHandlerTest, > HighlighterConfigTest, TestCustomSort, MultiTermTest, > VMParamsZkACLAndCredentialsProvidersTest, > IgnoreCommitOptimizeUpdateProcessorFactoryTest, CollectionReloadTest, > PrimUtilsTest, TestRecovery, TestWriterPerf, > AddSchemaFieldsUpdateProcessorFactoryTest, TimeZoneUtilsTest, > CurrencyFieldOpenExchangeTest, TestSolrCLIRunExample, > TestPHPSerializedResponseWriter, ChaosMonkeySafeLeaderTest, > TestIndexSearcher, EnumFieldTest, TestSolrIndexConfig, > TermVectorComponentDistributedTest, TestJoin, TestExpandComponent, > TestManagedResourceStorage, SortByFunctionTest, > TestDefaultSimilarityFactory, SuggesterTest, TestValueSourceCache, > SolrPluginUtilsTest, TermVectorComponentTest, TestFiltering, > TestQueryUtils, FileBasedSpellCheckerTest, BasicDistributedZk2Test, > CollectionsAPIDistributedZkTest, TestReplicationHandler, > TestDistributedSearch, BadIndexSchemaTest, ConvertedLegacyTest, > HighlighterTest, ShowFileRequestHandlerTest, SpellCheckCollatorTest, > SpatialFilterTest, NoCacheHeaderTest, WordBreakSolrSpellCheckerTest, > TestPseudoReturnFields, TestAtomicUpdateErrorCases, > TestWordDelimiterFilterFactory, DefaultValueUpdateProcessorTest, > TestRemoteStreaming, DebugComponentTest, TestSurroundQueryParser, > LukeRequestHandlerTest, TestSolrQueryParser, > IndexSchemaRuntimeFieldTest, RegexBoostProcessorTest, > TestJmxIntegration, QueryParsingTest, TestPartialUpdateDeduplication, > CSVRequestHandlerTest, TestBinaryResponseWriter, SOLR749Test, > CopyFieldTest, BadComponentTest, TestSolrDeletionPolicy2, SampleTest, > TestBinaryField, TestSearchPerf, NumericFieldsTest, MinimalSchemaTest, > TestFuzzyAnalyzedSuggestions, TestSolrCoreProperties, > TestPostingsSolrHighlighter, TestLuceneMatchVersion, > SpellPossibilityIteratorTest, TestCharFilters, SynonymTokenizerTest, > EchoParamsTest, TestSweetSpotSimilarityFactory, TestPerFieldSimilarity, > TestLMDirichletSimilarityFactory, ResourceLoaderTest, > TestFastOutputStream, ScriptEngineTest, > OpenExchangeRatesOrgProviderTest, PluginInfoTest, TestFastLRUCache, > ChaosMonkeyNothingIsSafeTest, TestHighlightDedupGrouping, > TestTolerantSearch, TestJettySolrRunner, AssignTest, > AsyncCallRequestStatusResponseTest, CollectionStateFormat2Test, > CollectionsAPIAsyncDistributedZkTest, DistribCursorPagingTest, > DistribJoinFromCollectionTest, LeaderInitiatedRecoveryOnCommitTest, > MigrateRouteKeyTest, OverseerStatusTest, ShardSplitTest, > TestConfigSetsAPIExclusivity, TestConfigSetsAPIZkFailure, > TestLeaderElectionZkExpiry, TestMiniSolrCloudCluster, > TestShortCircuitedRequests, HdfsRecoverLeaseTest, > CachingDirectoryFactoryTest, HdfsDirectoryFactoryTest, TestConfigOverlay, > TestConfigSetImmutable, TestImplicitCoreProperties, > TestInfoStreamLogging, TestInitParams, TestSolrDynamicMBean, > TestBlobHandler, TestConfigReload, TestReplicationHandlerBackup, > TestSolrConfigHandlerConcurrent, CoreAdminCreateDiscoverTest, > CoreAdminRequestStatusTest, CoreMergeIndexesAdminHandlerTest, > DistributedFacetPivotLargeTest, DistributedFacetPivotSmallTest, > FacetPivotSmallTest, SuggestComponentTest, JavabinLoaderTest, > SmileWriterTest, TestIntervalFaceting, TestChildDocTransformer, > TestCustomDocTransformer, TestSortingResponseWriter, > TestBulkSchemaAPI, TestFieldResource, > TestManagedSchemaDynamicFieldResource, TestBulkSchemaConcurrent, > TestCloudSchemaless, TestReloadDeadlock, TestSearcherReuse, > TestSimpleQParserPlugin, TestSmileRequest, TestSolr4Spatial2, > TestStandardQParsers, TestStressUserVersions, TestTrieFacet, > TestMinMaxOnMultiValuedField, TestOrdValues, > TestSortByMinMaxFunction, SimpleMLTQParserTest, TestDistribIDF, > TestExactSharedStatsCache, TestPKIAuthenticationPlugin, > TestBlendedInfixSuggestions, TestFileDictionaryLookup, > TestFreeTextSuggestions, TestHighFrequencyDictionaryFactory, > BlockCacheTest, HdfsDirectoryTest] > [junit4] 2> NOTE: reproduce with: ant test -Dtestcase=HdfsDirectoryTest > - > Dtests.seed=5D8F351977870E3F -Dtests.slow=true -Dtests.locale=es_BO - > Dtests.timezone=Antarctica/South_Pole -Dtests.asserts=true - > Dtests.file.encoding=UTF-8 > [junit4] ERROR 0.00s J1 | HdfsDirectoryTest (suite) <<< > [junit4] > Throwable #1: java.security.AccessControlException: access > denied ("java.io.FilePermission" "/export/home/jenkins/workspace/Lucene- > Solr-5.x-Solaris/solr/build/solr-core/test/J1" "write") > [junit4] > at > __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0) > [junit4] > at > java.security.AccessControlContext.checkPermission(AccessControlContext.j > ava:395) > [junit4] > at > java.security.AccessController.checkPermission(AccessController.java:559) > [junit4] > at > java.lang.SecurityManager.checkPermission(SecurityManager.java:549) > [junit4] > at > java.lang.SecurityManager.checkWrite(SecurityManager.java:979) > [junit4] > at java.io.File.canWrite(File.java:785) > [junit4] > at > org.apache.hadoop.fs.FileUtil.canWrite(FileUtil.java:1002) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.createPermissionsDiagnosisString( > MiniDFSCluster.java:856) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.j > ava:812) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:738) > [junit4] > at > org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:608) > [junit4] > at > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:98) > [junit4] > at > org.apache.solr.cloud.hdfs.HdfsTestUtil.setupClass(HdfsTestUtil.java:60) > [junit4] > at > org.apache.solr.store.hdfs.HdfsDirectoryTest.beforeClass(HdfsDirectoryTest. > java:62) > [junit4] > at java.lang.Thread.run(Thread.java:745)Throwable #2: > com.carrotsearch.randomizedtesting.ThreadLeakError: 1 thread leaked from > SUITE scope at org.apache.solr.store.hdfs.HdfsDirectoryTest: > [junit4] > 1) Thread[id=20389, name=IPC Server idle connection > scanner > for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest] > [junit4] > at java.lang.Object.wait(Native Method) > [junit4] > at java.lang.Object.wait(Object.java:503) > [junit4] > at java.util.TimerThread.mainLoop(Timer.java:526) > [junit4] > at java.util.TimerThread.run(Timer.java:505) > [junit4] > at > __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0)Throwable #3: > com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie > threads that couldn't be terminated: > [junit4] > 1) Thread[id=20389, name=IPC Server idle connection > scanner > for port 41610, state=WAITING, group=TGRP-HdfsDirectoryTest] > [junit4] > at java.lang.Object.wait(Native Method) > [junit4] > at java.lang.Object.wait(Object.java:503) > [junit4] > at java.util.TimerThread.mainLoop(Timer.java:526) > [junit4] > at java.util.TimerThread.run(Timer.java:505) > [junit4] > at > __randomizedtesting.SeedInfo.seed([5D8F351977870E3F]:0) > [junit4] Completed [521/536] on J1 in 33.05s, 0 tests, 3 errors <<< > FAILURES! > > [...truncated 64 lines...] > BUILD FAILED > /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:785: > The following error occurred while executing this line: > /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:729: > The following error occurred while executing this line: > /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/build.xml:59: The > following error occurred while executing this line: > /export/home/jenkins/workspace/Lucene-Solr-5.x- > Solaris/solr/build.xml:233: The following error occurred while executing this > line: > /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/solr/common- > build.xml:524: The following error occurred while executing this line: > /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/lucene/common- > build.xml:1452: The following error occurred while executing this line: > /export/home/jenkins/workspace/Lucene-Solr-5.x-Solaris/lucene/common- > build.xml:1006: There were test failures: 536 suites, 2123 tests, 4 > suite-level > errors, 108 ignored (34 assumptions) > > Total time: 77 minutes 51 seconds > Build step 'Invoke Ant' marked build as failure > Archiving artifacts > [WARNINGS] Skipping publisher since build result is FAILURE > Recording test results > Email was triggered for: Failure - Any > Sending email for trigger: Failure - Any > --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
