Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-7.x/389/
3 tests failed. FAILED: org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest.test Error Message: There are still nodes recoverying - waited for 320 seconds Stack Trace: java.lang.AssertionError: There are still nodes recoverying - waited for 320 seconds at __randomizedtesting.SeedInfo.seed([64E46779FF6ECF5F:ECB058A35192A2A7]:0) at org.junit.Assert.fail(Assert.java:93) at org.apache.solr.cloud.AbstractDistribZkTestBase.waitForRecoveriesToFinish(AbstractDistribZkTestBase.java:185) at org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForRecoveriesToFinish(AbstractFullDistribZkTestBase.java:920) at org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForThingsToLevelOut(AbstractFullDistribZkTestBase.java:1477) at org.apache.solr.cloud.RestartWhileUpdatingTest.test(RestartWhileUpdatingTest.java:145) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1742) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:935) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:971) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:985) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:1010) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:985) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:944) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:830) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:880) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.lang.Thread.run(Thread.java:748) FAILED: junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest Error Message: 7 threads leaked from SUITE scope at org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest: 1) Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 2) Thread[id=40196, name=searcherExecutor-12526-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 3) Thread[id=40237, name=searcherExecutor-12540-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 4) Thread[id=40318, name=searcherExecutor-12568-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 5) Thread[id=40399, name=searcherExecutor-12588-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 6) Thread[id=40274, name=searcherExecutor-12554-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 7) Thread[id=40434, name=searcherExecutor-12602-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Stack Trace: com.carrotsearch.randomizedtesting.ThreadLeakError: 7 threads leaked from SUITE scope at org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest: 1) Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 2) Thread[id=40196, name=searcherExecutor-12526-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 3) Thread[id=40237, name=searcherExecutor-12540-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 4) Thread[id=40318, name=searcherExecutor-12568-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 5) Thread[id=40399, name=searcherExecutor-12588-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 6) Thread[id=40274, name=searcherExecutor-12554-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 7) Thread[id=40434, name=searcherExecutor-12602-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) at __randomizedtesting.SeedInfo.seed([64E46779FF6ECF5F]:0) FAILED: junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest Error Message: There are still zombie threads that couldn't be terminated: 1) Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 2) Thread[id=40196, name=searcherExecutor-12526-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 3) Thread[id=40237, name=searcherExecutor-12540-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 4) Thread[id=40318, name=searcherExecutor-12568-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 5) Thread[id=40399, name=searcherExecutor-12588-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 6) Thread[id=40274, name=searcherExecutor-12554-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 7) Thread[id=40434, name=searcherExecutor-12602-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Stack Trace: com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie threads that couldn't be terminated: 1) Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 2) Thread[id=40196, name=searcherExecutor-12526-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 3) Thread[id=40237, name=searcherExecutor-12540-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 4) Thread[id=40318, name=searcherExecutor-12568-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 5) Thread[id=40399, name=searcherExecutor-12588-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 6) Thread[id=40274, name=searcherExecutor-12554-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 7) Thread[id=40434, name=searcherExecutor-12602-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) at __randomizedtesting.SeedInfo.seed([64E46779FF6ECF5F]:0) Build Log: [...truncated 15263 lines...] [junit4] Suite: org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest [junit4] 2> Creating dataDir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/init-core-data-001 [junit4] 2> 3455430 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true [junit4] 2> 3455433 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (true) via: @org.apache.solr.util.RandomizeSSL(reason=, ssl=NaN, value=NaN, clientAuth=NaN) [junit4] 2> 3455433 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom [junit4] 2> 3455433 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /pq_/z [junit4] 1> Formatting using clusterid: testClusterID [junit4] 2> 3455475 WARN (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties [junit4] 2> 3455481 WARN (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j [junit4] 2> 3455482 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.m.log jetty-6.1.26 [junit4] 2> 3455502 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs to ./temp/Jetty_localhost_38023_hdfs____.tovdyi/webapp [junit4] 2> 3455926 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:38023 [junit4] 2> 3456003 WARN (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j [junit4] 2> 3456005 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.m.log jetty-6.1.26 [junit4] 2> 3456025 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_36563_datanode____1av2u/webapp [junit4] 2> 3456424 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:36563 [junit4] 2> 3456483 WARN (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j [junit4] 2> 3456484 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.m.log jetty-6.1.26 [junit4] 2> 3456511 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_41352_datanode____rzdyvu/webapp [junit4] 2> 3456568 ERROR (DataNode: [[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data1/, [DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data2/]] heartbeating to localhost/127.0.0.1:34159) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000 [junit4] 2> 3456580 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xfe700a2551df0: from storage DS-8dc247b8-210c-4bb7-81f7-056a390b655d node DatanodeRegistration(127.0.0.1:35883, datanodeUuid=21573dcb-4239-46df-97c5-88f1d59ce34d, infoPort=35438, infoSecurePort=0, ipcPort=33352, storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs [junit4] 2> 3456580 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xfe700a2551df0: from storage DS-3c8bf2dd-d4d4-4ba3-89ef-97f75d414464 node DatanodeRegistration(127.0.0.1:35883, datanodeUuid=21573dcb-4239-46df-97c5-88f1d59ce34d, infoPort=35438, infoSecurePort=0, ipcPort=33352, storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs [junit4] 2> 3457073 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:41352 [junit4] 2> 3457173 ERROR (DataNode: [[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data3/, [DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data4/]] heartbeating to localhost/127.0.0.1:34159) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000 [junit4] 2> 3457185 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xfe700c65df150: from storage DS-12755719-cd60-4944-830c-54f7cb39a54f node DatanodeRegistration(127.0.0.1:46625, datanodeUuid=f2212e0f-6033-4d4b-8cc7-b19f374e3835, infoPort=43148, infoSecurePort=0, ipcPort=36074, storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs [junit4] 2> 3457185 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xfe700c65df150: from storage DS-d3aaa494-7d27-41ae-a4b5-5906476e046f node DatanodeRegistration(127.0.0.1:46625, datanodeUuid=f2212e0f-6033-4d4b-8cc7-b19f374e3835, infoPort=43148, infoSecurePort=0, ipcPort=36074, storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs [junit4] 2> 3457273 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER [junit4] 2> 3457274 INFO (Thread-12411) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0 [junit4] 2> 3457274 INFO (Thread-12411) [ ] o.a.s.c.ZkTestServer Starting server [junit4] 2> 3457276 ERROR (Thread-12411) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes [junit4] 2> 3457374 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.ZkTestServer start zk server on port:44631 [junit4] 2> 3457377 INFO (zkConnectionManagerCallback-10510-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3457382 INFO (zkConnectionManagerCallback-10512-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3457388 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml [junit4] 2> 3457391 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/schema15.xml to /configs/conf1/schema.xml [junit4] 2> 3457392 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml to /configs/conf1/solrconfig.snippet.randomindexconfig.xml [junit4] 2> 3457394 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt [junit4] 2> 3457396 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt [junit4] 2> 3457397 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml [junit4] 2> 3457399 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml to /configs/conf1/enumsConfig.xml [junit4] 2> 3457400 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json [junit4] 2> 3457402 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt [junit4] 2> 3457403 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt [junit4] 2> 3457405 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt [junit4] 2> 3457407 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractFullDistribZkTestBase Will use NRT replicas unless explicitly asked otherwise [junit4] 2> 3457590 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12 [junit4] 2> 3457591 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0 [junit4] 2> 3457591 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.session No SessionScavenger set, using defaults [junit4] 2> 3457591 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.session node0 Scavenging every 660000ms [junit4] 2> 3457594 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@2a9d160f{/pq_/z,null,AVAILABLE} [junit4] 2> 3457594 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@3f91f18a{HTTP/1.1,[http/1.1]}{127.0.0.1:36044} [junit4] 2> 3457594 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.Server Started @3457651ms [junit4] 2> 3457594 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=hdfs://localhost:34159/hdfs__localhost_34159__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J2_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001_tempDir-002_control_data, replicaType=NRT, hostContext=/pq_/z, hostPort=36044, coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores} [junit4] 2> 3457594 ERROR (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete. [junit4] 2> 3457594 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory [junit4] 2> 3457595 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.7.0 [junit4] 2> 3457595 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null [junit4] 2> 3457595 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null [junit4] 2> 3457595 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-11-27T17:55:33.178Z [junit4] 2> 3457597 INFO (zkConnectionManagerCallback-10514-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3457599 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper) [junit4] 2> 3457599 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/solr.xml [junit4] 2> 3457603 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored [junit4] 2> 3457603 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored [junit4] 2> 3457605 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3, but no JMX reporters were configured - adding default JMX reporter. [junit4] 2> 3457724 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44631/solr [junit4] 2> 3457726 INFO (zkConnectionManagerCallback-10518-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3457731 INFO (zkConnectionManagerCallback-10520-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3457867 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:36044_pq_%2Fz [junit4] 2> 3457868 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.Overseer Overseer (id=72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000) starting [junit4] 2> 3457876 INFO (zkConnectionManagerCallback-10527-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3457878 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready [junit4] 2> 3457879 INFO (OverseerStateUpdate-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.Overseer Starting to work on the main queue : 127.0.0.1:36044_pq_%2Fz [junit4] 2> 3457879 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:36044_pq_%2Fz [junit4] 2> 3457888 INFO (OverseerStateUpdate-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 3457891 INFO (zkCallback-10526-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 3457896 DEBUG (OverseerAutoScalingTriggerThread-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000) [ ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial livenodes: [127.0.0.1:36044_pq_%2Fz] [junit4] 2> 3457899 DEBUG (ScheduledTrigger-12519-thread-2) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1 [junit4] 2> 3457911 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory. [junit4] 2> 3457941 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3 [junit4] 2> 3457958 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3 [junit4] 2> 3457958 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3 [junit4] 2> 3457960 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores [junit4] 2> 3457998 INFO (zkConnectionManagerCallback-10533-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3457999 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 3458001 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready [junit4] 2> 3458002 INFO (qtp127098813-40156) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:36044_pq_%252Fz&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 3458005 INFO (OverseerThreadFactory-12521-thread-1-processing-n:127.0.0.1:36044_pq_%2Fz) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.a.c.CreateCollectionCmd Create collection control_collection [junit4] 2> 3458112 INFO (qtp127098813-40151) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core} status=0 QTime=0 [junit4] 2> 3458116 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz x:control_collection_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT [junit4] 2> 3458117 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz x:control_collection_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores [junit4] 2> 3458902 DEBUG (ScheduledTrigger-12519-thread-3) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1 [junit4] 2> 3459143 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.7.0 [junit4] 2> 3459162 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test [junit4] 2> 3459268 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id [junit4] 2> 3459289 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1' using configuration from collection control_collection, trusted=true [junit4] 2> 3459290 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.control_collection.shard1.replica_n1' (registry 'solr.core.control_collection.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3 [junit4] 2> 3459290 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore at [/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores/control_collection_shard1_replica_n1], dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores/control_collection_shard1_replica_n1/data/] [junit4] 2> 3459295 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=47, maxMergeAtOnceExplicit=22, maxMergedSegmentMB=1.4248046875, floorSegmentMB=0.9541015625, forceMergeDeletesPctAllowed=16.83237086906014, segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.4209624511310861, deletesPctAllowed=39.50145037974311 [junit4] 2> 3459304 WARN (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}} [junit4] 2> 3459383 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog [junit4] 2> 3459383 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536 [junit4] 2> 3459385 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: if uncommitted for 30000ms; [junit4] 2> 3459385 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: if uncommitted for 3000ms; [junit4] 2> 3459387 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=41, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.0] [junit4] 2> 3459388 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@1dd69426[control_collection_shard1_replica_n1] main] [junit4] 2> 3459389 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1 [junit4] 2> 3459390 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1 [junit4] 2> 3459391 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms. [junit4] 2> 3459391 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1618310683661697024 [junit4] 2> 3459393 INFO (searcherExecutor-12526-thread-1-processing-n:127.0.0.1:36044_pq_%2Fz x:control_collection_shard1_replica_n1 c:control_collection s:shard1) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [control_collection_shard1_replica_n1] Registered new searcher Searcher@1dd69426[control_collection_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())} [junit4] 2> 3459401 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/control_collection/terms/shard1 to Terms{values={core_node2=0}, version=0} [junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue. [junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync [junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:36044/pq_/z/control_collection_shard1_replica_n1/ [junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me [junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy http://127.0.0.1:36044/pq_/z/control_collection_shard1_replica_n1/ has no replicas [junit4] 2> 3459406 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR [junit4] 2> 3459412 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:36044/pq_/z/control_collection_shard1_replica_n1/ shard1 [junit4] 2> 3459515 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary [junit4] 2> 3459523 INFO (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1406 [junit4] 2> 3459526 INFO (qtp127098813-40156) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas [junit4] 2> 3459618 INFO (zkCallback-10519-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/control_collection/state.json] for collection [control_collection] has occurred - updating... (live nodes size: [1]) [junit4] 2> 3459902 DEBUG (ScheduledTrigger-12519-thread-4) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1 [junit4] 2> 3460008 INFO (OverseerCollectionConfigSetProcessor-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper [junit4] 2> 3460527 INFO (qtp127098813-40156) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:36044_pq_%252Fz&wt=javabin&version=2} status=0 QTime=2524 [junit4] 2> 3460532 INFO (zkConnectionManagerCallback-10538-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3460534 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 3460535 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready [junit4] 2> 3460535 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection loss:false [junit4] 2> 3460536 INFO (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 3460539 INFO (OverseerThreadFactory-12521-thread-2-processing-n:127.0.0.1:36044_pq_%2Fz) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.a.c.CreateCollectionCmd Create collection collection1 [junit4] 2> 3460744 WARN (OverseerThreadFactory-12521-thread-2-processing-n:127.0.0.1:36044_pq_%2Fz) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.a.c.CreateCollectionCmd It is unusual to create a collection (collection1) without cores. [junit4] 2> 3460747 INFO (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas [junit4] 2> 3460748 INFO (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2} status=0 QTime=211 [junit4] 2> 3460902 DEBUG (ScheduledTrigger-12519-thread-3) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1 [junit4] 2> 3460927 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001 of type NRT [junit4] 2> 3460931 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12 [junit4] 2> 3460933 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0 [junit4] 2> 3460933 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.session No SessionScavenger set, using defaults [junit4] 2> 3460934 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.session node0 Scavenging every 660000ms [junit4] 2> 3460934 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@4682a57{/pq_/z,null,AVAILABLE} [junit4] 2> 3460934 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@325d81d3{HTTP/1.1,[http/1.1]}{127.0.0.1:44392} [junit4] 2> 3460934 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.Server Started @3460991ms [junit4] 2> 3460934 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=hdfs://localhost:34159/hdfs__localhost_34159__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J2_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001_tempDir-002_jetty1, replicaType=NRT, solrconfig=solrconfig.xml, hostContext=/pq_/z, hostPort=44392, coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores} [junit4] 2> 3460935 ERROR (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete. [junit4] 2> 3460937 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory [junit4] 2> 3460937 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.7.0 [junit4] 2> 3460937 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null [junit4] 2> 3460937 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null [junit4] 2> 3460937 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-11-27T17:55:36.520Z [junit4] 2> 3460939 INFO (zkConnectionManagerCallback-10540-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3460940 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper) [junit4] 2> 3460940 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/solr.xml [junit4] 2> 3460945 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored [junit4] 2> 3460946 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored [junit4] 2> 3460948 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3, but no JMX reporters were configured - adding default JMX reporter. [junit4] 2> 3461155 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44631/solr [junit4] 2> 3461157 INFO (zkConnectionManagerCallback-10544-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3461159 INFO (zkConnectionManagerCallback-10546-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3461168 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 3461173 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores [junit4] 2> 3461174 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:44392_pq_%2Fz [junit4] 2> 3461175 INFO (zkCallback-10526-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 3461182 INFO (zkCallback-10519-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 3461182 INFO (zkCallback-10537-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 3461182 INFO (zkCallback-10545-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 3461204 INFO (zkConnectionManagerCallback-10553-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3461205 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2) [junit4] 2> 3461206 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready [junit4] 2> 3461206 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory. [junit4] 2> 3461237 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3 [junit4] 2> 3461255 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3 [junit4] 2> 3461255 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3 [junit4] 2> 3461258 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores [junit4] 2> 3461299 INFO (qtp127098813-40151) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:44392_pq_%252Fz&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 3461303 INFO (OverseerCollectionConfigSetProcessor-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000002 doesn't exist. Requestor may have disconnected from ZooKeeper [junit4] 2> 3461313 INFO (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics params={wt=javabin&version=2&key=solr.core.control_collection.shard1.replica_n1:INDEX.sizeInBytes} status=0 QTime=2 [junit4] 2> 3461315 INFO (qtp127098813-40157) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core} status=0 QTime=1 [junit4] 2> 3461318 INFO (qtp641447673-40209) [n:127.0.0.1:44392_pq_%2Fz ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core} status=0 QTime=0 [junit4] 2> 3461319 INFO (OverseerThreadFactory-12521-thread-3-processing-n:127.0.0.1:36044_pq_%2Fz) [n:127.0.0.1:36044_pq_%2Fz c:collection1 s:shard1 ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:44392_pq_%2Fz for creating new replica of shard shard1 for collection collection1 [junit4] 2> 3461324 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz x:collection1_shard1_replica_n21] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT [junit4] 2> 3461903 DEBUG (ScheduledTrigger-12519-thread-4) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2 [junit4] 2> 3462345 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.7.0 [junit4] 2> 3462361 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.s.IndexSchema [collection1_shard1_replica_n21] Schema name=test [junit4] 2> 3462456 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id [junit4] 2> 3462474 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard1_replica_n21' using configuration from collection collection1, trusted=true [junit4] 2> 3462475 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard1.replica_n21' (registry 'solr.core.collection1.shard1.replica_n21') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3 [junit4] 2> 3462475 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SolrCore [[collection1_shard1_replica_n21] ] Opening new SolrCore at [/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores/collection1_shard1_replica_n21], dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores/collection1_shard1_replica_n21/data/] [junit4] 2> 3462480 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=47, maxMergeAtOnceExplicit=22, maxMergedSegmentMB=1.4248046875, floorSegmentMB=0.9541015625, forceMergeDeletesPctAllowed=16.83237086906014, segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.4209624511310861, deletesPctAllowed=39.50145037974311 [junit4] 2> 3462497 WARN (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}} [junit4] 2> 3462568 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog [junit4] 2> 3462569 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536 [junit4] 2> 3462570 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.CommitTracker Hard AutoCommit: if uncommitted for 30000ms; [junit4] 2> 3462570 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.CommitTracker Soft AutoCommit: if uncommitted for 3000ms; [junit4] 2> 3462573 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=41, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.0] [junit4] 2> 3462573 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.s.SolrIndexSearcher Opening [Searcher@5fc5a521[collection1_shard1_replica_n21] main] [junit4] 2> 3462575 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1 [junit4] 2> 3462575 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1 [junit4] 2> 3462576 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms. [junit4] 2> 3462577 INFO (searcherExecutor-12540-thread-1-processing-n:127.0.0.1:44392_pq_%2Fz x:collection1_shard1_replica_n21 c:collection1 s:shard1) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SolrCore [collection1_shard1_replica_n21] Registered new searcher Searcher@5fc5a521[collection1_shard1_replica_n21] main{ExitableDirectoryReader(UninvertingDirectoryReader())} [junit4] 2> 3462577 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1618310687002460160 [junit4] 2> 3462584 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard1 to Terms{values={core_node22=0}, version=0} [junit4] 2> 3462588 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue. [junit4] 2> 3462588 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync [junit4] 2> 3462588 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:44392/pq_/z/collection1_shard1_replica_n21/ [junit4] 2> 3462588 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me [junit4] 2> 3462589 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy http://127.0.0.1:44392/pq_/z/collection1_shard1_replica_n21/ has no replicas [junit4] 2> 3462589 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR [junit4] 2> 3462593 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:44392/pq_/z/collection1_shard1_replica_n21/ shard1 [junit4] 2> 3462696 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ZkController I am the leader, no recovery necessary [junit4] 2> 3462699 INFO (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1375 [junit4] 2> 3462702 INFO (qtp127098813-40151) [n:127.0.0.1:36044_pq_%2Fz c:collection1 ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={node=127.0.0.1:44392_pq_%252Fz&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2} status=0 QTime=1402 [junit4] 2> 3462799 INFO (zkCallback-10545-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2]) [junit4] 2> 3462882 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-2-001 of type NRT [junit4] 2> 3462883 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12 [junit4] 2> 3462889 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0 [junit4] 2> 3462889 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.session No SessionScavenger set, using defaults [junit4] 2> 3462889 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.session node0 Scavenging every 660000ms [junit4] 2> 3462889 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@1aaa752d{/pq_/z,null,AVAILABLE} [junit4] 2> 3462889 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@70cb9b0f{HTTP/1.1,[http/1.1]}{127.0.0.1:35419} [junit4] 2> 3462889 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.e.j.s.Server Started @3462946ms [junit4] 2> 3462889 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=hdfs://localhost:34159/hdfs__localhost_34159__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J2_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001_tempDir-002_jetty2, solrconfig=solrconfig.xml, hostContext=/pq_/z, hostPort=35419, coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-2-001/cores} [junit4] 2> 3462890 ERROR (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete. [junit4] 2> 3462891 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory [junit4] 2> 3462891 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.7.0 [junit4] 2> 3462891 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null [junit4] 2> 3462891 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null [junit4] 2> 3462891 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-11-27T17:55:38.474Z [junit4] 2> 3462893 INFO (zkConnectionManagerCallback-10556-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3462895 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper) [junit4] 2> 3462895 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-2-001/solr.xml [junit4] 2> 3462899 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored [junit4] 2> 3462899 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored [junit4] 2> 3462901 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3, but no JMX reporters were configured - adding default JMX reporter. [junit4] 2> 3462904 DEBUG (ScheduledTrigger-12519-thread-3) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2 [junit4] 2> 3463304 INFO (OverseerCollectionConfigSetProcessor-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000) [n:127.0.0.1:36044_pq_%2Fz ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000004 doesn't exist. Requestor may have disconnected from ZooKeeper [junit4] 2> 3463361 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44631/solr [junit4] 2> 3463363 INFO (zkConnectionManagerCallback-10560-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3463365 INFO (zkConnectionManagerCallback-10562-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3463374 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2) [junit4] 2> 3463379 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores [junit4] 2> 3463379 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:35419_pq_%2Fz [junit4] 2> 3463381 INFO (zkCallback-10526-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 3463381 INFO (zkCallback-10537-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 3463381 INFO (zkCallback-10552-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 3463384 INFO (zkCallback-10545-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 3463384 INFO (zkCallback-10519-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 3463388 INFO (zkCallback-10561-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 3463409 INFO (zkConnectionManagerCallback-10569-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 3463411 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3) [junit4] 2> 3463412 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready [junit4] 2> 3463412 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory. [junit4] 2> 3463445 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3 [junit4] 2> 3463462 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3 [junit4] 2> 3463462 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3 [junit4] 2> 3463465 INFO (TEST-HdfsRestartWhileUpdatingTest. [...truncated too long message...] kout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: ivy-availability-check: [loadresource] Do not set property disallowed.ivy.jars.list as its length is 0. -ivy-fail-disallowed-ivy-version: ivy-fail: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml resolve: jar-checksums: [mkdir] Created dir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/null1170095894 [copy] Copying 241 files to /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/null1170095894 [delete] Deleting directory /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/null1170095894 check-working-copy: [ivy:cachepath] :: resolving dependencies :: org.eclipse.jgit#org.eclipse.jgit-caller;working [ivy:cachepath] confs: [default] [ivy:cachepath] found org.eclipse.jgit#org.eclipse.jgit;4.6.0.201612231935-r in public [ivy:cachepath] found com.jcraft#jsch;0.1.53 in public [ivy:cachepath] found com.googlecode.javaewah#JavaEWAH;1.1.6 in public [ivy:cachepath] found org.apache.httpcomponents#httpclient;4.3.6 in public [ivy:cachepath] found org.apache.httpcomponents#httpcore;4.3.3 in public [ivy:cachepath] found commons-logging#commons-logging;1.1.3 in public [ivy:cachepath] found commons-codec#commons-codec;1.6 in public [ivy:cachepath] found org.slf4j#slf4j-api;1.7.2 in public [ivy:cachepath] :: resolution report :: resolve 29ms :: artifacts dl 2ms --------------------------------------------------------------------- | | modules || artifacts | | conf | number| search|dwnlded|evicted|| number|dwnlded| --------------------------------------------------------------------- | default | 8 | 0 | 0 | 0 || 8 | 0 | --------------------------------------------------------------------- [wc-checker] Initializing working copy... [wc-checker] SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". [wc-checker] SLF4J: Defaulting to no-operation (NOP) logger implementation [wc-checker] SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details. [wc-checker] Checking working copy status... -jenkins-base: BUILD SUCCESSFUL Total time: 285 minutes 41 seconds Archiving artifacts java.lang.InterruptedException: no matches found within 10000 at hudson.FilePath$34.hasMatch(FilePath.java:2678) at hudson.FilePath$34.invoke(FilePath.java:2557) at hudson.FilePath$34.invoke(FilePath.java:2547) at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2918) Also: hudson.remoting.Channel$CallSiteStackTrace: Remote call to lucene at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1741) at hudson.remoting.UserRequest$ExceptionResponse.retrieve(UserRequest.java:357) at hudson.remoting.Channel.call(Channel.java:955) at hudson.FilePath.act(FilePath.java:1036) at hudson.FilePath.act(FilePath.java:1025) at hudson.FilePath.validateAntFileMask(FilePath.java:2547) at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243) at hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81) at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20) at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744) at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690) at hudson.model.Build$BuildExecution.post2(Build.java:186) at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635) at hudson.model.Run.execute(Run.java:1819) at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43) at hudson.model.ResourceController.execute(ResourceController.java:97) at hudson.model.Executor.run(Executor.java:429) Caused: hudson.FilePath$TunneledInterruptedException at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2920) at hudson.remoting.UserRequest.perform(UserRequest.java:212) at hudson.remoting.UserRequest.perform(UserRequest.java:54) at hudson.remoting.Request$2.run(Request.java:369) at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:748) Caused: java.lang.InterruptedException: java.lang.InterruptedException: no matches found within 10000 at hudson.FilePath.act(FilePath.java:1038) at hudson.FilePath.act(FilePath.java:1025) at hudson.FilePath.validateAntFileMask(FilePath.java:2547) at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243) at hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81) at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20) at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744) at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690) at hudson.model.Build$BuildExecution.post2(Build.java:186) at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635) at hudson.model.Run.execute(Run.java:1819) at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43) at hudson.model.ResourceController.execute(ResourceController.java:97) at hudson.model.Executor.run(Executor.java:429) No artifacts found that match the file pattern "**/*.events,heapdumps/**,**/hs_err_pid*". Configuration error? Recording test results Build step 'Publish JUnit test result report' changed build result to UNSTABLE Email was triggered for: Unstable (Test Failures) Sending email for trigger: Unstable (Test Failures)
--------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org