Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-7.x/389/

3 tests failed.
FAILED:  org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest.test

Error Message:
There are still nodes recoverying - waited for 320 seconds

Stack Trace:
java.lang.AssertionError: There are still nodes recoverying - waited for 320 
seconds
        at 
__randomizedtesting.SeedInfo.seed([64E46779FF6ECF5F:ECB058A35192A2A7]:0)
        at org.junit.Assert.fail(Assert.java:93)
        at 
org.apache.solr.cloud.AbstractDistribZkTestBase.waitForRecoveriesToFinish(AbstractDistribZkTestBase.java:185)
        at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForRecoveriesToFinish(AbstractFullDistribZkTestBase.java:920)
        at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForThingsToLevelOut(AbstractFullDistribZkTestBase.java:1477)
        at 
org.apache.solr.cloud.RestartWhileUpdatingTest.test(RestartWhileUpdatingTest.java:145)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1742)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:935)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:971)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:985)
        at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:1010)
        at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:985)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:944)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:830)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:880)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:891)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at java.lang.Thread.run(Thread.java:748)


FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest

Error Message:
7 threads leaked from SUITE scope at 
org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest:     1) 
Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    2) Thread[id=40196, 
name=searcherExecutor-12526-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    3) Thread[id=40237, 
name=searcherExecutor-12540-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    4) Thread[id=40318, 
name=searcherExecutor-12568-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    5) Thread[id=40399, 
name=searcherExecutor-12588-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    6) Thread[id=40274, 
name=searcherExecutor-12554-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    7) Thread[id=40434, 
name=searcherExecutor-12602-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: 7 threads leaked from SUITE 
scope at org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest: 
   1) Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   2) Thread[id=40196, name=searcherExecutor-12526-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   3) Thread[id=40237, name=searcherExecutor-12540-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   4) Thread[id=40318, name=searcherExecutor-12568-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   5) Thread[id=40399, name=searcherExecutor-12588-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   6) Thread[id=40274, name=searcherExecutor-12554-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   7) Thread[id=40434, name=searcherExecutor-12602-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
        at __randomizedtesting.SeedInfo.seed([64E46779FF6ECF5F]:0)


FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest

Error Message:
There are still zombie threads that couldn't be terminated:    1) 
Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    2) Thread[id=40196, 
name=searcherExecutor-12526-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    3) Thread[id=40237, 
name=searcherExecutor-12540-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    4) Thread[id=40318, 
name=searcherExecutor-12568-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    5) Thread[id=40399, 
name=searcherExecutor-12588-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    6) Thread[id=40274, 
name=searcherExecutor-12554-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    7) Thread[id=40434, 
name=searcherExecutor-12602-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie 
threads that couldn't be terminated:
   1) Thread[id=40469, name=searcherExecutor-12616-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   2) Thread[id=40196, name=searcherExecutor-12526-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   3) Thread[id=40237, name=searcherExecutor-12540-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   4) Thread[id=40318, name=searcherExecutor-12568-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   5) Thread[id=40399, name=searcherExecutor-12588-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   6) Thread[id=40274, name=searcherExecutor-12554-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   7) Thread[id=40434, name=searcherExecutor-12602-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
        at __randomizedtesting.SeedInfo.seed([64E46779FF6ECF5F]:0)




Build Log:
[...truncated 15263 lines...]
   [junit4] Suite: org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest
   [junit4]   2> Creating dataDir: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/init-core-data-001
   [junit4]   2> 3455430 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) 
w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 3455433 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (true) via: 
@org.apache.solr.util.RandomizeSSL(reason=, ssl=NaN, value=NaN, clientAuth=NaN)
   [junit4]   2> 3455433 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: 
test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> 3455433 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /pq_/z
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 3455475 WARN  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried 
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 3455481 WARN  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 3455482 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.m.log jetty-6.1.26
   [junit4]   2> 3455502 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs
 to ./temp/Jetty_localhost_38023_hdfs____.tovdyi/webapp
   [junit4]   2> 3455926 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:38023
   [junit4]   2> 3456003 WARN  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 3456005 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.m.log jetty-6.1.26
   [junit4]   2> 3456025 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode
 to ./temp/Jetty_localhost_36563_datanode____1av2u/webapp
   [junit4]   2> 3456424 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:36563
   [junit4]   2> 3456483 WARN  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 3456484 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.m.log jetty-6.1.26
   [junit4]   2> 3456511 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode
 to ./temp/Jetty_localhost_41352_datanode____rzdyvu/webapp
   [junit4]   2> 3456568 ERROR (DataNode: 
[[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data1/,
 
[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data2/]]
  heartbeating to localhost/127.0.0.1:34159) [    ] 
o.a.h.h.s.d.DirectoryScanner 
dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 
ms/sec. Assuming default value of 1000
   [junit4]   2> 3456580 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0xfe700a2551df0: from storage 
DS-8dc247b8-210c-4bb7-81f7-056a390b655d node 
DatanodeRegistration(127.0.0.1:35883, 
datanodeUuid=21573dcb-4239-46df-97c5-88f1d59ce34d, infoPort=35438, 
infoSecurePort=0, ipcPort=33352, 
storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0, 
hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 3456580 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0xfe700a2551df0: from storage 
DS-3c8bf2dd-d4d4-4ba3-89ef-97f75d414464 node 
DatanodeRegistration(127.0.0.1:35883, 
datanodeUuid=21573dcb-4239-46df-97c5-88f1d59ce34d, infoPort=35438, 
infoSecurePort=0, ipcPort=33352, 
storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0, 
hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 3457073 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[64E46779FF6ECF5F]-worker) [    ] 
o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:41352
   [junit4]   2> 3457173 ERROR (DataNode: 
[[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data3/,
 
[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/tempDir-001/hdfsBaseDir/data/data4/]]
  heartbeating to localhost/127.0.0.1:34159) [    ] 
o.a.h.h.s.d.DirectoryScanner 
dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 
ms/sec. Assuming default value of 1000
   [junit4]   2> 3457185 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0xfe700c65df150: from storage 
DS-12755719-cd60-4944-830c-54f7cb39a54f node 
DatanodeRegistration(127.0.0.1:46625, 
datanodeUuid=f2212e0f-6033-4d4b-8cc7-b19f374e3835, infoPort=43148, 
infoSecurePort=0, ipcPort=36074, 
storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0, 
hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 3457185 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0xfe700c65df150: from storage 
DS-d3aaa494-7d27-41ae-a4b5-5906476e046f node 
DatanodeRegistration(127.0.0.1:46625, 
datanodeUuid=f2212e0f-6033-4d4b-8cc7-b19f374e3835, infoPort=43148, 
infoSecurePort=0, ipcPort=36074, 
storageInfo=lv=-56;cid=testClusterID;nsid=1258357046;c=0), blocks: 0, 
hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 3457273 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 3457274 INFO  (Thread-12411) [    ] o.a.s.c.ZkTestServer 
client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 3457274 INFO  (Thread-12411) [    ] o.a.s.c.ZkTestServer 
Starting server
   [junit4]   2> 3457276 ERROR (Thread-12411) [    ] o.a.z.s.ZooKeeperServer 
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action 
on ERROR or SHUTDOWN server state changes
   [junit4]   2> 3457374 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.ZkTestServer start zk server on port:44631
   [junit4]   2> 3457377 INFO  (zkConnectionManagerCallback-10510-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3457382 INFO  (zkConnectionManagerCallback-10512-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3457388 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
 to /configs/conf1/solrconfig.xml
   [junit4]   2> 3457391 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/schema15.xml
 to /configs/conf1/schema.xml
   [junit4]   2> 3457392 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
 to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 3457394 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
 to /configs/conf1/stopwords.txt
   [junit4]   2> 3457396 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/protwords.txt
 to /configs/conf1/protwords.txt
   [junit4]   2> 3457397 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/currency.xml
 to /configs/conf1/currency.xml
   [junit4]   2> 3457399 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml
 to /configs/conf1/enumsConfig.xml
   [junit4]   2> 3457400 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json
 to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 3457402 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt
 to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 3457403 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt
 to /configs/conf1/old_synonyms.txt
   [junit4]   2> 3457405 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/synonyms.txt
 to /configs/conf1/synonyms.txt
   [junit4]   2> 3457407 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractFullDistribZkTestBase Will use NRT replicas unless explicitly 
asked otherwise
   [junit4]   2> 3457590 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: 
d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12
   [junit4]   2> 3457591 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 3457591 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 3457591 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 3457594 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@2a9d160f{/pq_/z,null,AVAILABLE}
   [junit4]   2> 3457594 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@3f91f18a{HTTP/1.1,[http/1.1]}{127.0.0.1:36044}
   [junit4]   2> 3457594 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.Server Started @3457651ms
   [junit4]   2> 3457594 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: 
{solr.data.dir=hdfs://localhost:34159/hdfs__localhost_34159__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J2_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001_tempDir-002_control_data,
 replicaType=NRT, hostContext=/pq_/z, hostPort=36044, 
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores}
   [junit4]   2> 3457594 ERROR 
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 3457594 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter Using logger factory 
org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 3457595 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 
7.7.0
   [junit4]   2> 3457595 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 3457595 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 3457595 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2018-11-27T17:55:33.178Z
   [junit4]   2> 3457597 INFO  (zkConnectionManagerCallback-10514-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3457599 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in 
ZooKeeper)
   [junit4]   2> 3457599 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig Loading container configuration from 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/solr.xml
   [junit4]   2> 3457603 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay 
is ignored
   [junit4]   2> 3457603 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig Configuration parameter 
autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 3457605 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 3457724 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44631/solr
   [junit4]   2> 3457726 INFO  (zkConnectionManagerCallback-10518-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3457731 INFO  (zkConnectionManagerCallback-10520-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3457867 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.OverseerElectionContext I am going to 
be the leader 127.0.0.1:36044_pq_%2Fz
   [junit4]   2> 3457868 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.Overseer Overseer 
(id=72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000) starting
   [junit4]   2> 3457876 INFO  (zkConnectionManagerCallback-10527-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3457878 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.s.i.ZkClientClusterStateProvider 
Cluster at 127.0.0.1:44631/solr ready
   [junit4]   2> 3457879 INFO  
(OverseerStateUpdate-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.Overseer Starting to work on the main 
queue : 127.0.0.1:36044_pq_%2Fz
   [junit4]   2> 3457879 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:36044_pq_%2Fz
   [junit4]   2> 3457888 INFO  
(OverseerStateUpdate-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.c.ZkStateReader Updated live nodes from 
ZooKeeper... (0) -> (1)
   [junit4]   2> 3457891 INFO  (zkCallback-10526-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 3457896 DEBUG 
(OverseerAutoScalingTriggerThread-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000)
 [    ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial 
livenodes: [127.0.0.1:36044_pq_%2Fz]
   [junit4]   2> 3457899 DEBUG (ScheduledTrigger-12519-thread-2) [    ] 
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with 
currently live nodes: 1
   [junit4]   2> 3457911 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.h.a.MetricsHistoryHandler No .system 
collection, keeping metrics history in memory.
   [junit4]   2> 3457941 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.node' (registry 'solr.node') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
   [junit4]   2> 3457958 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jvm' (registry 'solr.jvm') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
   [junit4]   2> 3457958 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jetty' (registry 'solr.jetty') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
   [junit4]   2> 3457960 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.CorePropertiesLocator Found 0 core 
definitions underneath 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores
   [junit4]   2> 3457998 INFO  (zkConnectionManagerCallback-10533-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3457999 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 3458001 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready
   [junit4]   2> 3458002 INFO  (qtp127098813-40156) [n:127.0.0.1:36044_pq_%2Fz  
  ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params 
collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:36044_pq_%252Fz&wt=javabin&version=2
 and sendToOCPQueue=true
   [junit4]   2> 3458005 INFO  
(OverseerThreadFactory-12521-thread-1-processing-n:127.0.0.1:36044_pq_%2Fz) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.a.c.CreateCollectionCmd Create 
collection control_collection
   [junit4]   2> 3458112 INFO  (qtp127098813-40151) [n:127.0.0.1:36044_pq_%2Fz  
  ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics 
params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core}
 status=0 QTime=0
   [junit4]   2> 3458116 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz  
  x:control_collection_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core 
create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 3458117 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz  
  x:control_collection_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault 
Allocating transient cache for 4 transient cores
   [junit4]   2> 3458902 DEBUG (ScheduledTrigger-12519-thread-3) [    ] 
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with 
currently live nodes: 1
   [junit4]   2> 3459143 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.7.0
   [junit4]   2> 3459162 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test
   [junit4]   2> 3459268 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
   [junit4]   2> 3459289 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1' 
using configuration from collection control_collection, trusted=true
   [junit4]   2> 3459290 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.core.control_collection.shard1.replica_n1' (registry 
'solr.core.control_collection.shard1.replica_n1') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
   [junit4]   2> 3459290 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore 
at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores/control_collection_shard1_replica_n1],
 
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/control-001/cores/control_collection_shard1_replica_n1/data/]
   [junit4]   2> 3459295 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: 
maxMergeAtOnce=47, maxMergeAtOnceExplicit=22, maxMergedSegmentMB=1.4248046875, 
floorSegmentMB=0.9541015625, forceMergeDeletesPctAllowed=16.83237086906014, 
segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.4209624511310861, deletesPctAllowed=39.50145037974311
   [junit4]   2> 3459304 WARN  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = 
requestHandler,name = /dump,class = DumpRequestHandler,attributes = 
{initParams=a, name=/dump, class=DumpRequestHandler},args = 
{defaults={a=A,b=B}}}
   [junit4]   2> 3459383 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.UpdateHandler Using UpdateLog implementation: 
org.apache.solr.update.UpdateLog
   [junit4]   2> 3459383 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH 
numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 3459385 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.CommitTracker Hard AutoCommit: if uncommitted for 30000ms; 
   [junit4]   2> 3459385 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.CommitTracker Soft AutoCommit: if uncommitted for 3000ms; 
   [junit4]   2> 3459387 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: 
minMergeSize=1677721, mergeFactor=41, maxMergeSize=2147483648, 
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, 
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.0]
   [junit4]   2> 3459388 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.s.SolrIndexSearcher Opening 
[Searcher@1dd69426[control_collection_shard1_replica_n1] main]
   [junit4]   2> 3459389 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 3459390 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 3459391 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 3459391 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.UpdateLog Could not find max version in index or recent updates, using 
new clock 1618310683661697024
   [junit4]   2> 3459393 INFO  
(searcherExecutor-12526-thread-1-processing-n:127.0.0.1:36044_pq_%2Fz 
x:control_collection_shard1_replica_n1 c:control_collection s:shard1) 
[n:127.0.0.1:36044_pq_%2Fz c:control_collection s:shard1  
x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore 
[control_collection_shard1_replica_n1] Registered new searcher 
Searcher@1dd69426[control_collection_shard1_replica_n1] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 3459401 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ZkShardTerms Successful update of terms at 
/collections/control_collection/terms/shard1 to Terms{values={core_node2=0}, 
version=0}
   [junit4]   2> 3459406 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 3459406 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 3459406 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.SyncStrategy Sync replicas to 
http://127.0.0.1:36044/pq_/z/control_collection_shard1_replica_n1/
   [junit4]   2> 3459406 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 3459406 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.SyncStrategy 
http://127.0.0.1:36044/pq_/z/control_collection_shard1_replica_n1/ has no 
replicas
   [junit4]   2> 3459406 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in 
election, clear LIR
   [junit4]   2> 3459412 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ShardLeaderElectionContext I am the new leader: 
http://127.0.0.1:36044/pq_/z/control_collection_shard1_replica_n1/ shard1
   [junit4]   2> 3459515 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 3459523 INFO  (qtp127098813-40152) [n:127.0.0.1:36044_pq_%2Fz 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores 
params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT}
 status=0 QTime=1406
   [junit4]   2> 3459526 INFO  (qtp127098813-40156) [n:127.0.0.1:36044_pq_%2Fz  
  ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at 
most 30 seconds. Check all shard replicas
   [junit4]   2> 3459618 INFO  (zkCallback-10519-thread-1) [    ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/control_collection/state.json] for collection 
[control_collection] has occurred - updating... (live nodes size: [1])
   [junit4]   2> 3459902 DEBUG (ScheduledTrigger-12519-thread-4) [    ] 
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with 
currently live nodes: 1
   [junit4]   2> 3460008 INFO  
(OverseerCollectionConfigSetProcessor-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000)
 [n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.OverseerTaskQueue Response ZK path: 
/overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may 
have disconnected from ZooKeeper
   [junit4]   2> 3460527 INFO  (qtp127098813-40156) [n:127.0.0.1:36044_pq_%2Fz  
  ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections 
params={collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:36044_pq_%252Fz&wt=javabin&version=2}
 status=0 QTime=2524
   [junit4]   2> 3460532 INFO  (zkConnectionManagerCallback-10538-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3460534 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 3460535 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready
   [junit4]   2> 3460535 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection 
loss:false
   [junit4]   2> 3460536 INFO  (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz  
  ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params 
collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2
 and sendToOCPQueue=true
   [junit4]   2> 3460539 INFO  
(OverseerThreadFactory-12521-thread-2-processing-n:127.0.0.1:36044_pq_%2Fz) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.a.c.CreateCollectionCmd Create 
collection collection1
   [junit4]   2> 3460744 WARN  
(OverseerThreadFactory-12521-thread-2-processing-n:127.0.0.1:36044_pq_%2Fz) 
[n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.a.c.CreateCollectionCmd It is unusual 
to create a collection (collection1) without cores.
   [junit4]   2> 3460747 INFO  (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz  
  ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at 
most 30 seconds. Check all shard replicas
   [junit4]   2> 3460748 INFO  (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz  
  ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections 
params={collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2}
 status=0 QTime=211
   [junit4]   2> 3460902 DEBUG (ScheduledTrigger-12519-thread-3) [    ] 
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with 
currently live nodes: 1
   [junit4]   2> 3460927 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001
 of type NRT
   [junit4]   2> 3460931 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: 
d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12
   [junit4]   2> 3460933 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 3460933 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 3460934 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 3460934 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@4682a57{/pq_/z,null,AVAILABLE}
   [junit4]   2> 3460934 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@325d81d3{HTTP/1.1,[http/1.1]}{127.0.0.1:44392}
   [junit4]   2> 3460934 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.Server Started @3460991ms
   [junit4]   2> 3460934 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: 
{solr.data.dir=hdfs://localhost:34159/hdfs__localhost_34159__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J2_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001_tempDir-002_jetty1,
 replicaType=NRT, solrconfig=solrconfig.xml, hostContext=/pq_/z, 
hostPort=44392, 
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores}
   [junit4]   2> 3460935 ERROR 
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 3460937 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter Using logger factory 
org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 3460937 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 
7.7.0
   [junit4]   2> 3460937 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 3460937 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 3460937 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2018-11-27T17:55:36.520Z
   [junit4]   2> 3460939 INFO  (zkConnectionManagerCallback-10540-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3460940 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in 
ZooKeeper)
   [junit4]   2> 3460940 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig Loading container configuration from 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/solr.xml
   [junit4]   2> 3460945 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay 
is ignored
   [junit4]   2> 3460946 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig Configuration parameter 
autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 3460948 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 3461155 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44631/solr
   [junit4]   2> 3461157 INFO  (zkConnectionManagerCallback-10544-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3461159 INFO  (zkConnectionManagerCallback-10546-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3461168 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 3461173 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 
transient cores
   [junit4]   2> 3461174 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:44392_pq_%2Fz
   [junit4]   2> 3461175 INFO  (zkCallback-10526-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 3461182 INFO  (zkCallback-10519-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 3461182 INFO  (zkCallback-10537-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 3461182 INFO  (zkCallback-10545-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 3461204 INFO  (zkConnectionManagerCallback-10553-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3461205 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 3461206 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready
   [junit4]   2> 3461206 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history 
in memory.
   [junit4]   2> 3461237 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') 
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
   [junit4]   2> 3461255 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') 
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
   [junit4]   2> 3461255 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 
'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
   [junit4]   2> 3461258 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores
   [junit4]   2> 3461299 INFO  (qtp127098813-40151) [n:127.0.0.1:36044_pq_%2Fz  
  ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with 
params 
node=127.0.0.1:44392_pq_%252Fz&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2
 and sendToOCPQueue=true
   [junit4]   2> 3461303 INFO  
(OverseerCollectionConfigSetProcessor-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000)
 [n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.OverseerTaskQueue Response ZK path: 
/overseer/collection-queue-work/qnr-0000000002 doesn't exist.  Requestor may 
have disconnected from ZooKeeper
   [junit4]   2> 3461313 INFO  (qtp127098813-40155) [n:127.0.0.1:36044_pq_%2Fz  
  ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics 
params={wt=javabin&version=2&key=solr.core.control_collection.shard1.replica_n1:INDEX.sizeInBytes}
 status=0 QTime=2
   [junit4]   2> 3461315 INFO  (qtp127098813-40157) [n:127.0.0.1:36044_pq_%2Fz  
  ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics 
params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core}
 status=0 QTime=1
   [junit4]   2> 3461318 INFO  (qtp641447673-40209) [n:127.0.0.1:44392_pq_%2Fz  
  ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics 
params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core}
 status=0 QTime=0
   [junit4]   2> 3461319 INFO  
(OverseerThreadFactory-12521-thread-3-processing-n:127.0.0.1:36044_pq_%2Fz) 
[n:127.0.0.1:36044_pq_%2Fz c:collection1 s:shard1  ] o.a.s.c.a.c.AddReplicaCmd 
Node Identified 127.0.0.1:44392_pq_%2Fz for creating new replica of shard 
shard1 for collection collection1
   [junit4]   2> 3461324 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz  
  x:collection1_shard1_replica_n21] o.a.s.h.a.CoreAdminOperation core create 
command 
qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 3461903 DEBUG (ScheduledTrigger-12519-thread-4) [    ] 
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with 
currently live nodes: 2
   [junit4]   2> 3462345 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.SolrConfig 
Using Lucene MatchVersion: 7.7.0
   [junit4]   2> 3462361 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.s.IndexSchema 
[collection1_shard1_replica_n21] Schema name=test
   [junit4]   2> 3462456 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.s.IndexSchema 
Loaded schema test/1.6 with uniqueid field id
   [junit4]   2> 3462474 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.CoreContainer 
Creating SolrCore 'collection1_shard1_replica_n21' using configuration from 
collection collection1, trusted=true
   [junit4]   2> 3462475 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.core.collection1.shard1.replica_n21' (registry 
'solr.core.collection1.shard1.replica_n21') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
   [junit4]   2> 3462475 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.SolrCore 
[[collection1_shard1_replica_n21] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores/collection1_shard1_replica_n21],
 
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-1-001/cores/collection1_shard1_replica_n21/data/]
   [junit4]   2> 3462480 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: 
maxMergeAtOnce=47, maxMergeAtOnceExplicit=22, maxMergedSegmentMB=1.4248046875, 
floorSegmentMB=0.9541015625, forceMergeDeletesPctAllowed=16.83237086906014, 
segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.4209624511310861, deletesPctAllowed=39.50145037974311
   [junit4]   2> 3462497 WARN  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = 
requestHandler,name = /dump,class = DumpRequestHandler,attributes = 
{initParams=a, name=/dump, class=DumpRequestHandler},args = 
{defaults={a=A,b=B}}}
   [junit4]   2> 3462568 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.u.UpdateHandler 
Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 3462569 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.u.UpdateLog 
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 3462570 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.u.CommitTracker 
Hard AutoCommit: if uncommitted for 30000ms; 
   [junit4]   2> 3462570 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.u.CommitTracker 
Soft AutoCommit: if uncommitted for 3000ms; 
   [junit4]   2> 3462573 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: 
minMergeSize=1677721, mergeFactor=41, maxMergeSize=2147483648, 
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, 
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.0]
   [junit4]   2> 3462573 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.s.SolrIndexSearcher Opening 
[Searcher@5fc5a521[collection1_shard1_replica_n21] main]
   [junit4]   2> 3462575 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 3462575 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 3462576 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 3462577 INFO  
(searcherExecutor-12540-thread-1-processing-n:127.0.0.1:44392_pq_%2Fz 
x:collection1_shard1_replica_n21 c:collection1 s:shard1) 
[n:127.0.0.1:44392_pq_%2Fz c:collection1 s:shard1  
x:collection1_shard1_replica_n21] o.a.s.c.SolrCore 
[collection1_shard1_replica_n21] Registered new searcher 
Searcher@5fc5a521[collection1_shard1_replica_n21] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 3462577 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.u.UpdateLog 
Could not find max version in index or recent updates, using new clock 
1618310687002460160
   [junit4]   2> 3462584 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.ZkShardTerms 
Successful update of terms at /collections/collection1/terms/shard1 to 
Terms{values={core_node22=0}, version=0}
   [junit4]   2> 3462588 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 3462588 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 3462588 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy 
Sync replicas to http://127.0.0.1:44392/pq_/z/collection1_shard1_replica_n21/
   [junit4]   2> 3462588 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy 
Sync Success - now sync replicas to me
   [junit4]   2> 3462589 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy 
http://127.0.0.1:44392/pq_/z/collection1_shard1_replica_n21/ has no replicas
   [junit4]   2> 3462589 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in 
election, clear LIR
   [junit4]   2> 3462593 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.c.ShardLeaderElectionContext I am the new leader: 
http://127.0.0.1:44392/pq_/z/collection1_shard1_replica_n21/ shard1
   [junit4]   2> 3462696 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.ZkController 
I am the leader, no recovery necessary
   [junit4]   2> 3462699 INFO  (qtp641447673-40204) [n:127.0.0.1:44392_pq_%2Fz 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.s.HttpSolrCall 
[admin] webapp=null path=/admin/cores 
params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT}
 status=0 QTime=1375
   [junit4]   2> 3462702 INFO  (qtp127098813-40151) [n:127.0.0.1:36044_pq_%2Fz 
c:collection1   ] o.a.s.s.HttpSolrCall [admin] webapp=null 
path=/admin/collections 
params={node=127.0.0.1:44392_pq_%252Fz&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2}
 status=0 QTime=1402
   [junit4]   2> 3462799 INFO  (zkCallback-10545-thread-1) [    ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/collection1/state.json] for collection [collection1] has 
occurred - updating... (live nodes size: [2])
   [junit4]   2> 3462882 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-2-001
 of type NRT
   [junit4]   2> 3462883 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: 
d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12
   [junit4]   2> 3462889 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 3462889 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 3462889 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 3462889 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@1aaa752d{/pq_/z,null,AVAILABLE}
   [junit4]   2> 3462889 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@70cb9b0f{HTTP/1.1,[http/1.1]}{127.0.0.1:35419}
   [junit4]   2> 3462889 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.e.j.s.Server Started @3462946ms
   [junit4]   2> 3462889 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: 
{solr.data.dir=hdfs://localhost:34159/hdfs__localhost_34159__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J2_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001_tempDir-002_jetty2,
 solrconfig=solrconfig.xml, hostContext=/pq_/z, hostPort=35419, 
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-2-001/cores}
   [junit4]   2> 3462890 ERROR 
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 3462891 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter Using logger factory 
org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 3462891 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 
7.7.0
   [junit4]   2> 3462891 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 3462891 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 3462891 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2018-11-27T17:55:38.474Z
   [junit4]   2> 3462893 INFO  (zkConnectionManagerCallback-10556-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3462895 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in 
ZooKeeper)
   [junit4]   2> 3462895 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig Loading container configuration from 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J2/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_64E46779FF6ECF5F-001/shard-2-001/solr.xml
   [junit4]   2> 3462899 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay 
is ignored
   [junit4]   2> 3462899 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig Configuration parameter 
autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 3462901 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 3462904 DEBUG (ScheduledTrigger-12519-thread-3) [    ] 
o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with 
currently live nodes: 2
   [junit4]   2> 3463304 INFO  
(OverseerCollectionConfigSetProcessor-72350940727279620-127.0.0.1:36044_pq_%2Fz-n_0000000000)
 [n:127.0.0.1:36044_pq_%2Fz    ] o.a.s.c.OverseerTaskQueue Response ZK path: 
/overseer/collection-queue-work/qnr-0000000004 doesn't exist.  Requestor may 
have disconnected from ZooKeeper
   [junit4]   2> 3463361 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44631/solr
   [junit4]   2> 3463363 INFO  (zkConnectionManagerCallback-10560-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3463365 INFO  (zkConnectionManagerCallback-10562-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3463374 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 3463379 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 
transient cores
   [junit4]   2> 3463379 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:35419_pq_%2Fz
   [junit4]   2> 3463381 INFO  (zkCallback-10526-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 3463381 INFO  (zkCallback-10537-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 3463381 INFO  (zkCallback-10552-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 3463384 INFO  (zkCallback-10545-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 3463384 INFO  (zkCallback-10519-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 3463388 INFO  (zkCallback-10561-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 3463409 INFO  (zkConnectionManagerCallback-10569-thread-1) [   
 ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3463411 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
   [junit4]   2> 3463412 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44631/solr ready
   [junit4]   2> 3463412 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history 
in memory.
   [junit4]   2> 3463445 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') 
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
   [junit4]   2> 3463462 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') 
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
   [junit4]   2> 3463462 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[64E46779FF6ECF5F]) [    ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 
'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a9ddff3
   [junit4]   2> 3463465 INFO  (TEST-HdfsRestartWhileUpdatingTest.

[...truncated too long message...]

kout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/top-level-ivy-settings.xml

resolve:

jar-checksums:
    [mkdir] Created dir: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/null1170095894
     [copy] Copying 241 files to 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/null1170095894
   [delete] Deleting directory 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/null1170095894

check-working-copy:
[ivy:cachepath] :: resolving dependencies :: 
org.eclipse.jgit#org.eclipse.jgit-caller;working
[ivy:cachepath]         confs: [default]
[ivy:cachepath]         found 
org.eclipse.jgit#org.eclipse.jgit;4.6.0.201612231935-r in public
[ivy:cachepath]         found com.jcraft#jsch;0.1.53 in public
[ivy:cachepath]         found com.googlecode.javaewah#JavaEWAH;1.1.6 in public
[ivy:cachepath]         found org.apache.httpcomponents#httpclient;4.3.6 in 
public
[ivy:cachepath]         found org.apache.httpcomponents#httpcore;4.3.3 in public
[ivy:cachepath]         found commons-logging#commons-logging;1.1.3 in public
[ivy:cachepath]         found commons-codec#commons-codec;1.6 in public
[ivy:cachepath]         found org.slf4j#slf4j-api;1.7.2 in public
[ivy:cachepath] :: resolution report :: resolve 29ms :: artifacts dl 2ms
        ---------------------------------------------------------------------
        |                  |            modules            ||   artifacts   |
        |       conf       | number| search|dwnlded|evicted|| number|dwnlded|
        ---------------------------------------------------------------------
        |      default     |   8   |   0   |   0   |   0   ||   8   |   0   |
        ---------------------------------------------------------------------
[wc-checker] Initializing working copy...
[wc-checker] SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
[wc-checker] SLF4J: Defaulting to no-operation (NOP) logger implementation
[wc-checker] SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for 
further details.
[wc-checker] Checking working copy status...

-jenkins-base:

BUILD SUCCESSFUL
Total time: 285 minutes 41 seconds
Archiving artifacts
java.lang.InterruptedException: no matches found within 10000
        at hudson.FilePath$34.hasMatch(FilePath.java:2678)
        at hudson.FilePath$34.invoke(FilePath.java:2557)
        at hudson.FilePath$34.invoke(FilePath.java:2547)
        at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2918)
Also:   hudson.remoting.Channel$CallSiteStackTrace: Remote call to lucene
                at 
hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1741)
                at 
hudson.remoting.UserRequest$ExceptionResponse.retrieve(UserRequest.java:357)
                at hudson.remoting.Channel.call(Channel.java:955)
                at hudson.FilePath.act(FilePath.java:1036)
                at hudson.FilePath.act(FilePath.java:1025)
                at hudson.FilePath.validateAntFileMask(FilePath.java:2547)
                at 
hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
                at 
hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
                at 
hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
                at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
                at 
hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
                at hudson.model.Build$BuildExecution.post2(Build.java:186)
                at 
hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
                at hudson.model.Run.execute(Run.java:1819)
                at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
                at 
hudson.model.ResourceController.execute(ResourceController.java:97)
                at hudson.model.Executor.run(Executor.java:429)
Caused: hudson.FilePath$TunneledInterruptedException
        at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2920)
        at hudson.remoting.UserRequest.perform(UserRequest.java:212)
        at hudson.remoting.UserRequest.perform(UserRequest.java:54)
        at hudson.remoting.Request$2.run(Request.java:369)
        at 
hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:748)
Caused: java.lang.InterruptedException: java.lang.InterruptedException: no 
matches found within 10000
        at hudson.FilePath.act(FilePath.java:1038)
        at hudson.FilePath.act(FilePath.java:1025)
        at hudson.FilePath.validateAntFileMask(FilePath.java:2547)
        at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
        at 
hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
        at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
        at hudson.model.Build$BuildExecution.post2(Build.java:186)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
        at hudson.model.Run.execute(Run.java:1819)
        at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
        at hudson.model.ResourceController.execute(ResourceController.java:97)
        at hudson.model.Executor.run(Executor.java:429)
No artifacts found that match the file pattern 
"**/*.events,heapdumps/**,**/hs_err_pid*". Configuration error?
Recording test results
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Email was triggered for: Unstable (Test Failures)
Sending email for trigger: Unstable (Test Failures)
---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

Reply via email to