Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-7.x/376/

4 tests failed.
FAILED:  org.apache.lucene.document.TestLatLonLineShapeQueries.testRandomBig

Error Message:
Java heap space

Stack Trace:
java.lang.OutOfMemoryError: Java heap space
        at 
__randomizedtesting.SeedInfo.seed([5186EC3284991EDD:D6D191BD15C0625D]:0)
        at org.apache.lucene.store.RAMFile.newBuffer(RAMFile.java:84)
        at org.apache.lucene.store.RAMFile.addBuffer(RAMFile.java:57)
        at 
org.apache.lucene.store.RAMOutputStream.switchCurrentBuffer(RAMOutputStream.java:168)
        at 
org.apache.lucene.store.RAMOutputStream.writeBytes(RAMOutputStream.java:154)
        at 
org.apache.lucene.store.MockIndexOutputWrapper.writeBytes(MockIndexOutputWrapper.java:137)
        at 
org.apache.lucene.util.bkd.OfflinePointWriter.append(OfflinePointWriter.java:75)
        at org.apache.lucene.util.bkd.BKDWriter.add(BKDWriter.java:287)
        at org.apache.lucene.index.RandomCodec$1$1$1.visit(RandomCodec.java:131)
        at 
org.apache.lucene.codecs.PointsWriter$1$1$1.visit(PointsWriter.java:117)
        at 
org.apache.lucene.index.AssertingLeafReader$AssertingIntersectVisitor.visit(AssertingLeafReader.java:1009)
        at 
org.apache.lucene.util.bkd.BKDReader.visitCompressedDocValues(BKDReader.java:709)
        at 
org.apache.lucene.util.bkd.BKDReader.visitDocValues(BKDReader.java:672)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:764)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
        at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:533)
        at 
org.apache.lucene.index.AssertingLeafReader$AssertingPointValues.intersect(AssertingLeafReader.java:898)
        at 
org.apache.lucene.codecs.PointsWriter$1$1.intersect(PointsWriter.java:105)
        at 
org.apache.lucene.index.RandomCodec$1$1.writeField(RandomCodec.java:124)
        at 
org.apache.lucene.codecs.PointsWriter.mergeOneField(PointsWriter.java:62)
        at org.apache.lucene.codecs.PointsWriter.merge(PointsWriter.java:191)


FAILED:  org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest.test

Error Message:
There are still nodes recoverying - waited for 320 seconds

Stack Trace:
java.lang.AssertionError: There are still nodes recoverying - waited for 320 
seconds
        at 
__randomizedtesting.SeedInfo.seed([EFC03E3D27357B3D:679401E789C916C5]:0)
        at org.junit.Assert.fail(Assert.java:93)
        at 
org.apache.solr.cloud.AbstractDistribZkTestBase.waitForRecoveriesToFinish(AbstractDistribZkTestBase.java:185)
        at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForRecoveriesToFinish(AbstractFullDistribZkTestBase.java:920)
        at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForThingsToLevelOut(AbstractFullDistribZkTestBase.java:1477)
        at 
org.apache.solr.cloud.RestartWhileUpdatingTest.test(RestartWhileUpdatingTest.java:145)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1742)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:935)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:971)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:985)
        at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:1010)
        at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:985)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:944)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:830)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:880)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:891)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at java.lang.Thread.run(Thread.java:748)


FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest

Error Message:
7 threads leaked from SUITE scope at 
org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest:     1) 
Thread[id=43682, name=searcherExecutor-2001-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    2) Thread[id=43559, 
name=searcherExecutor-1959-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    3) Thread[id=43879, 
name=searcherExecutor-2021-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    4) Thread[id=43914, 
name=searcherExecutor-2035-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    5) Thread[id=43949, 
name=searcherExecutor-2049-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    6) Thread[id=43600, 
name=searcherExecutor-1973-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    7) Thread[id=43637, 
name=searcherExecutor-1987-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: 7 threads leaked from SUITE 
scope at org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest: 
   1) Thread[id=43682, name=searcherExecutor-2001-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   2) Thread[id=43559, name=searcherExecutor-1959-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   3) Thread[id=43879, name=searcherExecutor-2021-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   4) Thread[id=43914, name=searcherExecutor-2035-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   5) Thread[id=43949, name=searcherExecutor-2049-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   6) Thread[id=43600, name=searcherExecutor-1973-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   7) Thread[id=43637, name=searcherExecutor-1987-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
        at __randomizedtesting.SeedInfo.seed([EFC03E3D27357B3D]:0)


FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest

Error Message:
There are still zombie threads that couldn't be terminated:    1) 
Thread[id=43682, name=searcherExecutor-2001-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    2) Thread[id=43559, 
name=searcherExecutor-1959-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    3) Thread[id=43879, 
name=searcherExecutor-2021-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    4) Thread[id=43914, 
name=searcherExecutor-2035-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    5) Thread[id=43949, 
name=searcherExecutor-2049-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    6) Thread[id=43600, 
name=searcherExecutor-1973-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)    7) Thread[id=43637, 
name=searcherExecutor-1987-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]         at sun.misc.Unsafe.park(Native 
Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)   
      at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) 
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
        at java.lang.Thread.run(Thread.java:748)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie 
threads that couldn't be terminated:
   1) Thread[id=43682, name=searcherExecutor-2001-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   2) Thread[id=43559, name=searcherExecutor-1959-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   3) Thread[id=43879, name=searcherExecutor-2021-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   4) Thread[id=43914, name=searcherExecutor-2035-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   5) Thread[id=43949, name=searcherExecutor-2049-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   6) Thread[id=43600, name=searcherExecutor-1973-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   7) Thread[id=43637, name=searcherExecutor-1987-thread-1, state=WAITING, 
group=TGRP-HdfsRestartWhileUpdatingTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
        at __randomizedtesting.SeedInfo.seed([EFC03E3D27357B3D]:0)




Build Log:
[...truncated 10294 lines...]
   [junit4] Suite: org.apache.lucene.document.TestLatLonLineShapeQueries
   [junit4]   2> NOTE: download the large Jenkins line-docs file by running 
'ant get-jenkins-line-docs' in the lucene directory.
   [junit4]   2> NOTE: reproduce with: ant test  
-Dtestcase=TestLatLonLineShapeQueries -Dtests.method=testRandomBig 
-Dtests.seed=5186EC3284991EDD -Dtests.multiplier=2 -Dtests.nightly=true 
-Dtests.slow=true 
-Dtests.linedocsfile=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/test-data/enwiki.random.lines.txt
 -Dtests.locale=de-CH -Dtests.timezone=America/Coral_Harbour 
-Dtests.asserts=true -Dtests.file.encoding=ISO-8859-1
   [junit4] ERROR    749s J2 | TestLatLonLineShapeQueries.testRandomBig <<<
   [junit4]    > Throwable #1: java.lang.OutOfMemoryError: Java heap space
   [junit4]    >        at 
__randomizedtesting.SeedInfo.seed([5186EC3284991EDD:D6D191BD15C0625D]:0)
   [junit4]    >        at 
org.apache.lucene.store.RAMFile.newBuffer(RAMFile.java:84)
   [junit4]    >        at 
org.apache.lucene.store.RAMFile.addBuffer(RAMFile.java:57)
   [junit4]    >        at 
org.apache.lucene.store.RAMOutputStream.switchCurrentBuffer(RAMOutputStream.java:168)
   [junit4]    >        at 
org.apache.lucene.store.RAMOutputStream.writeBytes(RAMOutputStream.java:154)
   [junit4]    >        at 
org.apache.lucene.store.MockIndexOutputWrapper.writeBytes(MockIndexOutputWrapper.java:137)
   [junit4]    >        at 
org.apache.lucene.util.bkd.OfflinePointWriter.append(OfflinePointWriter.java:75)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDWriter.add(BKDWriter.java:287)
   [junit4]    >        at 
org.apache.lucene.index.RandomCodec$1$1$1.visit(RandomCodec.java:131)
   [junit4]    >        at 
org.apache.lucene.codecs.PointsWriter$1$1$1.visit(PointsWriter.java:117)
   [junit4]    >        at 
org.apache.lucene.index.AssertingLeafReader$AssertingIntersectVisitor.visit(AssertingLeafReader.java:1009)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.visitCompressedDocValues(BKDReader.java:709)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.visitDocValues(BKDReader.java:672)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:764)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797)
   [junit4]    >        at 
org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:533)
   [junit4]    >        at 
org.apache.lucene.index.AssertingLeafReader$AssertingPointValues.intersect(AssertingLeafReader.java:898)
   [junit4]    >        at 
org.apache.lucene.codecs.PointsWriter$1$1.intersect(PointsWriter.java:105)
   [junit4]    >        at 
org.apache.lucene.index.RandomCodec$1$1.writeField(RandomCodec.java:124)
   [junit4]    >        at 
org.apache.lucene.codecs.PointsWriter.mergeOneField(PointsWriter.java:62)
   [junit4]    >        at 
org.apache.lucene.codecs.PointsWriter.merge(PointsWriter.java:191)
   [junit4]   2> NOTE: leaving temporary files on disk at: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/build/sandbox/test/J2/temp/lucene.document.TestLatLonLineShapeQueries_5186EC3284991EDD-001
   [junit4]   2> NOTE: test params are: codec=Asserting(Lucene70): {id=FST50}, 
docValues:{shape=DocValuesFormat(name=Asserting), 
id=DocValuesFormat(name=Lucene70)}, maxPointsInLeafNode=999, 
maxMBSortInHeap=6.293844669044266, sim=RandomSimilarity(queryNorm=true): {}, 
locale=de-CH, timezone=America/Coral_Harbour
   [junit4]   2> NOTE: Linux 4.4.0-137-generic amd64/Oracle Corporation 
1.8.0_191 (64-bit)/cpus=4,threads=1,free=285326408,total=440926208
   [junit4]   2> NOTE: All tests run in this JVM: 
[TestLatLonBoundingBoxQueries, TestLatLonShape, TestDoubleRangeField, 
TestLatLonLineShapeQueries]
   [junit4] Completed [19/22 (1!)] on J2 in 1096.68s, 4 tests, 1 error <<< 
FAILURES!

[...truncated 1 lines...]
   [junit4] JVM J2: stdout was not empty, see: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/build/sandbox/test/temp/junit4-J2-20181113_135933_6228772371157596158491.sysout
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] java.lang.OutOfMemoryError: Java heap space
   [junit4] Dumping heap to 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/heapdumps/java_pid15722.hprof
 ...
   [junit4] Heap dump file created [462447318 bytes in 0.805 secs]
   [junit4] <<< JVM J2: EOF ----

[...truncated 3234 lines...]
   [junit4] Suite: org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest
   [junit4]   2> Creating dataDir: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/init-core-data-001
   [junit4]   2> 884469 WARN  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=279 numCloses=279
   [junit4]   2> 884470 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) 
w/NUMERIC_DOCVALUES_SYSPROP=false
   [junit4]   2> 884472 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (true) via: 
@org.apache.solr.util.RandomizeSSL(reason=, ssl=NaN, value=NaN, clientAuth=NaN)
   [junit4]   2> 884473 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: 
test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> 884473 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /_ucp/t
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 886381 WARN  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried 
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 886708 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.m.log Logging to org.apache.logging.slf4j.Log4jLogger@385a6de4 via 
org.mortbay.log.Slf4jLog
   [junit4]   2> 886737 WARN  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 886840 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.m.log jetty-6.1.26
   [junit4]   2> 886913 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs
 to ./temp/Jetty_localhost_46064_hdfs____.ao98ve/webapp
   [junit4]   2> 887807 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:46064
   [junit4]   2> 889145 WARN  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 889156 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.m.log jetty-6.1.26
   [junit4]   2> 889177 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode
 to ./temp/Jetty_localhost_46434_datanode____hprwhm/webapp
   [junit4]   2> 889627 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:46434
   [junit4]   2> 890396 WARN  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 890398 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.m.log jetty-6.1.26
   [junit4]   2> 890465 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode
 to ./temp/Jetty_localhost_41325_datanode____.wise4g/webapp
   [junit4]   2> 891344 INFO  
(SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [    ] 
o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:41325
   [junit4]   2> 891885 ERROR (DataNode: 
[[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/tempDir-001/hdfsBaseDir/data/data1/,
 
[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/tempDir-001/hdfsBaseDir/data/data2/]]
  heartbeating to localhost/127.0.0.1:37822) [    ] 
o.a.h.h.s.d.DirectoryScanner 
dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 
ms/sec. Assuming default value of 1000
   [junit4]   2> 891887 ERROR (DataNode: 
[[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/tempDir-001/hdfsBaseDir/data/data3/,
 
[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/tempDir-001/hdfsBaseDir/data/data4/]]
  heartbeating to localhost/127.0.0.1:37822) [    ] 
o.a.h.h.s.d.DirectoryScanner 
dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 
ms/sec. Assuming default value of 1000
   [junit4]   2> 892135 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0xb91f918a9cb56: from storage 
DS-320ccfa3-68d6-4847-abf1-9fd841aea974 node 
DatanodeRegistration(127.0.0.1:34595, 
datanodeUuid=d43953ae-b8dc-464d-8815-026a85d6d206, infoPort=38832, 
infoSecurePort=0, ipcPort=41708, 
storageInfo=lv=-56;cid=testClusterID;nsid=1189364975;c=0), blocks: 0, 
hasStaleStorage: true, processing time: 6 msecs
   [junit4]   2> 892136 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0xb91f91845f1b7: from storage 
DS-a8c82a40-70f3-4f1a-8323-4186be7d4843 node 
DatanodeRegistration(127.0.0.1:39605, 
datanodeUuid=49051a15-2b0b-488c-95aa-dffb4e893cb4, infoPort=42551, 
infoSecurePort=0, ipcPort=43971, 
storageInfo=lv=-56;cid=testClusterID;nsid=1189364975;c=0), blocks: 0, 
hasStaleStorage: true, processing time: 1 msecs
   [junit4]   2> 892136 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0xb91f918a9cb56: from storage 
DS-cdecb5ea-8e3a-49b8-bd0f-7ce1646c22c3 node 
DatanodeRegistration(127.0.0.1:34595, 
datanodeUuid=d43953ae-b8dc-464d-8815-026a85d6d206, infoPort=38832, 
infoSecurePort=0, ipcPort=41708, 
storageInfo=lv=-56;cid=testClusterID;nsid=1189364975;c=0), blocks: 0, 
hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 892141 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0xb91f91845f1b7: from storage 
DS-760520a3-b796-4e7e-953e-5da6d48b7a15 node 
DatanodeRegistration(127.0.0.1:39605, 
datanodeUuid=49051a15-2b0b-488c-95aa-dffb4e893cb4, infoPort=42551, 
infoSecurePort=0, ipcPort=43971, 
storageInfo=lv=-56;cid=testClusterID;nsid=1189364975;c=0), blocks: 0, 
hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 892982 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 892983 INFO  (Thread-37972) [    ] o.a.s.c.ZkTestServer client 
port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 892983 INFO  (Thread-37972) [    ] o.a.s.c.ZkTestServer 
Starting server
   [junit4]   2> 892987 ERROR (Thread-37972) [    ] o.a.z.s.ZooKeeperServer 
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action 
on ERROR or SHUTDOWN server state changes
   [junit4]   2> 893083 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.ZkTestServer start zk server on port:41000
   [junit4]   2> 893108 INFO  (zkConnectionManagerCallback-3229-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 893112 WARN  (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [    ] 
o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 
0x100c21dacb70000, likely client has closed socket
   [junit4]   2> 893123 INFO  (zkConnectionManagerCallback-3231-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 893130 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
 to /configs/conf1/solrconfig.xml
   [junit4]   2> 893133 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/schema15.xml
 to /configs/conf1/schema.xml
   [junit4]   2> 893134 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
 to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 893136 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
 to /configs/conf1/stopwords.txt
   [junit4]   2> 893138 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/protwords.txt
 to /configs/conf1/protwords.txt
   [junit4]   2> 893139 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/currency.xml
 to /configs/conf1/currency.xml
   [junit4]   2> 893141 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml
 to /configs/conf1/enumsConfig.xml
   [junit4]   2> 893142 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json
 to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 893151 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt
 to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 893156 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt
 to /configs/conf1/old_synonyms.txt
   [junit4]   2> 893157 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/synonyms.txt
 to /configs/conf1/synonyms.txt
   [junit4]   2> 893161 INFO  (zkConnectionManagerCallback-3235-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 893164 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractFullDistribZkTestBase Will use NRT replicas unless explicitly 
asked otherwise
   [junit4]   2> 893427 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: 
d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12
   [junit4]   2> 893489 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 893489 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 893489 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 893492 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@c6ebc75{/_ucp/t,null,AVAILABLE}
   [junit4]   2> 893500 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.AbstractConnector Started ServerConnector@3e77f6ea{SSL,[ssl, 
http/1.1]}{127.0.0.1:35905}
   [junit4]   2> 893500 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.Server Started @893601ms
   [junit4]   2> 893500 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: 
{solr.data.dir=hdfs://localhost:37822/hdfs__localhost_37822__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J1_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001_tempDir-002_control_data,
 replicaType=NRT, hostContext=/_ucp/t, hostPort=35905, 
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/control-001/cores}
   [junit4]   2> 893501 ERROR 
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 893501 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter Using logger factory 
org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 893501 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 
7.7.0
   [junit4]   2> 893501 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 893501 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 893501 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2018-11-13T15:12:20.733Z
   [junit4]   2> 893530 INFO  (zkConnectionManagerCallback-3237-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 893532 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in 
ZooKeeper)
   [junit4]   2> 893532 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.SolrXmlConfig Loading container configuration from 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/control-001/solr.xml
   [junit4]   2> 893538 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay 
is ignored
   [junit4]   2> 893538 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.SolrXmlConfig Configuration parameter 
autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 893540 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 893764 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41000/solr
   [junit4]   2> 893766 INFO  (zkConnectionManagerCallback-3241-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 893769 INFO  (zkConnectionManagerCallback-3243-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 893904 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.c.OverseerElectionContext I am going to 
be the leader 127.0.0.1:35905__ucp%2Ft
   [junit4]   2> 893906 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.c.Overseer Overseer 
(id=72271026745442309-127.0.0.1:35905__ucp%2Ft-n_0000000000) starting
   [junit4]   2> 893931 INFO  (zkConnectionManagerCallback-3250-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 893934 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.c.s.i.ZkClientClusterStateProvider 
Cluster at 127.0.0.1:41000/solr ready
   [junit4]   2> 893938 INFO  
(OverseerStateUpdate-72271026745442309-127.0.0.1:35905__ucp%2Ft-n_0000000000) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.c.Overseer Starting to work on the main 
queue : 127.0.0.1:35905__ucp%2Ft
   [junit4]   2> 893942 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:35905__ucp%2Ft
   [junit4]   2> 893968 INFO  (zkCallback-3242-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 893969 INFO  (zkCallback-3249-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 893991 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.h.a.MetricsHistoryHandler No .system 
collection, keeping metrics history in memory.
   [junit4]   2> 894020 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.node' (registry 'solr.node') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3
   [junit4]   2> 894034 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jvm' (registry 'solr.jvm') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3
   [junit4]   2> 894034 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jetty' (registry 'solr.jetty') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3
   [junit4]   2> 894036 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.c.CorePropertiesLocator Found 0 core 
definitions underneath 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/control-001/cores
   [junit4]   2> 894066 INFO  (zkConnectionManagerCallback-3256-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 894067 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 894069 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41000/solr ready
   [junit4]   2> 894110 INFO  (qtp1040856369-43520) [n:127.0.0.1:35905__ucp%2Ft 
   ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params 
collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:35905__ucp%252Ft&wt=javabin&version=2
 and sendToOCPQueue=true
   [junit4]   2> 894120 INFO  
(OverseerThreadFactory-1954-thread-1-processing-n:127.0.0.1:35905__ucp%2Ft) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.c.a.c.CreateCollectionCmd Create 
collection control_collection
   [junit4]   2> 894237 INFO  (qtp1040856369-43514) [n:127.0.0.1:35905__ucp%2Ft 
   ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics 
params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core}
 status=0 QTime=0
   [junit4]   2> 894249 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
   x:control_collection_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core 
create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 894249 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
   x:control_collection_shard1_replica_n1] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 
transient cores
   [junit4]   2> 895279 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.7.0
   [junit4]   2> 895296 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test
   [junit4]   2> 895423 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
   [junit4]   2> 895463 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1' 
using configuration from collection control_collection, trusted=true
   [junit4]   2> 895463 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.core.control_collection.shard1.replica_n1' (registry 
'solr.core.control_collection.shard1.replica_n1') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3
   [junit4]   2> 895464 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore 
at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/control-001/cores/control_collection_shard1_replica_n1],
 
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/control-001/cores/control_collection_shard1_replica_n1/data/]
   [junit4]   2> 895469 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: 
maxMergeAtOnce=33, maxMergeAtOnceExplicit=49, maxMergedSegmentMB=3.9248046875, 
floorSegmentMB=0.2802734375, forceMergeDeletesPctAllowed=26.06478682172601, 
segmentsPerTier=43.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, 
deletesPctAllowed=36.5804592175983
   [junit4]   2> 895481 WARN  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = 
requestHandler,name = /dump,class = DumpRequestHandler,attributes = 
{initParams=a, name=/dump, class=DumpRequestHandler},args = 
{defaults={a=A,b=B}}}
   [junit4]   2> 895555 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.UpdateHandler Using UpdateLog implementation: 
org.apache.solr.update.UpdateLog
   [junit4]   2> 895555 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH 
numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 895556 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.CommitTracker Hard AutoCommit: if uncommitted for 30000ms; 
   [junit4]   2> 895556 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.CommitTracker Soft AutoCommit: if uncommitted for 3000ms; 
   [junit4]   2> 895560 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: 
maxMergeAtOnce=14, maxMergeAtOnceExplicit=11, maxMergedSegmentMB=18.544921875, 
floorSegmentMB=0.8408203125, forceMergeDeletesPctAllowed=10.035465587719308, 
segmentsPerTier=14.0, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.14629428284484744, deletesPctAllowed=40.71873956406105
   [junit4]   2> 895561 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.s.SolrIndexSearcher Opening 
[Searcher@4f901b73[control_collection_shard1_replica_n1] main]
   [junit4]   2> 895562 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 895563 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 895564 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 895564 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.u.UpdateLog Could not find max version in index or recent updates, using 
new clock 1617032058289258496
   [junit4]   2> 895568 INFO  
(searcherExecutor-1959-thread-1-processing-n:127.0.0.1:35905__ucp%2Ft 
x:control_collection_shard1_replica_n1 c:control_collection s:shard1) 
[n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1  
x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore 
[control_collection_shard1_replica_n1] Registered new searcher 
Searcher@4f901b73[control_collection_shard1_replica_n1] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 895573 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ZkShardTerms Successful update of terms at 
/collections/control_collection/terms/shard1 to Terms{values={core_node2=0}, 
version=0}
   [junit4]   2> 895579 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 895579 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 895579 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.SyncStrategy Sync replicas to 
https://127.0.0.1:35905/_ucp/t/control_collection_shard1_replica_n1/
   [junit4]   2> 895579 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 895580 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.SyncStrategy 
https://127.0.0.1:35905/_ucp/t/control_collection_shard1_replica_n1/ has no 
replicas
   [junit4]   2> 895580 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in 
election, clear LIR
   [junit4]   2> 895586 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ShardLeaderElectionContext I am the new leader: 
https://127.0.0.1:35905/_ucp/t/control_collection_shard1_replica_n1/ shard1
   [junit4]   2> 895741 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 895745 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
c:control_collection s:shard1  x:control_collection_shard1_replica_n1] 
o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores 
params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT}
 status=0 QTime=1497
   [junit4]   2> 895762 INFO  (qtp1040856369-43520) [n:127.0.0.1:35905__ucp%2Ft 
   ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at 
most 30 seconds. Check all shard replicas
   [junit4]   2> 895856 INFO  (zkCallback-3242-thread-1) [    ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/control_collection/state.json] for collection 
[control_collection] has occurred - updating... (live nodes size: [1])
   [junit4]   2> 896126 INFO  
(OverseerCollectionConfigSetProcessor-72271026745442309-127.0.0.1:35905__ucp%2Ft-n_0000000000)
 [n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.c.OverseerTaskQueue Response ZK path: 
/overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may 
have disconnected from ZooKeeper
   [junit4]   2> 896762 INFO  (qtp1040856369-43520) [n:127.0.0.1:35905__ucp%2Ft 
   ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections 
params={collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:35905__ucp%252Ft&wt=javabin&version=2}
 status=0 QTime=2651
   [junit4]   2> 896768 INFO  (zkConnectionManagerCallback-3261-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 896769 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 896770 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41000/solr ready
   [junit4]   2> 896770 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection 
loss:false
   [junit4]   2> 896773 INFO  (qtp1040856369-43521) [n:127.0.0.1:35905__ucp%2Ft 
   ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params 
collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2
 and sendToOCPQueue=true
   [junit4]   2> 896777 INFO  
(OverseerThreadFactory-1954-thread-2-processing-n:127.0.0.1:35905__ucp%2Ft) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.c.a.c.CreateCollectionCmd Create 
collection collection1
   [junit4]   2> 896983 WARN  
(OverseerThreadFactory-1954-thread-2-processing-n:127.0.0.1:35905__ucp%2Ft) 
[n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.c.a.c.CreateCollectionCmd It is unusual 
to create a collection (collection1) without cores.
   [junit4]   2> 896986 INFO  (qtp1040856369-43521) [n:127.0.0.1:35905__ucp%2Ft 
   ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at 
most 30 seconds. Check all shard replicas
   [junit4]   2> 896988 INFO  (qtp1040856369-43521) [n:127.0.0.1:35905__ucp%2Ft 
   ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections 
params={collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2}
 status=0 QTime=214
   [junit4]   2> 897161 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001
 of type NRT
   [junit4]   2> 897163 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: 
d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12
   [junit4]   2> 897164 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 897164 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 897164 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 897165 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@153af2ab{/_ucp/t,null,AVAILABLE}
   [junit4]   2> 897165 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.AbstractConnector Started ServerConnector@682bcad2{SSL,[ssl, 
http/1.1]}{127.0.0.1:39772}
   [junit4]   2> 897165 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.Server Started @897266ms
   [junit4]   2> 897166 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: 
{solr.data.dir=hdfs://localhost:37822/hdfs__localhost_37822__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J1_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001_tempDir-002_jetty1,
 replicaType=NRT, solrconfig=solrconfig.xml, hostContext=/_ucp/t, 
hostPort=39772, 
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001/cores}
   [junit4]   2> 897166 ERROR 
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 897166 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter Using logger factory 
org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 897166 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 
7.7.0
   [junit4]   2> 897166 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 897166 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 897166 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2018-11-13T15:12:24.398Z
   [junit4]   2> 897178 INFO  (zkConnectionManagerCallback-3263-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 897179 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in 
ZooKeeper)
   [junit4]   2> 897179 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.SolrXmlConfig Loading container configuration from 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001/solr.xml
   [junit4]   2> 897184 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay 
is ignored
   [junit4]   2> 897184 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.SolrXmlConfig Configuration parameter 
autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 897186 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 897390 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41000/solr
   [junit4]   2> 897398 INFO  (zkConnectionManagerCallback-3267-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 897409 INFO  (zkConnectionManagerCallback-3269-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 897420 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 897425 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 
transient cores
   [junit4]   2> 897425 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:39772__ucp%2Ft
   [junit4]   2> 897426 INFO  (zkCallback-3242-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 897427 INFO  (zkCallback-3249-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 897433 INFO  (zkCallback-3260-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 897434 INFO  (zkCallback-3268-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 897460 INFO  (zkConnectionManagerCallback-3276-thread-1) [    
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 897463 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 897464 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41000/solr ready
   [junit4]   2> 897465 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history 
in memory.
   [junit4]   2> 897513 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') 
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3
   [junit4]   2> 897532 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') 
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3
   [junit4]   2> 897532 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 
'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3
   [junit4]   2> 897534 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001/cores
   [junit4]   2> 897574 INFO  (qtp1040856369-43520) [n:127.0.0.1:35905__ucp%2Ft 
   ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with 
params 
node=127.0.0.1:39772__ucp%252Ft&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2
 and sendToOCPQueue=true
   [junit4]   2> 897579 INFO  
(OverseerCollectionConfigSetProcessor-72271026745442309-127.0.0.1:35905__ucp%2Ft-n_0000000000)
 [n:127.0.0.1:35905__ucp%2Ft    ] o.a.s.c.OverseerTaskQueue Response ZK path: 
/overseer/collection-queue-work/qnr-0000000002 doesn't exist.  Requestor may 
have disconnected from ZooKeeper
   [junit4]   2> 897587 INFO  (qtp1040856369-43521) [n:127.0.0.1:35905__ucp%2Ft 
   ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics 
params={wt=javabin&version=2&key=solr.core.control_collection.shard1.replica_n1:INDEX.sizeInBytes}
 status=0 QTime=0
   [junit4]   2> 897589 INFO  (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft 
   ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics 
params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core}
 status=0 QTime=1
   [junit4]   2> 897627 INFO  (qtp1600864120-43574) [n:127.0.0.1:39772__ucp%2Ft 
   ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics 
params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core}
 status=0 QTime=0
   [junit4]   2> 897632 INFO  
(OverseerThreadFactory-1954-thread-3-processing-n:127.0.0.1:35905__ucp%2Ft) 
[n:127.0.0.1:35905__ucp%2Ft c:collection1 s:shard1  ] o.a.s.c.a.c.AddReplicaCmd 
Node Identified 127.0.0.1:39772__ucp%2Ft for creating new replica of shard 
shard1 for collection collection1
   [junit4]   2> 897642 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
   x:collection1_shard1_replica_n21] o.a.s.h.a.CoreAdminOperation core create 
command 
qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 898675 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.SolrConfig 
Using Lucene MatchVersion: 7.7.0
   [junit4]   2> 898713 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.s.IndexSchema 
[collection1_shard1_replica_n21] Schema name=test
   [junit4]   2> 898814 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.s.IndexSchema 
Loaded schema test/1.6 with uniqueid field id
   [junit4]   2> 898838 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.CoreContainer 
Creating SolrCore 'collection1_shard1_replica_n21' using configuration from 
collection collection1, trusted=true
   [junit4]   2> 898839 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.core.collection1.shard1.replica_n21' (registry 
'solr.core.collection1.shard1.replica_n21') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3
   [junit4]   2> 898840 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.SolrCore 
[[collection1_shard1_replica_n21] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001/cores/collection1_shard1_replica_n21],
 
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001/cores/collection1_shard1_replica_n21/data/]
   [junit4]   2> 898846 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: 
maxMergeAtOnce=33, maxMergeAtOnceExplicit=49, maxMergedSegmentMB=3.9248046875, 
floorSegmentMB=0.2802734375, forceMergeDeletesPctAllowed=26.06478682172601, 
segmentsPerTier=43.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, 
deletesPctAllowed=36.5804592175983
   [junit4]   2> 898867 WARN  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = 
requestHandler,name = /dump,class = DumpRequestHandler,attributes = 
{initParams=a, name=/dump, class=DumpRequestHandler},args = 
{defaults={a=A,b=B}}}
   [junit4]   2> 898996 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.u.UpdateHandler 
Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 898996 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.u.UpdateLog 
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 898999 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.u.CommitTracker 
Hard AutoCommit: if uncommitted for 30000ms; 
   [junit4]   2> 898999 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.u.CommitTracker 
Soft AutoCommit: if uncommitted for 3000ms; 
   [junit4]   2> 899002 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: 
maxMergeAtOnce=14, maxMergeAtOnceExplicit=11, maxMergedSegmentMB=18.544921875, 
floorSegmentMB=0.8408203125, forceMergeDeletesPctAllowed=10.035465587719308, 
segmentsPerTier=14.0, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.14629428284484744, deletesPctAllowed=40.71873956406105
   [junit4]   2> 899003 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.s.SolrIndexSearcher Opening 
[Searcher@3d7e87a4[collection1_shard1_replica_n21] main]
   [junit4]   2> 899005 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 899006 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 899007 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 899009 INFO  
(searcherExecutor-1973-thread-1-processing-n:127.0.0.1:39772__ucp%2Ft 
x:collection1_shard1_replica_n21 c:collection1 s:shard1) 
[n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1  
x:collection1_shard1_replica_n21] o.a.s.c.SolrCore 
[collection1_shard1_replica_n21] Registered new searcher 
Searcher@3d7e87a4[collection1_shard1_replica_n21] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 899010 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.u.UpdateLog 
Could not find max version in index or recent updates, using new clock 
1617032061902651392
   [junit4]   2> 899030 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.ZkShardTerms 
Successful update of terms at /collections/collection1/terms/shard1 to 
Terms{values={core_node22=0}, version=0}
   [junit4]   2> 899036 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 899036 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 899036 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy 
Sync replicas to https://127.0.0.1:39772/_ucp/t/collection1_shard1_replica_n21/
   [junit4]   2> 899037 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy 
Sync Success - now sync replicas to me
   [junit4]   2> 899037 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy 
https://127.0.0.1:39772/_ucp/t/collection1_shard1_replica_n21/ has no replicas
   [junit4]   2> 899037 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in 
election, clear LIR
   [junit4]   2> 899044 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] 
o.a.s.c.ShardLeaderElectionContext I am the new leader: 
https://127.0.0.1:39772/_ucp/t/collection1_shard1_replica_n21/ shard1
   [junit4]   2> 899149 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.c.ZkController 
I am the leader, no recovery necessary
   [junit4]   2> 899154 INFO  (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft 
c:collection1 s:shard1  x:collection1_shard1_replica_n21] o.a.s.s.HttpSolrCall 
[admin] webapp=null path=/admin/cores 
params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT}
 status=0 QTime=1512
   [junit4]   2> 899163 INFO  (qtp1040856369-43520) [n:127.0.0.1:35905__ucp%2Ft 
c:collection1   ] o.a.s.s.HttpSolrCall [admin] webapp=null 
path=/admin/collections 
params={node=127.0.0.1:39772__ucp%252Ft&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2}
 status=0 QTime=1588
   [junit4]   2> 899252 INFO  (zkCallback-3268-thread-1) [    ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/collection1/state.json] for collection [collection1] has 
occurred - updating... (live nodes size: [2])
   [junit4]   2> 899370 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-2-001
 of type NRT
   [junit4]   2> 899373 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: 
d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12
   [junit4]   2> 899374 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 899374 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 899374 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 899375 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@4ac3511b{/_ucp/t,null,AVAILABLE}
   [junit4]   2> 899375 INFO  
(TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [    ] 
o.e.j.s.AbstractConnector Started Se

[...truncated too long message...]

x/checkout/solr/core/test-lib/apacheds-interceptors-admin-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-authn-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-authz-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-changelog-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-collective-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-event-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-exception-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-journal-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-normalization-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-operational-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-referral-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-schema-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-subtree-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-trigger-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-jdbm-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-jdbm1-2.0.0-M2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-kerberos-codec-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-ldif-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-mavibot-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-protocol-kerberos-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-protocol-ldap-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-protocol-shared-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-xdbm-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/api-all-1.0.0-M20.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/bcprov-jdk15on-1.60.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/byte-buddy-1.6.2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/commons-collections-3.2.2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/ehcache-core-2.4.4.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/hadoop-common-2.7.4-tests.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/hadoop-hdfs-2.7.4-tests.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/hadoop-minikdc-2.7.4.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/htrace-core-3.2.0-incubating.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/jersey-core-1.9.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/jersey-server-1.9.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/jetty-6.1.26.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/jetty-sslengine-6.1.26.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/jetty-util-6.1.26.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/mina-core-2.0.0-M5.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/mockito-core-2.6.2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/netty-all-4.0.36.Final.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/objenesis-2.5.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/build/analysis/icu/lucene-analyzers-icu-7.7.0-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/contrib/solr-analysis-extras/classes/java:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/build/queryparser/classes/test:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/build/backward-codecs/classes/test:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/contrib/analysis-extras/lib/icu4j-62.1.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-launcher.jar:/x1/jenkins/.ant/lib/ivy-2.4.0.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-junit.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-log4j.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-junit4.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jai.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-javamail.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-bsf.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-commons-net.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-antlr.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jsch.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-oro.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-commons-logging.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-netrexx.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-testutil.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jdepend.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-bcel.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-xalan2.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-resolver.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jmf.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-regexp.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-swing.jar:/usr/local/asfpackages/java/jdk1.8.0_191/lib/tools.jar:/x1/jenkins/.ivy2/cache/com.carrotsearch.randomizedtesting/junit4-ant/jars/junit4-ant-2.6.4.jar
 com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe -eventsfile 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/temp/junit4-J0-20181113_145727_0205648669687998503296.events
 
@/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/temp/junit4-J0-20181113_145727_0201119673956783150849.suites
 -stdin
   [junit4] ERROR: JVM J0 ended with an exception: Quit event not received from 
the forked process? This may indicate JVM crash or runner bugs.
   [junit4]     at 
com.carrotsearch.ant.tasks.junit4.JUnit4.executeSlave(JUnit4.java:1563)
   [junit4]     at 
com.carrotsearch.ant.tasks.junit4.JUnit4.access$000(JUnit4.java:123)
   [junit4]     at 
com.carrotsearch.ant.tasks.junit4.JUnit4$2.call(JUnit4.java:997)
   [junit4]     at 
com.carrotsearch.ant.tasks.junit4.JUnit4$2.call(JUnit4.java:994)
   [junit4]     at java.util.concurrent.FutureTask.run(FutureTask.java:266)
   [junit4]     at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
   [junit4]     at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
   [junit4]     at java.lang.Thread.run(Thread.java:748)

BUILD FAILED
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/build.xml:651:
 The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/build.xml:577:
 The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/build.xml:59:
 The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build.xml:267:
 The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/common-build.xml:558:
 The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/common-build.xml:1567:
 The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/common-build.xml:1092:
 At least one slave process threw an exception, first: Quit event not received 
from the forked process? This may indicate JVM crash or runner bugs.

Total time: 222 minutes 37 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
[Fast Archiver] No artifacts from Lucene-Solr-NightlyTests-7.x #375 to compare, 
so performing full copy of artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

Reply via email to