Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-7.x/376/
4 tests failed. FAILED: org.apache.lucene.document.TestLatLonLineShapeQueries.testRandomBig Error Message: Java heap space Stack Trace: java.lang.OutOfMemoryError: Java heap space at __randomizedtesting.SeedInfo.seed([5186EC3284991EDD:D6D191BD15C0625D]:0) at org.apache.lucene.store.RAMFile.newBuffer(RAMFile.java:84) at org.apache.lucene.store.RAMFile.addBuffer(RAMFile.java:57) at org.apache.lucene.store.RAMOutputStream.switchCurrentBuffer(RAMOutputStream.java:168) at org.apache.lucene.store.RAMOutputStream.writeBytes(RAMOutputStream.java:154) at org.apache.lucene.store.MockIndexOutputWrapper.writeBytes(MockIndexOutputWrapper.java:137) at org.apache.lucene.util.bkd.OfflinePointWriter.append(OfflinePointWriter.java:75) at org.apache.lucene.util.bkd.BKDWriter.add(BKDWriter.java:287) at org.apache.lucene.index.RandomCodec$1$1$1.visit(RandomCodec.java:131) at org.apache.lucene.codecs.PointsWriter$1$1$1.visit(PointsWriter.java:117) at org.apache.lucene.index.AssertingLeafReader$AssertingIntersectVisitor.visit(AssertingLeafReader.java:1009) at org.apache.lucene.util.bkd.BKDReader.visitCompressedDocValues(BKDReader.java:709) at org.apache.lucene.util.bkd.BKDReader.visitDocValues(BKDReader.java:672) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:764) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:533) at org.apache.lucene.index.AssertingLeafReader$AssertingPointValues.intersect(AssertingLeafReader.java:898) at org.apache.lucene.codecs.PointsWriter$1$1.intersect(PointsWriter.java:105) at org.apache.lucene.index.RandomCodec$1$1.writeField(RandomCodec.java:124) at org.apache.lucene.codecs.PointsWriter.mergeOneField(PointsWriter.java:62) at org.apache.lucene.codecs.PointsWriter.merge(PointsWriter.java:191) FAILED: org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest.test Error Message: There are still nodes recoverying - waited for 320 seconds Stack Trace: java.lang.AssertionError: There are still nodes recoverying - waited for 320 seconds at __randomizedtesting.SeedInfo.seed([EFC03E3D27357B3D:679401E789C916C5]:0) at org.junit.Assert.fail(Assert.java:93) at org.apache.solr.cloud.AbstractDistribZkTestBase.waitForRecoveriesToFinish(AbstractDistribZkTestBase.java:185) at org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForRecoveriesToFinish(AbstractFullDistribZkTestBase.java:920) at org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForThingsToLevelOut(AbstractFullDistribZkTestBase.java:1477) at org.apache.solr.cloud.RestartWhileUpdatingTest.test(RestartWhileUpdatingTest.java:145) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1742) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:935) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:971) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:985) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:1010) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:985) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:944) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:830) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:880) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.lang.Thread.run(Thread.java:748) FAILED: junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest Error Message: 7 threads leaked from SUITE scope at org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest: 1) Thread[id=43682, name=searcherExecutor-2001-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 2) Thread[id=43559, name=searcherExecutor-1959-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 3) Thread[id=43879, name=searcherExecutor-2021-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 4) Thread[id=43914, name=searcherExecutor-2035-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 5) Thread[id=43949, name=searcherExecutor-2049-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 6) Thread[id=43600, name=searcherExecutor-1973-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 7) Thread[id=43637, name=searcherExecutor-1987-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Stack Trace: com.carrotsearch.randomizedtesting.ThreadLeakError: 7 threads leaked from SUITE scope at org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest: 1) Thread[id=43682, name=searcherExecutor-2001-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 2) Thread[id=43559, name=searcherExecutor-1959-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 3) Thread[id=43879, name=searcherExecutor-2021-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 4) Thread[id=43914, name=searcherExecutor-2035-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 5) Thread[id=43949, name=searcherExecutor-2049-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 6) Thread[id=43600, name=searcherExecutor-1973-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 7) Thread[id=43637, name=searcherExecutor-1987-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) at __randomizedtesting.SeedInfo.seed([EFC03E3D27357B3D]:0) FAILED: junit.framework.TestSuite.org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest Error Message: There are still zombie threads that couldn't be terminated: 1) Thread[id=43682, name=searcherExecutor-2001-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 2) Thread[id=43559, name=searcherExecutor-1959-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 3) Thread[id=43879, name=searcherExecutor-2021-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 4) Thread[id=43914, name=searcherExecutor-2035-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 5) Thread[id=43949, name=searcherExecutor-2049-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 6) Thread[id=43600, name=searcherExecutor-1973-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 7) Thread[id=43637, name=searcherExecutor-1987-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Stack Trace: com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie threads that couldn't be terminated: 1) Thread[id=43682, name=searcherExecutor-2001-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 2) Thread[id=43559, name=searcherExecutor-1959-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 3) Thread[id=43879, name=searcherExecutor-2021-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 4) Thread[id=43914, name=searcherExecutor-2035-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 5) Thread[id=43949, name=searcherExecutor-2049-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 6) Thread[id=43600, name=searcherExecutor-1973-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 7) Thread[id=43637, name=searcherExecutor-1987-thread-1, state=WAITING, group=TGRP-HdfsRestartWhileUpdatingTest] at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) at __randomizedtesting.SeedInfo.seed([EFC03E3D27357B3D]:0) Build Log: [...truncated 10294 lines...] [junit4] Suite: org.apache.lucene.document.TestLatLonLineShapeQueries [junit4] 2> NOTE: download the large Jenkins line-docs file by running 'ant get-jenkins-line-docs' in the lucene directory. [junit4] 2> NOTE: reproduce with: ant test -Dtestcase=TestLatLonLineShapeQueries -Dtests.method=testRandomBig -Dtests.seed=5186EC3284991EDD -Dtests.multiplier=2 -Dtests.nightly=true -Dtests.slow=true -Dtests.linedocsfile=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/test-data/enwiki.random.lines.txt -Dtests.locale=de-CH -Dtests.timezone=America/Coral_Harbour -Dtests.asserts=true -Dtests.file.encoding=ISO-8859-1 [junit4] ERROR 749s J2 | TestLatLonLineShapeQueries.testRandomBig <<< [junit4] > Throwable #1: java.lang.OutOfMemoryError: Java heap space [junit4] > at __randomizedtesting.SeedInfo.seed([5186EC3284991EDD:D6D191BD15C0625D]:0) [junit4] > at org.apache.lucene.store.RAMFile.newBuffer(RAMFile.java:84) [junit4] > at org.apache.lucene.store.RAMFile.addBuffer(RAMFile.java:57) [junit4] > at org.apache.lucene.store.RAMOutputStream.switchCurrentBuffer(RAMOutputStream.java:168) [junit4] > at org.apache.lucene.store.RAMOutputStream.writeBytes(RAMOutputStream.java:154) [junit4] > at org.apache.lucene.store.MockIndexOutputWrapper.writeBytes(MockIndexOutputWrapper.java:137) [junit4] > at org.apache.lucene.util.bkd.OfflinePointWriter.append(OfflinePointWriter.java:75) [junit4] > at org.apache.lucene.util.bkd.BKDWriter.add(BKDWriter.java:287) [junit4] > at org.apache.lucene.index.RandomCodec$1$1$1.visit(RandomCodec.java:131) [junit4] > at org.apache.lucene.codecs.PointsWriter$1$1$1.visit(PointsWriter.java:117) [junit4] > at org.apache.lucene.index.AssertingLeafReader$AssertingIntersectVisitor.visit(AssertingLeafReader.java:1009) [junit4] > at org.apache.lucene.util.bkd.BKDReader.visitCompressedDocValues(BKDReader.java:709) [junit4] > at org.apache.lucene.util.bkd.BKDReader.visitDocValues(BKDReader.java:672) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:764) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:787) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:797) [junit4] > at org.apache.lucene.util.bkd.BKDReader.intersect(BKDReader.java:533) [junit4] > at org.apache.lucene.index.AssertingLeafReader$AssertingPointValues.intersect(AssertingLeafReader.java:898) [junit4] > at org.apache.lucene.codecs.PointsWriter$1$1.intersect(PointsWriter.java:105) [junit4] > at org.apache.lucene.index.RandomCodec$1$1.writeField(RandomCodec.java:124) [junit4] > at org.apache.lucene.codecs.PointsWriter.mergeOneField(PointsWriter.java:62) [junit4] > at org.apache.lucene.codecs.PointsWriter.merge(PointsWriter.java:191) [junit4] 2> NOTE: leaving temporary files on disk at: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/build/sandbox/test/J2/temp/lucene.document.TestLatLonLineShapeQueries_5186EC3284991EDD-001 [junit4] 2> NOTE: test params are: codec=Asserting(Lucene70): {id=FST50}, docValues:{shape=DocValuesFormat(name=Asserting), id=DocValuesFormat(name=Lucene70)}, maxPointsInLeafNode=999, maxMBSortInHeap=6.293844669044266, sim=RandomSimilarity(queryNorm=true): {}, locale=de-CH, timezone=America/Coral_Harbour [junit4] 2> NOTE: Linux 4.4.0-137-generic amd64/Oracle Corporation 1.8.0_191 (64-bit)/cpus=4,threads=1,free=285326408,total=440926208 [junit4] 2> NOTE: All tests run in this JVM: [TestLatLonBoundingBoxQueries, TestLatLonShape, TestDoubleRangeField, TestLatLonLineShapeQueries] [junit4] Completed [19/22 (1!)] on J2 in 1096.68s, 4 tests, 1 error <<< FAILURES! [...truncated 1 lines...] [junit4] JVM J2: stdout was not empty, see: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/build/sandbox/test/temp/junit4-J2-20181113_135933_6228772371157596158491.sysout [junit4] >>> JVM J2 emitted unexpected output (verbatim) ---- [junit4] java.lang.OutOfMemoryError: Java heap space [junit4] Dumping heap to /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/heapdumps/java_pid15722.hprof ... [junit4] Heap dump file created [462447318 bytes in 0.805 secs] [junit4] <<< JVM J2: EOF ---- [...truncated 3234 lines...] [junit4] Suite: org.apache.solr.cloud.hdfs.HdfsRestartWhileUpdatingTest [junit4] 2> Creating dataDir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/init-core-data-001 [junit4] 2> 884469 WARN (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=279 numCloses=279 [junit4] 2> 884470 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=false [junit4] 2> 884472 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (true) via: @org.apache.solr.util.RandomizeSSL(reason=, ssl=NaN, value=NaN, clientAuth=NaN) [junit4] 2> 884473 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom [junit4] 2> 884473 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /_ucp/t [junit4] 1> Formatting using clusterid: testClusterID [junit4] 2> 886381 WARN (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties [junit4] 2> 886708 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.m.log Logging to org.apache.logging.slf4j.Log4jLogger@385a6de4 via org.mortbay.log.Slf4jLog [junit4] 2> 886737 WARN (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j [junit4] 2> 886840 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.m.log jetty-6.1.26 [junit4] 2> 886913 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs to ./temp/Jetty_localhost_46064_hdfs____.ao98ve/webapp [junit4] 2> 887807 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:46064 [junit4] 2> 889145 WARN (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j [junit4] 2> 889156 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.m.log jetty-6.1.26 [junit4] 2> 889177 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_46434_datanode____hprwhm/webapp [junit4] 2> 889627 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:46434 [junit4] 2> 890396 WARN (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j [junit4] 2> 890398 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.m.log jetty-6.1.26 [junit4] 2> 890465 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_41325_datanode____.wise4g/webapp [junit4] 2> 891344 INFO (SUITE-HdfsRestartWhileUpdatingTest-seed#[EFC03E3D27357B3D]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:41325 [junit4] 2> 891885 ERROR (DataNode: [[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/tempDir-001/hdfsBaseDir/data/data1/, [DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/tempDir-001/hdfsBaseDir/data/data2/]] heartbeating to localhost/127.0.0.1:37822) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000 [junit4] 2> 891887 ERROR (DataNode: [[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/tempDir-001/hdfsBaseDir/data/data3/, [DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/tempDir-001/hdfsBaseDir/data/data4/]] heartbeating to localhost/127.0.0.1:37822) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000 [junit4] 2> 892135 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xb91f918a9cb56: from storage DS-320ccfa3-68d6-4847-abf1-9fd841aea974 node DatanodeRegistration(127.0.0.1:34595, datanodeUuid=d43953ae-b8dc-464d-8815-026a85d6d206, infoPort=38832, infoSecurePort=0, ipcPort=41708, storageInfo=lv=-56;cid=testClusterID;nsid=1189364975;c=0), blocks: 0, hasStaleStorage: true, processing time: 6 msecs [junit4] 2> 892136 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xb91f91845f1b7: from storage DS-a8c82a40-70f3-4f1a-8323-4186be7d4843 node DatanodeRegistration(127.0.0.1:39605, datanodeUuid=49051a15-2b0b-488c-95aa-dffb4e893cb4, infoPort=42551, infoSecurePort=0, ipcPort=43971, storageInfo=lv=-56;cid=testClusterID;nsid=1189364975;c=0), blocks: 0, hasStaleStorage: true, processing time: 1 msecs [junit4] 2> 892136 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xb91f918a9cb56: from storage DS-cdecb5ea-8e3a-49b8-bd0f-7ce1646c22c3 node DatanodeRegistration(127.0.0.1:34595, datanodeUuid=d43953ae-b8dc-464d-8815-026a85d6d206, infoPort=38832, infoSecurePort=0, ipcPort=41708, storageInfo=lv=-56;cid=testClusterID;nsid=1189364975;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs [junit4] 2> 892141 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xb91f91845f1b7: from storage DS-760520a3-b796-4e7e-953e-5da6d48b7a15 node DatanodeRegistration(127.0.0.1:39605, datanodeUuid=49051a15-2b0b-488c-95aa-dffb4e893cb4, infoPort=42551, infoSecurePort=0, ipcPort=43971, storageInfo=lv=-56;cid=testClusterID;nsid=1189364975;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs [junit4] 2> 892982 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER [junit4] 2> 892983 INFO (Thread-37972) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0 [junit4] 2> 892983 INFO (Thread-37972) [ ] o.a.s.c.ZkTestServer Starting server [junit4] 2> 892987 ERROR (Thread-37972) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes [junit4] 2> 893083 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.ZkTestServer start zk server on port:41000 [junit4] 2> 893108 INFO (zkConnectionManagerCallback-3229-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 893112 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100c21dacb70000, likely client has closed socket [junit4] 2> 893123 INFO (zkConnectionManagerCallback-3231-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 893130 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml [junit4] 2> 893133 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/schema15.xml to /configs/conf1/schema.xml [junit4] 2> 893134 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml to /configs/conf1/solrconfig.snippet.randomindexconfig.xml [junit4] 2> 893136 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt [junit4] 2> 893138 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt [junit4] 2> 893139 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml [junit4] 2> 893141 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml to /configs/conf1/enumsConfig.xml [junit4] 2> 893142 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json [junit4] 2> 893151 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt [junit4] 2> 893156 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt [junit4] 2> 893157 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/src/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt [junit4] 2> 893161 INFO (zkConnectionManagerCallback-3235-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 893164 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractFullDistribZkTestBase Will use NRT replicas unless explicitly asked otherwise [junit4] 2> 893427 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12 [junit4] 2> 893489 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0 [junit4] 2> 893489 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.session No SessionScavenger set, using defaults [junit4] 2> 893489 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.session node0 Scavenging every 600000ms [junit4] 2> 893492 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@c6ebc75{/_ucp/t,null,AVAILABLE} [junit4] 2> 893500 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@3e77f6ea{SSL,[ssl, http/1.1]}{127.0.0.1:35905} [junit4] 2> 893500 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.Server Started @893601ms [junit4] 2> 893500 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=hdfs://localhost:37822/hdfs__localhost_37822__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J1_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001_tempDir-002_control_data, replicaType=NRT, hostContext=/_ucp/t, hostPort=35905, coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/control-001/cores} [junit4] 2> 893501 ERROR (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete. [junit4] 2> 893501 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory [junit4] 2> 893501 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr? version 7.7.0 [junit4] 2> 893501 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null [junit4] 2> 893501 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null [junit4] 2> 893501 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-11-13T15:12:20.733Z [junit4] 2> 893530 INFO (zkConnectionManagerCallback-3237-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 893532 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper) [junit4] 2> 893532 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/control-001/solr.xml [junit4] 2> 893538 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored [junit4] 2> 893538 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored [junit4] 2> 893540 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3, but no JMX reporters were configured - adding default JMX reporter. [junit4] 2> 893764 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41000/solr [junit4] 2> 893766 INFO (zkConnectionManagerCallback-3241-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 893769 INFO (zkConnectionManagerCallback-3243-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 893904 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:35905__ucp%2Ft [junit4] 2> 893906 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.c.Overseer Overseer (id=72271026745442309-127.0.0.1:35905__ucp%2Ft-n_0000000000) starting [junit4] 2> 893931 INFO (zkConnectionManagerCallback-3250-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 893934 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41000/solr ready [junit4] 2> 893938 INFO (OverseerStateUpdate-72271026745442309-127.0.0.1:35905__ucp%2Ft-n_0000000000) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.c.Overseer Starting to work on the main queue : 127.0.0.1:35905__ucp%2Ft [junit4] 2> 893942 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:35905__ucp%2Ft [junit4] 2> 893968 INFO (zkCallback-3242-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 893969 INFO (zkCallback-3249-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 893991 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory. [junit4] 2> 894020 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3 [junit4] 2> 894034 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3 [junit4] 2> 894034 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3 [junit4] 2> 894036 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/control-001/cores [junit4] 2> 894066 INFO (zkConnectionManagerCallback-3256-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 894067 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 894069 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41000/solr ready [junit4] 2> 894110 INFO (qtp1040856369-43520) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:35905__ucp%252Ft&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 894120 INFO (OverseerThreadFactory-1954-thread-1-processing-n:127.0.0.1:35905__ucp%2Ft) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.c.a.c.CreateCollectionCmd Create collection control_collection [junit4] 2> 894237 INFO (qtp1040856369-43514) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core} status=0 QTime=0 [junit4] 2> 894249 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft x:control_collection_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT [junit4] 2> 894249 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft x:control_collection_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores [junit4] 2> 895279 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.7.0 [junit4] 2> 895296 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test [junit4] 2> 895423 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id [junit4] 2> 895463 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1' using configuration from collection control_collection, trusted=true [junit4] 2> 895463 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.control_collection.shard1.replica_n1' (registry 'solr.core.control_collection.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3 [junit4] 2> 895464 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore at [/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/control-001/cores/control_collection_shard1_replica_n1], dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/../../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/control-001/cores/control_collection_shard1_replica_n1/data/] [junit4] 2> 895469 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=33, maxMergeAtOnceExplicit=49, maxMergedSegmentMB=3.9248046875, floorSegmentMB=0.2802734375, forceMergeDeletesPctAllowed=26.06478682172601, segmentsPerTier=43.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, deletesPctAllowed=36.5804592175983 [junit4] 2> 895481 WARN (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}} [junit4] 2> 895555 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog [junit4] 2> 895555 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536 [junit4] 2> 895556 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: if uncommitted for 30000ms; [junit4] 2> 895556 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: if uncommitted for 3000ms; [junit4] 2> 895560 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=14, maxMergeAtOnceExplicit=11, maxMergedSegmentMB=18.544921875, floorSegmentMB=0.8408203125, forceMergeDeletesPctAllowed=10.035465587719308, segmentsPerTier=14.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.14629428284484744, deletesPctAllowed=40.71873956406105 [junit4] 2> 895561 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@4f901b73[control_collection_shard1_replica_n1] main] [junit4] 2> 895562 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1 [junit4] 2> 895563 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1 [junit4] 2> 895564 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms. [junit4] 2> 895564 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1617032058289258496 [junit4] 2> 895568 INFO (searcherExecutor-1959-thread-1-processing-n:127.0.0.1:35905__ucp%2Ft x:control_collection_shard1_replica_n1 c:control_collection s:shard1) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [control_collection_shard1_replica_n1] Registered new searcher Searcher@4f901b73[control_collection_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())} [junit4] 2> 895573 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/control_collection/terms/shard1 to Terms{values={core_node2=0}, version=0} [junit4] 2> 895579 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue. [junit4] 2> 895579 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync [junit4] 2> 895579 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:35905/_ucp/t/control_collection_shard1_replica_n1/ [junit4] 2> 895579 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me [junit4] 2> 895580 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:35905/_ucp/t/control_collection_shard1_replica_n1/ has no replicas [junit4] 2> 895580 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR [junit4] 2> 895586 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:35905/_ucp/t/control_collection_shard1_replica_n1/ shard1 [junit4] 2> 895741 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary [junit4] 2> 895745 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1497 [junit4] 2> 895762 INFO (qtp1040856369-43520) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas [junit4] 2> 895856 INFO (zkCallback-3242-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/control_collection/state.json] for collection [control_collection] has occurred - updating... (live nodes size: [1]) [junit4] 2> 896126 INFO (OverseerCollectionConfigSetProcessor-72271026745442309-127.0.0.1:35905__ucp%2Ft-n_0000000000) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper [junit4] 2> 896762 INFO (qtp1040856369-43520) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:35905__ucp%252Ft&wt=javabin&version=2} status=0 QTime=2651 [junit4] 2> 896768 INFO (zkConnectionManagerCallback-3261-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 896769 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 896770 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41000/solr ready [junit4] 2> 896770 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection loss:false [junit4] 2> 896773 INFO (qtp1040856369-43521) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 896777 INFO (OverseerThreadFactory-1954-thread-2-processing-n:127.0.0.1:35905__ucp%2Ft) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.c.a.c.CreateCollectionCmd Create collection collection1 [junit4] 2> 896983 WARN (OverseerThreadFactory-1954-thread-2-processing-n:127.0.0.1:35905__ucp%2Ft) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.c.a.c.CreateCollectionCmd It is unusual to create a collection (collection1) without cores. [junit4] 2> 896986 INFO (qtp1040856369-43521) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas [junit4] 2> 896988 INFO (qtp1040856369-43521) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2} status=0 QTime=214 [junit4] 2> 897161 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001 of type NRT [junit4] 2> 897163 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12 [junit4] 2> 897164 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0 [junit4] 2> 897164 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.session No SessionScavenger set, using defaults [junit4] 2> 897164 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.session node0 Scavenging every 660000ms [junit4] 2> 897165 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@153af2ab{/_ucp/t,null,AVAILABLE} [junit4] 2> 897165 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@682bcad2{SSL,[ssl, http/1.1]}{127.0.0.1:39772} [junit4] 2> 897165 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.Server Started @897266ms [junit4] 2> 897166 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=hdfs://localhost:37822/hdfs__localhost_37822__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-7.x_checkout_solr_build_solr-core_test_J1_temp_solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001_tempDir-002_jetty1, replicaType=NRT, solrconfig=solrconfig.xml, hostContext=/_ucp/t, hostPort=39772, coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001/cores} [junit4] 2> 897166 ERROR (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete. [junit4] 2> 897166 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory [junit4] 2> 897166 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr? version 7.7.0 [junit4] 2> 897166 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null [junit4] 2> 897166 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null [junit4] 2> 897166 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-11-13T15:12:24.398Z [junit4] 2> 897178 INFO (zkConnectionManagerCallback-3263-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 897179 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper) [junit4] 2> 897179 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001/solr.xml [junit4] 2> 897184 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored [junit4] 2> 897184 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored [junit4] 2> 897186 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3, but no JMX reporters were configured - adding default JMX reporter. [junit4] 2> 897390 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:41000/solr [junit4] 2> 897398 INFO (zkConnectionManagerCallback-3267-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 897409 INFO (zkConnectionManagerCallback-3269-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 897420 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 897425 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores [junit4] 2> 897425 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:39772__ucp%2Ft [junit4] 2> 897426 INFO (zkCallback-3242-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 897427 INFO (zkCallback-3249-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 897433 INFO (zkCallback-3260-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 897434 INFO (zkCallback-3268-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 897460 INFO (zkConnectionManagerCallback-3276-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 897463 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2) [junit4] 2> 897464 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:41000/solr ready [junit4] 2> 897465 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory. [junit4] 2> 897513 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3 [junit4] 2> 897532 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3 [junit4] 2> 897532 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3 [junit4] 2> 897534 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001/cores [junit4] 2> 897574 INFO (qtp1040856369-43520) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:39772__ucp%252Ft&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 897579 INFO (OverseerCollectionConfigSetProcessor-72271026745442309-127.0.0.1:35905__ucp%2Ft-n_0000000000) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000002 doesn't exist. Requestor may have disconnected from ZooKeeper [junit4] 2> 897587 INFO (qtp1040856369-43521) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics params={wt=javabin&version=2&key=solr.core.control_collection.shard1.replica_n1:INDEX.sizeInBytes} status=0 QTime=0 [junit4] 2> 897589 INFO (qtp1040856369-43518) [n:127.0.0.1:35905__ucp%2Ft ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core} status=0 QTime=1 [junit4] 2> 897627 INFO (qtp1600864120-43574) [n:127.0.0.1:39772__ucp%2Ft ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/metrics params={prefix=CONTAINER.fs.usableSpace,CONTAINER.fs.totalSpace,CORE.coreName&wt=javabin&version=2&group=solr.node,solr.core} status=0 QTime=0 [junit4] 2> 897632 INFO (OverseerThreadFactory-1954-thread-3-processing-n:127.0.0.1:35905__ucp%2Ft) [n:127.0.0.1:35905__ucp%2Ft c:collection1 s:shard1 ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:39772__ucp%2Ft for creating new replica of shard shard1 for collection collection1 [junit4] 2> 897642 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft x:collection1_shard1_replica_n21] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT [junit4] 2> 898675 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.7.0 [junit4] 2> 898713 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.s.IndexSchema [collection1_shard1_replica_n21] Schema name=test [junit4] 2> 898814 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id [junit4] 2> 898838 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard1_replica_n21' using configuration from collection collection1, trusted=true [junit4] 2> 898839 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard1.replica_n21' (registry 'solr.core.collection1.shard1.replica_n21') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4ba098d3 [junit4] 2> 898840 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SolrCore [[collection1_shard1_replica_n21] ] Opening new SolrCore at [/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001/cores/collection1_shard1_replica_n21], dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-1-001/cores/collection1_shard1_replica_n21/data/] [junit4] 2> 898846 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=33, maxMergeAtOnceExplicit=49, maxMergedSegmentMB=3.9248046875, floorSegmentMB=0.2802734375, forceMergeDeletesPctAllowed=26.06478682172601, segmentsPerTier=43.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, deletesPctAllowed=36.5804592175983 [junit4] 2> 898867 WARN (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}} [junit4] 2> 898996 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog [junit4] 2> 898996 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536 [junit4] 2> 898999 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.CommitTracker Hard AutoCommit: if uncommitted for 30000ms; [junit4] 2> 898999 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.CommitTracker Soft AutoCommit: if uncommitted for 3000ms; [junit4] 2> 899002 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=14, maxMergeAtOnceExplicit=11, maxMergedSegmentMB=18.544921875, floorSegmentMB=0.8408203125, forceMergeDeletesPctAllowed=10.035465587719308, segmentsPerTier=14.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.14629428284484744, deletesPctAllowed=40.71873956406105 [junit4] 2> 899003 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.s.SolrIndexSearcher Opening [Searcher@3d7e87a4[collection1_shard1_replica_n21] main] [junit4] 2> 899005 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1 [junit4] 2> 899006 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1 [junit4] 2> 899007 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms. [junit4] 2> 899009 INFO (searcherExecutor-1973-thread-1-processing-n:127.0.0.1:39772__ucp%2Ft x:collection1_shard1_replica_n21 c:collection1 s:shard1) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SolrCore [collection1_shard1_replica_n21] Registered new searcher Searcher@3d7e87a4[collection1_shard1_replica_n21] main{ExitableDirectoryReader(UninvertingDirectoryReader())} [junit4] 2> 899010 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1617032061902651392 [junit4] 2> 899030 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard1 to Terms{values={core_node22=0}, version=0} [junit4] 2> 899036 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue. [junit4] 2> 899036 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync [junit4] 2> 899036 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:39772/_ucp/t/collection1_shard1_replica_n21/ [junit4] 2> 899037 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me [junit4] 2> 899037 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.SyncStrategy https://127.0.0.1:39772/_ucp/t/collection1_shard1_replica_n21/ has no replicas [junit4] 2> 899037 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR [junit4] 2> 899044 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:39772/_ucp/t/collection1_shard1_replica_n21/ shard1 [junit4] 2> 899149 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.c.ZkController I am the leader, no recovery necessary [junit4] 2> 899154 INFO (qtp1600864120-43567) [n:127.0.0.1:39772__ucp%2Ft c:collection1 s:shard1 x:collection1_shard1_replica_n21] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1512 [junit4] 2> 899163 INFO (qtp1040856369-43520) [n:127.0.0.1:35905__ucp%2Ft c:collection1 ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={node=127.0.0.1:39772__ucp%252Ft&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2} status=0 QTime=1588 [junit4] 2> 899252 INFO (zkCallback-3268-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2]) [junit4] 2> 899370 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/J1/temp/solr.cloud.hdfs.HdfsRestartWhileUpdatingTest_EFC03E3D27357B3D-001/shard-2-001 of type NRT [junit4] 2> 899373 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_191-b12 [junit4] 2> 899374 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0 [junit4] 2> 899374 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.session No SessionScavenger set, using defaults [junit4] 2> 899374 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.session node0 Scavenging every 660000ms [junit4] 2> 899375 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@4ac3511b{/_ucp/t,null,AVAILABLE} [junit4] 2> 899375 INFO (TEST-HdfsRestartWhileUpdatingTest.test-seed#[EFC03E3D27357B3D]) [ ] o.e.j.s.AbstractConnector Started Se [...truncated too long message...] x/checkout/solr/core/test-lib/apacheds-interceptors-admin-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-authn-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-authz-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-changelog-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-collective-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-event-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-exception-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-journal-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-normalization-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-operational-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-referral-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-schema-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-subtree-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-interceptors-trigger-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-jdbm-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-jdbm1-2.0.0-M2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-kerberos-codec-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-ldif-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-mavibot-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-protocol-kerberos-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-protocol-ldap-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-protocol-shared-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/apacheds-xdbm-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/api-all-1.0.0-M20.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/bcprov-jdk15on-1.60.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/byte-buddy-1.6.2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/commons-collections-3.2.2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/ehcache-core-2.4.4.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/hadoop-common-2.7.4-tests.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/hadoop-hdfs-2.7.4-tests.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/hadoop-minikdc-2.7.4.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/htrace-core-3.2.0-incubating.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/jersey-core-1.9.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/jersey-server-1.9.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/jetty-6.1.26.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/jetty-sslengine-6.1.26.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/jetty-util-6.1.26.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/mina-core-2.0.0-M5.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/mockito-core-2.6.2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/netty-all-4.0.36.Final.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/core/test-lib/objenesis-2.5.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/build/analysis/icu/lucene-analyzers-icu-7.7.0-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/contrib/solr-analysis-extras/classes/java:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/build/queryparser/classes/test:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/build/backward-codecs/classes/test:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/contrib/analysis-extras/lib/icu4j-62.1.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-launcher.jar:/x1/jenkins/.ant/lib/ivy-2.4.0.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-junit.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-log4j.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-junit4.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jai.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-javamail.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-bsf.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-commons-net.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-antlr.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jsch.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-oro.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-commons-logging.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-netrexx.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-testutil.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jdepend.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-bcel.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-xalan2.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-resolver.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jmf.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-regexp.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-swing.jar:/usr/local/asfpackages/java/jdk1.8.0_191/lib/tools.jar:/x1/jenkins/.ivy2/cache/com.carrotsearch.randomizedtesting/junit4-ant/jars/junit4-ant-2.6.4.jar com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe -eventsfile /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/temp/junit4-J0-20181113_145727_0205648669687998503296.events @/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build/solr-core/test/temp/junit4-J0-20181113_145727_0201119673956783150849.suites -stdin [junit4] ERROR: JVM J0 ended with an exception: Quit event not received from the forked process? This may indicate JVM crash or runner bugs. [junit4] at com.carrotsearch.ant.tasks.junit4.JUnit4.executeSlave(JUnit4.java:1563) [junit4] at com.carrotsearch.ant.tasks.junit4.JUnit4.access$000(JUnit4.java:123) [junit4] at com.carrotsearch.ant.tasks.junit4.JUnit4$2.call(JUnit4.java:997) [junit4] at com.carrotsearch.ant.tasks.junit4.JUnit4$2.call(JUnit4.java:994) [junit4] at java.util.concurrent.FutureTask.run(FutureTask.java:266) [junit4] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [junit4] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [junit4] at java.lang.Thread.run(Thread.java:748) BUILD FAILED /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/build.xml:651: The following error occurred while executing this line: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/build.xml:577: The following error occurred while executing this line: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/build.xml:59: The following error occurred while executing this line: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/build.xml:267: The following error occurred while executing this line: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/solr/common-build.xml:558: The following error occurred while executing this line: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/common-build.xml:1567: The following error occurred while executing this line: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-7.x/checkout/lucene/common-build.xml:1092: At least one slave process threw an exception, first: Quit event not received from the forked process? This may indicate JVM crash or runner bugs. Total time: 222 minutes 37 seconds Build step 'Invoke Ant' marked build as failure Archiving artifacts [Fast Archiver] No artifacts from Lucene-Solr-NightlyTests-7.x #375 to compare, so performing full copy of artifacts Recording test results Email was triggered for: Failure - Any Sending email for trigger: Failure - Any
--------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org