Build: https://builds.apache.org/job/Lucene-Solr-Tests-master/2415/
6 tests failed. FAILED: junit.framework.TestSuite.org.apache.solr.cloud.ZkControllerTest Error Message: ObjectTracker found 1 object(s) that were not released!!! [Overseer] org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: org.apache.solr.cloud.Overseer at org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42) at org.apache.solr.cloud.Overseer.start(Overseer.java:545) at org.apache.solr.cloud.OverseerElectionContext.runLeaderProcess(ElectionContext.java:850) at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:170) at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:135) at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:307) at org.apache.solr.cloud.LeaderElector.retryElection(LeaderElector.java:393) at org.apache.solr.cloud.ZkController.rejoinOverseerElection(ZkController.java:2055) at org.apache.solr.cloud.Overseer$ClusterStateUpdater.checkIfIamStillLeader(Overseer.java:331) at java.lang.Thread.run(Thread.java:748) Stack Trace: java.lang.AssertionError: ObjectTracker found 1 object(s) that were not released!!! [Overseer] org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: org.apache.solr.cloud.Overseer at org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42) at org.apache.solr.cloud.Overseer.start(Overseer.java:545) at org.apache.solr.cloud.OverseerElectionContext.runLeaderProcess(ElectionContext.java:850) at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:170) at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:135) at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:307) at org.apache.solr.cloud.LeaderElector.retryElection(LeaderElector.java:393) at org.apache.solr.cloud.ZkController.rejoinOverseerElection(ZkController.java:2055) at org.apache.solr.cloud.Overseer$ClusterStateUpdater.checkIfIamStillLeader(Overseer.java:331) at java.lang.Thread.run(Thread.java:748) at __randomizedtesting.SeedInfo.seed([82B128B6ED4F7220]:0) at org.junit.Assert.fail(Assert.java:93) at org.junit.Assert.assertTrue(Assert.java:43) at org.junit.Assert.assertNull(Assert.java:551) at org.apache.solr.SolrTestCaseJ4.teardownTestCases(SolrTestCaseJ4.java:301) at sun.reflect.GeneratedMethodAccessor48.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:897) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.lang.Thread.run(Thread.java:748) FAILED: junit.framework.TestSuite.org.apache.solr.cloud.ZkControllerTest Error Message: 1 thread leaked from SUITE scope at org.apache.solr.cloud.ZkControllerTest: 1) Thread[id=14946, name=OverseerAutoScalingTriggerThread-72262182530908162-127.0.0.1:8983_solr-n_0000000001, state=RUNNABLE, group=Overseer autoscaling triggers] at java.security.AccessController.doPrivileged(Native Method) at com.carrotsearch.ant.tasks.junit4.events.Serializer.flushQueue(Serializer.java:114) at com.carrotsearch.ant.tasks.junit4.events.Serializer.serialize(Serializer.java:99) at com.carrotsearch.ant.tasks.junit4.slave.SlaveMain$3$2.write(SlaveMain.java:472) at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140) at java.io.PrintStream.flush(PrintStream.java:338) at java.io.FilterOutputStream.flush(FilterOutputStream.java:140) at java.io.PrintStream.write(PrintStream.java:482) at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:221) at sun.nio.cs.StreamEncoder.implFlushBuffer(StreamEncoder.java:291) at sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:295) at sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:141) at java.io.OutputStreamWriter.flush(OutputStreamWriter.java:229) at org.apache.log4j.helpers.QuietWriter.flush(QuietWriter.java:59) at org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:324) at org.apache.log4j.WriterAppender.append(WriterAppender.java:162) at org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251) at org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66) at org.apache.log4j.Category.callAppenders(Category.java:206) at org.apache.log4j.Category.forcedLog(Category.java:391) at org.apache.log4j.Category.log(Category.java:856) at org.slf4j.impl.Log4jLoggerAdapter.error(Log4jLoggerAdapter.java:576) at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:142) at java.lang.Thread.run(Thread.java:748) Stack Trace: com.carrotsearch.randomizedtesting.ThreadLeakError: 1 thread leaked from SUITE scope at org.apache.solr.cloud.ZkControllerTest: 1) Thread[id=14946, name=OverseerAutoScalingTriggerThread-72262182530908162-127.0.0.1:8983_solr-n_0000000001, state=RUNNABLE, group=Overseer autoscaling triggers] at java.security.AccessController.doPrivileged(Native Method) at com.carrotsearch.ant.tasks.junit4.events.Serializer.flushQueue(Serializer.java:114) at com.carrotsearch.ant.tasks.junit4.events.Serializer.serialize(Serializer.java:99) at com.carrotsearch.ant.tasks.junit4.slave.SlaveMain$3$2.write(SlaveMain.java:472) at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140) at java.io.PrintStream.flush(PrintStream.java:338) at java.io.FilterOutputStream.flush(FilterOutputStream.java:140) at java.io.PrintStream.write(PrintStream.java:482) at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:221) at sun.nio.cs.StreamEncoder.implFlushBuffer(StreamEncoder.java:291) at sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:295) at sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:141) at java.io.OutputStreamWriter.flush(OutputStreamWriter.java:229) at org.apache.log4j.helpers.QuietWriter.flush(QuietWriter.java:59) at org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:324) at org.apache.log4j.WriterAppender.append(WriterAppender.java:162) at org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251) at org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66) at org.apache.log4j.Category.callAppenders(Category.java:206) at org.apache.log4j.Category.forcedLog(Category.java:391) at org.apache.log4j.Category.log(Category.java:856) at org.slf4j.impl.Log4jLoggerAdapter.error(Log4jLoggerAdapter.java:576) at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:142) at java.lang.Thread.run(Thread.java:748) at __randomizedtesting.SeedInfo.seed([82B128B6ED4F7220]:0) FAILED: junit.framework.TestSuite.org.apache.solr.cloud.ZkControllerTest Error Message: There are still zombie threads that couldn't be terminated: 1) Thread[id=14946, name=OverseerAutoScalingTriggerThread-72262182530908162-127.0.0.1:8983_solr-n_0000000001, state=RUNNABLE, group=Overseer autoscaling triggers] at java.security.AccessController.doPrivileged(Native Method) at com.carrotsearch.ant.tasks.junit4.events.Serializer.flushQueue(Serializer.java:114) at com.carrotsearch.ant.tasks.junit4.events.Serializer.serialize(Serializer.java:99) at com.carrotsearch.ant.tasks.junit4.slave.SlaveMain$3$2.write(SlaveMain.java:472) at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140) at java.io.PrintStream.flush(PrintStream.java:338) at java.io.FilterOutputStream.flush(FilterOutputStream.java:140) at java.io.PrintStream.write(PrintStream.java:482) at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:221) at sun.nio.cs.StreamEncoder.implFlushBuffer(StreamEncoder.java:291) at sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:295) at sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:141) at java.io.OutputStreamWriter.flush(OutputStreamWriter.java:229) at org.apache.log4j.helpers.QuietWriter.flush(QuietWriter.java:59) at org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:324) at org.apache.log4j.WriterAppender.append(WriterAppender.java:162) at org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251) at org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66) at org.apache.log4j.Category.callAppenders(Category.java:206) at org.apache.log4j.Category.forcedLog(Category.java:391) at org.apache.log4j.Category.log(Category.java:856) at org.slf4j.impl.Log4jLoggerAdapter.error(Log4jLoggerAdapter.java:576) at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:142) at java.lang.Thread.run(Thread.java:748) Stack Trace: com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie threads that couldn't be terminated: 1) Thread[id=14946, name=OverseerAutoScalingTriggerThread-72262182530908162-127.0.0.1:8983_solr-n_0000000001, state=RUNNABLE, group=Overseer autoscaling triggers] at java.security.AccessController.doPrivileged(Native Method) at com.carrotsearch.ant.tasks.junit4.events.Serializer.flushQueue(Serializer.java:114) at com.carrotsearch.ant.tasks.junit4.events.Serializer.serialize(Serializer.java:99) at com.carrotsearch.ant.tasks.junit4.slave.SlaveMain$3$2.write(SlaveMain.java:472) at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140) at java.io.PrintStream.flush(PrintStream.java:338) at java.io.FilterOutputStream.flush(FilterOutputStream.java:140) at java.io.PrintStream.write(PrintStream.java:482) at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:221) at sun.nio.cs.StreamEncoder.implFlushBuffer(StreamEncoder.java:291) at sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:295) at sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:141) at java.io.OutputStreamWriter.flush(OutputStreamWriter.java:229) at org.apache.log4j.helpers.QuietWriter.flush(QuietWriter.java:59) at org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:324) at org.apache.log4j.WriterAppender.append(WriterAppender.java:162) at org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251) at org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66) at org.apache.log4j.Category.callAppenders(Category.java:206) at org.apache.log4j.Category.forcedLog(Category.java:391) at org.apache.log4j.Category.log(Category.java:856) at org.slf4j.impl.Log4jLoggerAdapter.error(Log4jLoggerAdapter.java:576) at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:142) at java.lang.Thread.run(Thread.java:748) at __randomizedtesting.SeedInfo.seed([82B128B6ED4F7220]:0) FAILED: org.apache.solr.logging.TestLogWatcher.testLog4jWatcher Error Message: expected:<47> but was:<1> Stack Trace: java.lang.AssertionError: expected:<47> but was:<1> at org.junit.Assert.fail(Assert.java:93) at org.junit.Assert.failNotEquals(Assert.java:647) at org.junit.Assert.assertEquals(Assert.java:128) at org.junit.Assert.assertEquals(Assert.java:472) at org.junit.Assert.assertEquals(Assert.java:456) at org.apache.solr.logging.TestLogWatcher.testLog4jWatcher(TestLogWatcher.java:52) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:45) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:42) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:263) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:68) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:47) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:231) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:60) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:229) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:50) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:222) at org.junit.runners.ParentRunner.run(ParentRunner.java:300) at com.carrotsearch.ant.tasks.junit4.slave.SlaveMain.execute(SlaveMain.java:251) at com.carrotsearch.ant.tasks.junit4.slave.SlaveMain.main(SlaveMain.java:368) at com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe.main(SlaveMainSafe.java:13) FAILED: junit.framework.TestSuite.org.apache.solr.search.join.BlockJoinFacetDistribTest Error Message: Error from server at http://127.0.0.1:32837/solr: KeeperErrorCode = NoNode for /overseer/collection-queue-work/qnr-0000000000 Stack Trace: org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at http://127.0.0.1:32837/solr: KeeperErrorCode = NoNode for /overseer/collection-queue-work/qnr-0000000000 at __randomizedtesting.SeedInfo.seed([82B128B6ED4F7220]:0) at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:643) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244) at org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:483) at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:413) at org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1105) at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:885) at org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:818) at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194) at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:211) at org.apache.solr.search.join.BlockJoinFacetDistribTest.setupCluster(BlockJoinFacetDistribTest.java:69) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:874) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.lang.Thread.run(Thread.java:748) FAILED: org.apache.solr.cloud.hdfs.StressHdfsTest.test Error Message: Error from server at http://127.0.0.1:42805/n_k: collection already exists: delete_data_dir Stack Trace: org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at http://127.0.0.1:42805/n_k: collection already exists: delete_data_dir at __randomizedtesting.SeedInfo.seed([82B128B6ED4F7220:AE5176C43B31FD8]:0) at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:643) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244) at org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1219) at org.apache.solr.cloud.BasicDistributedZkTest.createCollection(BasicDistributedZkTest.java:643) at org.apache.solr.cloud.AbstractFullDistribZkTestBase.createCollection(AbstractFullDistribZkTestBase.java:1612) at org.apache.solr.cloud.hdfs.StressHdfsTest.createAndDeleteCollection(StressHdfsTest.java:159) at org.apache.solr.cloud.hdfs.StressHdfsTest.test(StressHdfsTest.java:105) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:993) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:968) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.lang.Thread.run(Thread.java:748) Build Log: [...truncated 12766 lines...] [junit4] Suite: org.apache.solr.cloud.hdfs.StressHdfsTest [junit4] 2> Creating dataDir: /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/init-core-data-001 [junit4] 2> 1470667 WARN (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=6 numCloses=6 [junit4] 2> 1470668 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=false [junit4] 2> 1470669 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via: @org.apache.solr.SolrTestCaseJ4$SuppressSSL(bugUrl=https://issues.apache.org/jira/browse/SOLR-5776) [junit4] 2> 1470669 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom [junit4] 2> 1470669 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /n_k/ [junit4] 1> Formatting using clusterid: testClusterID [junit4] 2> 1471032 WARN (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties [junit4] 2> 1471097 WARN (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j [junit4] 2> 1471132 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.m.log jetty-6.1.26 [junit4] 2> 1471324 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs to ./temp/Jetty_lucene2.us.west_apache_org_34186_hdfs____6wokm0/webapp [junit4] 2> 1472439 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.m.log Started HttpServer2$selectchannelconnectorwithsafestar...@lucene2-us-west.apache.org:34186 [junit4] 2> 1473245 WARN (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j [junit4] 2> 1473295 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.m.log jetty-6.1.26 [junit4] 2> 1473570 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_35844_datanode____o44mwt/webapp [junit4] 2> 1474416 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:35844 [junit4] 2> 1474744 WARN (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j [junit4] 2> 1474745 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.m.log jetty-6.1.26 [junit4] 2> 1474869 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_37729_datanode____.p8rxvj/webapp [junit4] 2> 1475330 ERROR (DataNode: [[[DISK]file:/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/tempDir-001/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/tempDir-001/hdfsBaseDir/data/data2/]] heartbeating to lucene2-us-west.apache.org/127.0.0.1:42713) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000 [junit4] 2> 1475487 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xb15e01689c0dc: from storage DS-2597eb82-7aa7-45f9-aab1-a0feeae495f1 node DatanodeRegistration(127.0.0.1:43644, datanodeUuid=09eb17c7-302d-4e16-94a9-508f597e9270, infoPort=40139, infoSecurePort=0, ipcPort=40517, storageInfo=lv=-56;cid=testClusterID;nsid=1502172852;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs [junit4] 2> 1475487 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xb15e01689c0dc: from storage DS-a47493d7-e999-49ee-963c-9b9590939dbc node DatanodeRegistration(127.0.0.1:43644, datanodeUuid=09eb17c7-302d-4e16-94a9-508f597e9270, infoPort=40139, infoSecurePort=0, ipcPort=40517, storageInfo=lv=-56;cid=testClusterID;nsid=1502172852;c=0), blocks: 0, hasStaleStorage: false, processing time: 1 msecs [junit4] 2> 1475860 INFO (SUITE-StressHdfsTest-seed#[82B128B6ED4F7220]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:37729 [junit4] 2> 1476633 ERROR (DataNode: [[[DISK]file:/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/tempDir-001/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/tempDir-001/hdfsBaseDir/data/data4/]] heartbeating to lucene2-us-west.apache.org/127.0.0.1:42713) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000 [junit4] 2> 1476725 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xb15e060559e8d: from storage DS-66691d3e-e180-4130-a0eb-add9330edf48 node DatanodeRegistration(127.0.0.1:39744, datanodeUuid=941394f7-1afb-4c24-90f4-ede6f2db7745, infoPort=37179, infoSecurePort=0, ipcPort=33432, storageInfo=lv=-56;cid=testClusterID;nsid=1502172852;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs [junit4] 2> 1476725 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xb15e060559e8d: from storage DS-f2951b19-39c3-4e93-a666-abe0e3ce74a4 node DatanodeRegistration(127.0.0.1:39744, datanodeUuid=941394f7-1afb-4c24-90f4-ede6f2db7745, infoPort=37179, infoSecurePort=0, ipcPort=33432, storageInfo=lv=-56;cid=testClusterID;nsid=1502172852;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs [junit4] 2> 1477395 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER [junit4] 2> 1477396 INFO (Thread-1358) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0 [junit4] 2> 1477396 INFO (Thread-1358) [ ] o.a.s.c.ZkTestServer Starting server [junit4] 2> 1477402 ERROR (Thread-1358) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes [junit4] 2> 1477496 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.ZkTestServer start zk server on port:44358 [junit4] 2> 1477532 INFO (zkConnectionManagerCallback-1132-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1477588 INFO (zkConnectionManagerCallback-1134-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1477631 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml [junit4] 2> 1477634 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/core/src/test-files/solr/collection1/conf/schema.xml to /configs/conf1/schema.xml [junit4] 2> 1477664 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml to /configs/conf1/solrconfig.snippet.randomindexconfig.xml [junit4] 2> 1477666 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/core/src/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt [junit4] 2> 1477680 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/core/src/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt [junit4] 2> 1477680 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/core/src/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml [junit4] 2> 1477681 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml to /configs/conf1/enumsConfig.xml [junit4] 2> 1477682 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json [junit4] 2> 1477692 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt [junit4] 2> 1477695 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt [junit4] 2> 1477708 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/core/src/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt [junit4] 2> 1477710 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100b9fbad500001, likely client has closed socket [junit4] 2> 1477711 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractFullDistribZkTestBase Will use TLOG replicas unless explicitly asked otherwise [junit4] 2> 1478266 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.Server jetty-9.4.8.v20171121, build timestamp: 2017-11-21T15:27:37-06:00, git hash: 82b8fb23f757335bb3329d540ce37a2a2615f0a8 [junit4] 2> 1478267 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0 [junit4] 2> 1478267 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.session No SessionScavenger set, using defaults [junit4] 2> 1478267 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.session Scavenging every 600000ms [junit4] 2> 1478268 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@424ca966{/n_k,null,AVAILABLE} [junit4] 2> 1478280 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@862fe57{HTTP/1.1,[http/1.1]}{127.0.0.1:42091} [junit4] 2> 1478280 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.Server Started @1489787ms [junit4] 2> 1478281 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=hdfs://lucene2-us-west.apache.org:42713/hdfs__lucene2-us-west.apache.org_42713__home_jenkins_jenkins-slave_workspace_Lucene-Solr-Tests-master_solr_build_solr-core_test_J0_temp_solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001_tempDir-002_control_data, replicaType=NRT, hostContext=/n_k, hostPort=42091, coreRootDirectory=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/../../../../../../../../../../home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/control-001/cores} [junit4] 2> 1478281 ERROR (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete. [junit4] 2> 1478281 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr? version 8.0.0 [junit4] 2> 1478281 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null [junit4] 2> 1478281 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null [junit4] 2> 1478281 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-03-11T15:41:18.811Z [junit4] 2> 1478289 INFO (zkConnectionManagerCallback-1136-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1478289 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper) [junit4] 2> 1478290 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/control-001/solr.xml [junit4] 2> 1478310 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored [junit4] 2> 1478310 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored [junit4] 2> 1478311 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@5990b685, but no JMX reporters were configured - adding default JMX reporter. [junit4] 2> 1478346 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44358/solr [junit4] 2> 1478393 INFO (zkConnectionManagerCallback-1140-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1478459 INFO (zkConnectionManagerCallback-1142-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1478982 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42091_n_k ] o.a.s.c.Overseer Overseer (id=null) closing [junit4] 2> 1478983 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42091_n_k ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:42091_n_k [junit4] 2> 1478983 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42091_n_k ] o.a.s.c.Overseer Overseer (id=72262084633559044-127.0.0.1:42091_n_k-n_0000000000) starting [junit4] 2> 1479008 INFO (zkConnectionManagerCallback-1147-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1479010 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42091_n_k ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44358/solr ready [junit4] 2> 1479010 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42091_n_k ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:42091_n_k [junit4] 2> 1479068 INFO (zkCallback-1146-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 1479095 INFO (OverseerStateUpdate-72262084633559044-127.0.0.1:42091_n_k-n_0000000000) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 1479319 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42091_n_k ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5990b685 [junit4] 2> 1479392 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42091_n_k ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5990b685 [junit4] 2> 1479392 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42091_n_k ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5990b685 [junit4] 2> 1479410 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42091_n_k ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/../../../../../../../../../../home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/control-001/cores [junit4] 2> 1479507 INFO (zkConnectionManagerCallback-1151-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1479508 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 1479509 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44358/solr ready [junit4] 2> 1479511 INFO (qtp779735766-5077) [n:127.0.0.1:42091_n_k ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params replicationFactor=1&collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:42091_n_k&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 1479557 INFO (OverseerThreadFactory-1636-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.a.c.CreateCollectionCmd Create collection control_collection [junit4] 2> 1479690 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT [junit4] 2> 1479690 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores [junit4] 2> 1479906 INFO (zkCallback-1141-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/control_collection/state.json] for collection [control_collection] has occurred - updating... (live nodes size: [1]) [junit4] 2> 1480970 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0 [junit4] 2> 1481065 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test [junit4] 2> 1485620 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema test/1.0 with uniqueid field id [junit4] 2> 1485742 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1' using configuration from collection control_collection, trusted=true [junit4] 2> 1485743 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.control_collection.shard1.replica_n1' (registry 'solr.core.control_collection.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5990b685 [junit4] 2> 1485743 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://lucene2-us-west.apache.org:42713/solr_hdfs_home [junit4] 2> 1485743 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled [junit4] 2> 1485743 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder [junit4] 2> 1485743 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/control-001/cores/control_collection_shard1_replica_n1], dataDir=[hdfs://lucene2-us-west.apache.org:42713/solr_hdfs_home/control_collection/core_node2/data/] [junit4] 2> 1485797 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://lucene2-us-west.apache.org:42713/solr_hdfs_home/control_collection/core_node2/data/snapshot_metadata [junit4] 2> 1485885 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct memory allocation set to [true] [junit4] 2> 1485886 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of [8388608] will allocate [1] slabs and use ~[8388608] bytes [junit4] 2> 1485886 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache [junit4] 2> 1486040 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.b.BlockDirectory Block cache on write is disabled [junit4] 2> 1486068 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://lucene2-us-west.apache.org:42713/solr_hdfs_home/control_collection/core_node2/data [junit4] 2> 1486176 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://lucene2-us-west.apache.org:42713/solr_hdfs_home/control_collection/core_node2/data/index [junit4] 2> 1486211 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct memory allocation set to [true] [junit4] 2> 1486211 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of [8388608] will allocate [1] slabs and use ~[8388608] bytes [junit4] 2> 1486211 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache [junit4] 2> 1488590 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.b.BlockDirectory Block cache on write is disabled [junit4] 2> 1488591 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=34, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=0.6650390625, noCFSRatio=0.0] [junit4] 2> 1488906 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:43644 is added to blk_1073741825_1001{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-f2951b19-39c3-4e93-a666-abe0e3ce74a4:NORMAL:127.0.0.1:39744|RBW], ReplicaUC[[DISK]DS-2597eb82-7aa7-45f9-aab1-a0feeae495f1:NORMAL:127.0.0.1:43644|RBW]]} size 69 [junit4] 2> 1488906 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:39744 is added to blk_1073741825_1001 size 69 [junit4] 2> 1489328 WARN (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}} [junit4] 2> 1489581 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog [junit4] 2> 1489581 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536 [junit4] 2> 1489581 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=2 [junit4] 2> 1489635 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled [junit4] 2> 1489635 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled [junit4] 2> 1489646 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=17, maxMergeAtOnceExplicit=44, maxMergedSegmentMB=95.1953125, floorSegmentMB=0.8076171875, forceMergeDeletesPctAllowed=20.97836991332744, segmentsPerTier=22.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0 [junit4] 2> 1489707 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@3a38451c[control_collection_shard1_replica_n1] main] [junit4] 2> 1489725 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1 [junit4] 2> 1489725 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1 [junit4] 2> 1489726 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms. [junit4] 2> 1489729 INFO (searcherExecutor-1639-thread-1-processing-n:127.0.0.1:42091_n_k x:control_collection_shard1_replica_n1 s:shard1 c:control_collection) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [control_collection_shard1_replica_n1] Registered new searcher Searcher@3a38451c[control_collection_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())} [junit4] 2> 1489729 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1594656439936221184 [junit4] 2> 1489757 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update terms at /collections/control_collection/terms/shard1 to Terms{values={core_node2=0}, version=0} [junit4] 2> 1489772 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue. [junit4] 2> 1489772 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync [junit4] 2> 1489772 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:42091/n_k/control_collection_shard1_replica_n1/ [junit4] 2> 1489772 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me [junit4] 2> 1489772 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy http://127.0.0.1:42091/n_k/control_collection_shard1_replica_n1/ has no replicas [junit4] 2> 1489772 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR [junit4] 2> 1489775 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:42091/n_k/control_collection_shard1_replica_n1/ shard1 [junit4] 2> 1489784 INFO (zkCallback-1141-thread-2-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/control_collection/state.json] for collection [control_collection] has occurred - updating... (live nodes size: [1]) [junit4] 2> 1489834 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary [junit4] 2> 1489836 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=10146 [junit4] 2> 1489838 INFO (qtp779735766-5077) [n:127.0.0.1:42091_n_k ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas [junit4] 2> 1489936 INFO (zkCallback-1141-thread-2-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/control_collection/state.json] for collection [control_collection] has occurred - updating... (live nodes size: [1]) [junit4] 2> 1490595 INFO (OverseerCollectionConfigSetProcessor-72262084633559044-127.0.0.1:42091_n_k-n_0000000000) [n:127.0.0.1:42091_n_k ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper [junit4] 2> 1490839 INFO (qtp779735766-5077) [n:127.0.0.1:42091_n_k ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={replicationFactor=1&collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:42091_n_k&wt=javabin&version=2} status=0 QTime=11327 [junit4] 2> 1490867 INFO (zkConnectionManagerCallback-1155-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1490868 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 1490884 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44358/solr ready [junit4] 2> 1490884 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection loss:false [junit4] 2> 1490885 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params replicationFactor=1&collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 1490909 INFO (OverseerThreadFactory-1636-thread-2-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.a.c.CreateCollectionCmd Create collection collection1 [junit4] 2> 1490909 WARN (OverseerThreadFactory-1636-thread-2-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.a.c.CreateCollectionCmd It is unusual to create a collection (collection1) without cores. [junit4] 2> 1491117 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas [junit4] 2> 1491142 INFO (qtp779735766-5081) [n:127.0.0.1:42091_n_k ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={replicationFactor=1&collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=2&wt=javabin&version=2} status=0 QTime=256 [junit4] 2> 1492042 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/shard-1-001 of type TLOG [junit4] 2> 1492043 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.Server jetty-9.4.8.v20171121, build timestamp: 2017-11-21T15:27:37-06:00, git hash: 82b8fb23f757335bb3329d540ce37a2a2615f0a8 [junit4] 2> 1492044 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0 [junit4] 2> 1492044 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.session No SessionScavenger set, using defaults [junit4] 2> 1492044 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.session Scavenging every 660000ms [junit4] 2> 1492044 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@74bfce94{/n_k,null,AVAILABLE} [junit4] 2> 1492057 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@7ac1decb{HTTP/1.1,[http/1.1]}{127.0.0.1:42805} [junit4] 2> 1492057 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.e.j.s.Server Started @1503564ms [junit4] 2> 1492057 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=hdfs://lucene2-us-west.apache.org:42713/hdfs__lucene2-us-west.apache.org_42713__home_jenkins_jenkins-slave_workspace_Lucene-Solr-Tests-master_solr_build_solr-core_test_J0_temp_solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001_tempDir-002_jetty1, replicaType=TLOG, solrconfig=solrconfig.xml, hostContext=/n_k, hostPort=42805, coreRootDirectory=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/shard-1-001/cores} [junit4] 2> 1492058 ERROR (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete. [junit4] 2> 1492058 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr? version 8.0.0 [junit4] 2> 1492058 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null [junit4] 2> 1492059 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null [junit4] 2> 1492059 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-03-11T15:41:32.589Z [junit4] 2> 1492101 INFO (zkConnectionManagerCallback-1157-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1492107 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper) [junit4] 2> 1492107 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/shard-1-001/solr.xml [junit4] 2> 1492127 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored [junit4] 2> 1492127 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored [junit4] 2> 1492129 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@5990b685, but no JMX reporters were configured - adding default JMX reporter. [junit4] 2> 1492147 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44358/solr [junit4] 2> 1492191 INFO (zkConnectionManagerCallback-1161-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1492212 INFO (zkConnectionManagerCallback-1163-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1492225 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 1492226 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42805_n_k ] o.a.s.c.Overseer Overseer (id=null) closing [junit4] 2> 1492241 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42805_n_k ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:42805_n_k [junit4] 2> 1492253 INFO (zkCallback-1141-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 1492265 INFO (zkCallback-1154-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 1492276 INFO (zkCallback-1146-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 1492369 INFO (zkCallback-1162-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 1492702 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42805_n_k ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5990b685 [junit4] 2> 1492761 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42805_n_k ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5990b685 [junit4] 2> 1492761 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42805_n_k ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5990b685 [junit4] 2> 1492763 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42805_n_k ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/shard-1-001/cores [junit4] 2> 1492832 INFO (zkConnectionManagerCallback-1168-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1492833 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2) [junit4] 2> 1492834 INFO (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]) [n:127.0.0.1:42805_n_k ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44358/solr ready [junit4] 2> 1493088 INFO (OverseerCollectionConfigSetProcessor-72262084633559044-127.0.0.1:42091_n_k-n_0000000000) [n:127.0.0.1:42091_n_k ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000002 doesn't exist. Requestor may have disconnected from ZooKeeper [junit4] 2> 1493101 INFO (qtp779735766-5079) [n:127.0.0.1:42091_n_k ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:42805_n_k&action=ADDREPLICA&collection=collection1&shard=shard1&type=TLOG&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 1493200 INFO (OverseerThreadFactory-1636-thread-3-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:42805_n_k for creating new replica [junit4] 2> 1493228 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_t21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=TLOG [junit4] 2> 1493228 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores [junit4] 2> 1493389 INFO (zkCallback-1162-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2]) [junit4] 2> 1494321 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0 [junit4] 2> 1494431 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.s.IndexSchema [collection1_shard1_replica_t21] Schema name=test [junit4] 2> 1495132 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.s.IndexSchema Loaded schema test/1.0 with uniqueid field id [junit4] 2> 1495268 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard1_replica_t21' using configuration from collection collection1, trusted=true [junit4] 2> 1495301 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard1.replica_t21' (registry 'solr.core.collection1.shard1.replica_t21') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5990b685 [junit4] 2> 1495302 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://lucene2-us-west.apache.org:42713/solr_hdfs_home [junit4] 2> 1495302 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled [junit4] 2> 1495302 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder [junit4] 2> 1495302 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SolrCore [[collection1_shard1_replica_t21] ] Opening new SolrCore at [/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-master/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_82B128B6ED4F7220-001/shard-1-001/cores/collection1_shard1_replica_t21], dataDir=[hdfs://lucene2-us-west.apache.org:42713/solr_hdfs_home/collection1/core_node22/data/] [junit4] 2> 1495303 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://lucene2-us-west.apache.org:42713/solr_hdfs_home/collection1/core_node22/data/snapshot_metadata [junit4] 2> 1495397 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct memory allocation set to [true] [junit4] 2> 1495397 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of [8388608] will allocate [1] slabs and use ~[8388608] bytes [junit4] 2> 1495397 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache [junit4] 2> 1495543 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.s.b.BlockDirectory Block cache on write is disabled [junit4] 2> 1495544 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://lucene2-us-west.apache.org:42713/solr_hdfs_home/collection1/core_node22/data [junit4] 2> 1495695 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://lucene2-us-west.apache.org:42713/solr_hdfs_home/collection1/core_node22/data/index [junit4] 2> 1495718 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct memory allocation set to [true] [junit4] 2> 1495718 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of [8388608] will allocate [1] slabs and use ~[8388608] bytes [junit4] 2> 1495718 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache [junit4] 2> 1503450 WARN (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]-SendThread(127.0.0.1:44358)) [ ] o.a.z.ClientCnxn Client session timed out, have not heard from server in 7836ms for sessionid 0x100b9fbad500007 [junit4] 2> 1503465 WARN (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]-SendThread(127.0.0.1:44358)) [n:127.0.0.1:42091_n_k ] o.a.z.ClientCnxn Client session timed out, have not heard from server in 10113ms for sessionid 0x100b9fbad500005 [junit4] 2> 1503468 WARN (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]-SendThread(127.0.0.1:44358)) [n:127.0.0.1:42805_n_k ] o.a.z.ClientCnxn Client session timed out, have not heard from server in 10634ms for sessionid 0x100b9fbad50000b [junit4] 2> 1503466 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100b9fbad500007, likely client has closed socket [junit4] 2> 1503469 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100b9fbad500005, likely client has closed socket [junit4] 2> 1503518 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100b9fbad50000b, likely client has closed socket [junit4] 2> 1503592 WARN (zkConnectionManagerCallback-1147-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@7fd3a644 name: ZooKeeperConnection Watcher:127.0.0.1:44358/solr got event WatchedEvent state:Disconnected type:None path:null path: null type: None [junit4] 2> 1503592 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.s.b.BlockDirectory Block cache on write is disabled [junit4] 2> 1503592 WARN (zkConnectionManagerCallback-1147-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ConnectionManager zkClient has disconnected [junit4] 2> 1503612 WARN (zkConnectionManagerCallback-1155-thread-1) [ ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@285fec09 name: ZooKeeperConnection Watcher:127.0.0.1:44358/solr got event WatchedEvent state:Disconnected type:None path:null path: null type: None [junit4] 2> 1503612 WARN (zkConnectionManagerCallback-1155-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has disconnected [junit4] 2> 1503628 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=34, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=0.6650390625, noCFSRatio=0.0] [junit4] 2> 1503664 WARN (zkConnectionManagerCallback-1168-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@37f90046 name: ZooKeeperConnection Watcher:127.0.0.1:44358/solr got event WatchedEvent state:Disconnected type:None path:null path: null type: None [junit4] 2> 1503664 WARN (zkConnectionManagerCallback-1168-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ConnectionManager zkClient has disconnected [junit4] 2> 1503936 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:39744 is added to blk_1073741826_1002{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-2597eb82-7aa7-45f9-aab1-a0feeae495f1:NORMAL:127.0.0.1:43644|RBW], ReplicaUC[[DISK]DS-f2951b19-39c3-4e93-a666-abe0e3ce74a4:NORMAL:127.0.0.1:39744|FINALIZED]]} size 0 [junit4] 2> 1503989 INFO (Block report processor) [ ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:43644 is added to blk_1073741826_1002 size 69 [junit4] 2> 1504104 WARN (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}} [junit4] 2> 1504578 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog [junit4] 2> 1504578 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536 [junit4] 2> 1504578 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=2 [junit4] 2> 1504695 INFO (zkConnectionManagerCallback-1155-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1504732 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.CommitTracker Hard AutoCommit: disabled [junit4] 2> 1504732 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.CommitTracker Soft AutoCommit: disabled [junit4] 2> 1504752 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=17, maxMergeAtOnceExplicit=44, maxMergedSegmentMB=95.1953125, floorSegmentMB=0.8076171875, forceMergeDeletesPctAllowed=20.97836991332744, segmentsPerTier=22.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0 [junit4] 2> 1504994 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.s.SolrIndexSearcher Opening [Searcher@5e37626a[collection1_shard1_replica_t21] main] [junit4] 2> 1505009 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1 [junit4] 2> 1505009 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1 [junit4] 2> 1505010 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms. [junit4] 2> 1505011 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1594656455960559616 [junit4] 2> 1505132 INFO (searcherExecutor-1650-thread-1-processing-n:127.0.0.1:42805_n_k x:collection1_shard1_replica_t21 s:shard1 c:collection1) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SolrCore [collection1_shard1_replica_t21] Registered new searcher Searcher@5e37626a[collection1_shard1_replica_t21] main{ExitableDirectoryReader(UninvertingDirectoryReader())} [junit4] 2> 1505172 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ZkShardTerms Successful update terms at /collections/collection1/terms/shard1 to Terms{values={core_node22=0}, version=0} [junit4] 2> 1505228 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue. [junit4] 2> 1505228 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync [junit4] 2> 1505228 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:42805/n_k/collection1_shard1_replica_t21/ [junit4] 2> 1505228 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me [junit4] 2> 1505228 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SyncStrategy http://127.0.0.1:42805/n_k/collection1_shard1_replica_t21/ has no replicas [junit4] 2> 1505228 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR [junit4] 2> 1505228 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ZkController collection1_shard1_replica_t21 stopping background replication from leader [junit4] 2> 1505238 WARN (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]-SendThread(127.0.0.1:44358)) [n:127.0.0.1:42805_n_k ] o.a.z.ClientCnxn Unable to reconnect to ZooKeeper service, session 0x100b9fbad50000b has expired [junit4] 2> 1505239 WARN (zkConnectionManagerCallback-1168-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@37f90046 name: ZooKeeperConnection Watcher:127.0.0.1:44358/solr got event WatchedEvent state:Expired type:None path:null path: null type: None [junit4] 2> 1505239 WARN (zkConnectionManagerCallback-1168-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ConnectionManager Our previous ZooKeeper session was expired. Attempting to reconnect to recover relationship with ZooKeeper... [junit4] 2> 1505240 WARN (zkConnectionManagerCallback-1168-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.DefaultConnectionStrategy Connection expired - starting a new one... [junit4] 2> 1505263 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:42805/n_k/collection1_shard1_replica_t21/ shard1 [junit4] 2> 1505277 INFO (zkConnectionManagerCallback-1168-thread-1-processing-n:127.0.0.1:42805_n_k-EventThread) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1505277 INFO (zkConnectionManagerCallback-1168-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ConnectionManager Connection with ZooKeeper reestablished. [junit4] 2> 1505279 INFO (zkCallback-1162-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2]) [junit4] 2> 1505528 INFO (zkConnectionManagerCallback-1168-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.DefaultConnectionStrategy Reconnected to ZooKeeper [junit4] 2> 1505529 INFO (zkConnectionManagerCallback-1168-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ConnectionManager zkClient Connected:true [junit4] 2> 1505531 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ZkController I am the leader, no recovery necessary [junit4] 2> 1505533 WARN (TEST-StressHdfsTest.test-seed#[82B128B6ED4F7220]-SendThread(127.0.0.1:44358)) [n:127.0.0.1:42091_n_k ] o.a.z.ClientCnxn Unable to reconnect to ZooKeeper service, session 0x100b9fbad500005 has expired [junit4] 2> 1505533 WARN (zkConnectionManagerCallback-1147-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@7fd3a644 name: ZooKeeperConnection Watcher:127.0.0.1:44358/solr got event WatchedEvent state:Expired type:None path:null path: null type: None [junit4] 2> 1505533 WARN (zkConnectionManagerCallback-1147-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ConnectionManager Our previous ZooKeeper session was expired. Attempting to reconnect to recover relationship with ZooKeeper... [junit4] 2> 1505533 WARN (zkConnectionManagerCallback-1147-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.DefaultConnectionStrategy Connection expired - starting a new one... [junit4] 2> 1505549 INFO (qtp1713953838-5139) [n:127.0.0.1:42805_n_k c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_t21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=TLOG} status=0 QTime=12321 [junit4] 2> 1505592 INFO (zkConnectionManagerCallback-1147-thread-1-processing-n:127.0.0.1:42091_n_k-EventThread) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 1505592 INFO (zkConnectionManagerCallback-1147-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ConnectionManager Connection with ZooKeeper reestablished. [junit4] 2> 1505873 INFO (zkConnectionManagerCallback-1147-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.DefaultConnectionStrategy Reconnected to ZooKeeper [junit4] 2> 1505873 INFO (zkConnectionManagerCallback-1147-thread-1-processing-n:127.0.0.1:42091_n_k) [n:127.0.0.1:42091_n_k ] o.a.s.c.c.ConnectionManager zkClient Connected:true [junit4] 2> 1505886 INFO (zkCallback-1162-thread-1-processing-n:127.0.0.1:42805_n_k) [n:127.0.0.1:42805_n_k ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2]) [junit4] 2> 1505900 INFO (qtp779735766-5079) [n:127.0.0.1:42091_n_k ] o.a.s.s.HttpSolrCall [admin] webapp=null pat [...truncated too long message...] scaling.json [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:183) [junit4] 2> at org.apache.solr.client.solrj.cloud.autoscaling.DistribStateManager.getAutoScalingConfig(DistribStateManager.java:78) [junit4] 2> at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:127) [junit4] 2> at java.lang.Thread.run(Thread.java:748) [junit4] 2> Caused by: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /autoscaling.json [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:130) [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:54) [junit4] 2> at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:1215) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.lambda$getData$5(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.getData(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:176) [junit4] 2> ... 3 more [junit4] 2> 3312781 ERROR (OverseerAutoScalingTriggerThread-72262182530908162-127.0.0.1:8983_solr-n_0000000001) [n:127.0.0.1:8983_solr ] o.a.s.c.a.OverseerTriggerThread A ZK error has occurred [junit4] 2> java.io.IOException: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /autoscaling.json [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:183) [junit4] 2> at org.apache.solr.client.solrj.cloud.autoscaling.DistribStateManager.getAutoScalingConfig(DistribStateManager.java:78) [junit4] 2> at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:127) [junit4] 2> at java.lang.Thread.run(Thread.java:748) [junit4] 2> Caused by: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /autoscaling.json [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:130) [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:54) [junit4] 2> at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:1215) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.lambda$getData$5(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.getData(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:176) [junit4] 2> ... 3 more [junit4] 2> 3312782 ERROR (OverseerAutoScalingTriggerThread-72262182530908162-127.0.0.1:8983_solr-n_0000000001) [n:127.0.0.1:8983_solr ] o.a.s.c.a.OverseerTriggerThread A ZK error has occurred [junit4] 2> java.io.IOException: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /autoscaling.json [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:183) [junit4] 2> at org.apache.solr.client.solrj.cloud.autoscaling.DistribStateManager.getAutoScalingConfig(DistribStateManager.java:78) [junit4] 2> at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:127) [junit4] 2> at java.lang.Thread.run(Thread.java:748) [junit4] 2> Caused by: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /autoscaling.json [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:130) [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:54) [junit4] 2> at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:1215) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.lambda$getData$5(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.getData(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:176) [junit4] 2> ... 3 more [junit4] 2> 3312782 ERROR (OverseerAutoScalingTriggerThread-72262182530908162-127.0.0.1:8983_solr-n_0000000001) [n:127.0.0.1:8983_solr ] o.a.s.c.a.OverseerTriggerThread A ZK error has occurred [junit4] 2> java.io.IOException: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /autoscaling.json [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:183) [junit4] 2> at org.apache.solr.client.solrj.cloud.autoscaling.DistribStateManager.getAutoScalingConfig(DistribStateManager.java:78) [junit4] 2> at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:127) [junit4] 2> at java.lang.Thread.run(Thread.java:748) [junit4] 2> Caused by: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /autoscaling.json [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:130) [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:54) [junit4] 2> at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:1215) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.lambda$getData$5(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.getData(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:176) [junit4] 2> ... 3 more [junit4] 2> 3312782 ERROR (OverseerAutoScalingTriggerThread-72262182530908162-127.0.0.1:8983_solr-n_0000000001) [n:127.0.0.1:8983_solr ] o.a.s.c.a.OverseerTriggerThread A ZK error has occurred [junit4] 2> java.io.IOException: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /autoscaling.json [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:183) [junit4] 2> at org.apache.solr.client.solrj.cloud.autoscaling.DistribStateManager.getAutoScalingConfig(DistribStateManager.java:78) [junit4] 2> at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:127) [junit4] 2> at java.lang.Thread.run(Thread.java:748) [junit4] 2> Caused by: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /autoscaling.json [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:130) [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:54) [junit4] 2> at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:1215) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.lambda$getData$5(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.getData(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:176) [junit4] 2> ... 3 more [junit4] 2> 3312782 ERROR (OverseerAutoScalingTriggerThread-72262182530908162-127.0.0.1:8983_solr-n_0000000001) [n:127.0.0.1:8983_solr ] o.a.s.c.a.OverseerTriggerThread A ZK error has occurred [junit4] 2> java.io.IOException: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /autoscaling.json [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:183) [junit4] 2> at org.apache.solr.client.solrj.cloud.autoscaling.DistribStateManager.getAutoScalingConfig(DistribStateManager.java:78) [junit4] 2> at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:127) [junit4] 2> at java.lang.Thread.run(Thread.java:748) [junit4] 2> Caused by: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /autoscaling.json [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:130) [junit4] 2> at org.apache.zookeeper.KeeperException.create(KeeperException.java:54) [junit4] 2> at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:1215) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.lambda$getData$5(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60) [junit4] 2> at org.apache.solr.common.cloud.SolrZkClient.getData(SolrZkClient.java:340) [junit4] 2> at org.apache.solr.client.solrj.impl.ZkDistribStateManager.getAutoScalingConfig(ZkDistribStateManager.java:176) [junit4] 2> ... 3 more [junit4] FAILURE 0.12s J2 | TestLogWatcher.testLog4jWatcher <<< [junit4] > Throwable #1: java.lang.AssertionError: expected:<47> but was:<1> [junit4] > at org.apache.solr.logging.TestLogWatcher.testLog4jWatcher(TestLogWatcher.java:52) [junit4] Completed [646/785 (4!)] on J2 in 0.12s, 1 test, 1 failure <<< FAILURES! [...truncated 49900 lines...]
--------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org