Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/2294/ Java: 64bit/jdk-10.0.1 -XX:+UseCompressedOops -XX:+UseG1GC
2 tests failed. FAILED: org.apache.solr.cloud.MoveReplicaHDFSTest.testFailedMove Error Message: No live SolrServers available to handle this request:[https://127.0.0.1:38555/solr/MoveReplicaHDFSTest_failed_coll_true, https://127.0.0.1:46435/solr/MoveReplicaHDFSTest_failed_coll_true] Stack Trace: org.apache.solr.client.solrj.SolrServerException: No live SolrServers available to handle this request:[https://127.0.0.1:38555/solr/MoveReplicaHDFSTest_failed_coll_true, https://127.0.0.1:46435/solr/MoveReplicaHDFSTest_failed_coll_true] at __randomizedtesting.SeedInfo.seed([EF501846732A343:A438D276D0E17693]:0) at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:462) at org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1106) at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:886) at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:993) at org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:819) at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194) at org.apache.solr.client.solrj.SolrClient.query(SolrClient.java:942) at org.apache.solr.cloud.MoveReplicaTest.testFailedMove(MoveReplicaTest.java:288) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:564) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.base/java.lang.Thread.run(Thread.java:844) Caused by: org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at https://127.0.0.1:46435/solr/MoveReplicaHDFSTest_failed_coll_true: no servers hosting shard: shard1 at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:643) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255) at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244) at org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:483) at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:413) ... 45 more FAILED: org.apache.solr.cloud.RecoveryAfterSoftCommitTest.test Error Message: Didn't see all replicas for shard shard1 in collection1 come up within 30000 ms! ClusterState: { "control_collection":{ "pullReplicas":"0", "replicationFactor":"1", "shards":{"shard1":{ "range":"80000000-7fffffff", "state":"active", "replicas":{"core_node2":{ "core":"control_collection_shard1_replica_n1", "base_url":"http://127.0.0.1:34913", "node_name":"127.0.0.1:34913_", "state":"active", "type":"NRT", "leader":"true"}}}}, "router":{"name":"compositeId"}, "maxShardsPerNode":"1", "autoAddReplicas":"false", "nrtReplicas":"1", "tlogReplicas":"0"}, "collection1":{ "pullReplicas":"0", "replicationFactor":"1", "shards":{"shard1":{ "range":"80000000-7fffffff", "state":"active", "replicas":{ "core_node22":{ "core":"collection1_shard1_replica_t21", "base_url":"http://127.0.0.1:32995", "node_name":"127.0.0.1:32995_", "state":"active", "type":"TLOG", "leader":"true"}, "core_node24":{ "core":"collection1_shard1_replica_t23", "base_url":"http://127.0.0.1:44299", "node_name":"127.0.0.1:44299_", "state":"recovering", "type":"TLOG"}}}}, "router":{"name":"compositeId"}, "maxShardsPerNode":"1", "autoAddReplicas":"false", "nrtReplicas":"1", "tlogReplicas":"0"}} Stack Trace: java.lang.AssertionError: Didn't see all replicas for shard shard1 in collection1 come up within 30000 ms! ClusterState: { "control_collection":{ "pullReplicas":"0", "replicationFactor":"1", "shards":{"shard1":{ "range":"80000000-7fffffff", "state":"active", "replicas":{"core_node2":{ "core":"control_collection_shard1_replica_n1", "base_url":"http://127.0.0.1:34913", "node_name":"127.0.0.1:34913_", "state":"active", "type":"NRT", "leader":"true"}}}}, "router":{"name":"compositeId"}, "maxShardsPerNode":"1", "autoAddReplicas":"false", "nrtReplicas":"1", "tlogReplicas":"0"}, "collection1":{ "pullReplicas":"0", "replicationFactor":"1", "shards":{"shard1":{ "range":"80000000-7fffffff", "state":"active", "replicas":{ "core_node22":{ "core":"collection1_shard1_replica_t21", "base_url":"http://127.0.0.1:32995", "node_name":"127.0.0.1:32995_", "state":"active", "type":"TLOG", "leader":"true"}, "core_node24":{ "core":"collection1_shard1_replica_t23", "base_url":"http://127.0.0.1:44299", "node_name":"127.0.0.1:44299_", "state":"recovering", "type":"TLOG"}}}}, "router":{"name":"compositeId"}, "maxShardsPerNode":"1", "autoAddReplicas":"false", "nrtReplicas":"1", "tlogReplicas":"0"}} at __randomizedtesting.SeedInfo.seed([EF501846732A343:86A13E5EC9CECEBB]:0) at org.junit.Assert.fail(Assert.java:93) at org.apache.solr.cloud.AbstractFullDistribZkTestBase.ensureAllReplicasAreActive(AbstractFullDistribZkTestBase.java:2005) at org.apache.solr.cloud.RecoveryAfterSoftCommitTest.test(RecoveryAfterSoftCommitTest.java:116) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:564) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:1008) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:983) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) at java.base/java.lang.Thread.run(Thread.java:844) Build Log: [...truncated 13138 lines...] [junit4] Suite: org.apache.solr.cloud.RecoveryAfterSoftCommitTest [junit4] 2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/init-core-data-001 [junit4] 2> 595127 WARN (SUITE-RecoveryAfterSoftCommitTest-seed#[EF501846732A343]-worker) [ ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=28 numCloses=28 [junit4] 2> 595127 INFO (SUITE-RecoveryAfterSoftCommitTest-seed#[EF501846732A343]-worker) [ ] o.a.s.SolrTestCaseJ4 Using TrieFields (NUMERIC_POINTS_SYSPROP=false) w/NUMERIC_DOCVALUES_SYSPROP=false [junit4] 2> 595128 INFO (SUITE-RecoveryAfterSoftCommitTest-seed#[EF501846732A343]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via: @org.apache.solr.SolrTestCaseJ4$SuppressSSL(bugUrl="None") [junit4] 2> 595128 INFO (SUITE-RecoveryAfterSoftCommitTest-seed#[EF501846732A343]-worker) [ ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom [junit4] 2> 595128 INFO (SUITE-RecoveryAfterSoftCommitTest-seed#[EF501846732A343]-worker) [ ] o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: / [junit4] 2> 595129 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER [junit4] 2> 595129 INFO (Thread-1834) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0 [junit4] 2> 595129 INFO (Thread-1834) [ ] o.a.s.c.ZkTestServer Starting server [junit4] 2> 595154 ERROR (Thread-1834) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes [junit4] 2> 595229 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.ZkTestServer start zk server on port:38373 [junit4] 2> 595232 INFO (zkConnectionManagerCallback-2141-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 595233 INFO (zkConnectionManagerCallback-2143-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 595236 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml [junit4] 2> 595237 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/schema.xml to /configs/conf1/schema.xml [junit4] 2> 595237 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml to /configs/conf1/solrconfig.snippet.randomindexconfig.xml [junit4] 2> 595238 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt [junit4] 2> 595238 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt [junit4] 2> 595239 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml [junit4] 2> 595239 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml to /configs/conf1/enumsConfig.xml [junit4] 2> 595240 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json [junit4] 2> 595241 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt [junit4] 2> 595241 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt [junit4] 2> 595242 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/core/src/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt [junit4] 2> 595242 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractFullDistribZkTestBase Will use TLOG replicas unless explicitly asked otherwise [junit4] 2> 595327 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 10.0.1+10 [junit4] 2> 595344 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0 [junit4] 2> 595344 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.session No SessionScavenger set, using defaults [junit4] 2> 595344 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.session node0 Scavenging every 600000ms [junit4] 2> 595344 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@591e71bb{/,null,AVAILABLE} [junit4] 2> 595346 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@3986b917{HTTP/1.1,[http/1.1]}{127.0.0.1:44877} [junit4] 2> 595346 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.Server Started @595382ms [junit4] 2> 595346 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/, solr.data.dir=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/tempDir-001/control/data, hostPort=34913, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/control-001/cores, replicaType=NRT} [junit4] 2> 595346 ERROR (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete. [junit4] 2> 595346 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory [junit4] 2> 595346 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0 [junit4] 2> 595346 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null [junit4] 2> 595346 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null [junit4] 2> 595346 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-07-10T23:50:49.155643Z [junit4] 2> 595354 INFO (zkConnectionManagerCallback-2145-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 595355 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper) [junit4] 2> 595355 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/control-001/solr.xml [junit4] 2> 595357 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored [junit4] 2> 595357 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored [junit4] 2> 595358 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f, but no JMX reporters were configured - adding default JMX reporter. [junit4] 2> 595462 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:38373/solr [junit4] 2> 595463 INFO (zkConnectionManagerCallback-2149-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 595464 INFO (zkConnectionManagerCallback-2151-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 595544 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:34913_ ] o.a.s.c.Overseer Overseer (id=null) closing [junit4] 2> 595544 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:34913_ ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:34913_ [junit4] 2> 595544 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:34913_ ] o.a.s.c.Overseer Overseer (id=72105817116246020-127.0.0.1:34913_-n_0000000000) starting [junit4] 2> 595549 INFO (zkConnectionManagerCallback-2158-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 595550 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:34913_ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38373/solr ready [junit4] 2> 595553 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:34913_ ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:34913_ [junit4] 2> 595553 INFO (zkCallback-2150-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 595553 INFO (zkCallback-2157-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 595563 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:34913_ ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory. [junit4] 2> 595572 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:34913_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 595579 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:34913_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 595579 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:34913_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 595580 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:34913_ ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/control-001/cores [junit4] 2> 595604 INFO (zkConnectionManagerCallback-2164-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 595604 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 595605 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38373/solr ready [junit4] 2> 595606 INFO (SocketProxy-Acceptor-34913) [ ] o.a.s.c.SocketProxy accepted Socket[addr=/127.0.0.1,port=59150,localport=34913], receiveBufferSize:531000 [junit4] 2> 595618 INFO (SocketProxy-Acceptor-34913) [ ] o.a.s.c.SocketProxy proxy connection Socket[addr=/127.0.0.1,port=44877,localport=51992], receiveBufferSize=530904 [junit4] 2> 595619 INFO (qtp1195114600-8107) [n:127.0.0.1:34913_ ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:34913_&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 595623 INFO (OverseerThreadFactory-3556-thread-1) [ ] o.a.s.c.a.c.CreateCollectionCmd Create collection control_collection [junit4] 2> 595728 INFO (SocketProxy-Acceptor-34913) [ ] o.a.s.c.SocketProxy accepted Socket[addr=/127.0.0.1,port=59154,localport=34913], receiveBufferSize:531000 [junit4] 2> 595729 INFO (SocketProxy-Acceptor-34913) [ ] o.a.s.c.SocketProxy proxy connection Socket[addr=/127.0.0.1,port=44877,localport=51996], receiveBufferSize=530904 [junit4] 2> 595753 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ x:control_collection_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT [junit4] 2> 595753 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ x:control_collection_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores [junit4] 2> 596770 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0 [junit4] 2> 596798 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test [junit4] 2> 596883 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema test/1.0 with uniqueid field id [junit4] 2> 596898 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1' using configuration from collection control_collection, trusted=true [junit4] 2> 596898 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.control_collection.shard1.replica_n1' (registry 'solr.core.control_collection.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 596898 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder [junit4] 2> 596898 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/control-001/cores/control_collection_shard1_replica_n1], dataDir=[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/control-001/cores/control_collection_shard1_replica_n1/data/] [junit4] 2> 596902 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy: minMergeSize=1000, mergeFactor=29, maxMergeSize=9223372036854775807, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.4020472703188249] [junit4] 2> 596910 WARN (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}} [junit4] 2> 596948 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog [junit4] 2> 596948 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=2 maxNumLogsToKeep=10 numVersionBuckets=65536 [junit4] 2> 596949 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled [junit4] 2> 596949 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled [junit4] 2> 596949 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=24, maxMergeAtOnceExplicit=29, maxMergedSegmentMB=34.2255859375, floorSegmentMB=0.5888671875, forceMergeDeletesPctAllowed=20.25650626245356, segmentsPerTier=30.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, reclaimDeletesWeight=0.38370259958152886 [junit4] 2> 596950 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@410faf01[control_collection_shard1_replica_n1] main] [junit4] 2> 596950 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1 [junit4] 2> 596951 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1 [junit4] 2> 596951 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms. [junit4] 2> 596952 INFO (searcherExecutor-3561-thread-1-processing-n:127.0.0.1:34913_ x:control_collection_shard1_replica_n1 c:control_collection s:shard1) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [control_collection_shard1_replica_n1] Registered new searcher Searcher@410faf01[control_collection_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())} [junit4] 2> 596952 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1605649459588366336 [junit4] 2> 596956 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/control_collection/terms/shard1 to Terms{values={core_node2=0}, version=0} [junit4] 2> 596963 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue. [junit4] 2> 596963 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync [junit4] 2> 596963 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:34913/control_collection_shard1_replica_n1/ [junit4] 2> 596963 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me [junit4] 2> 596963 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy http://127.0.0.1:34913/control_collection_shard1_replica_n1/ has no replicas [junit4] 2> 596963 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR [junit4] 2> 596964 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:34913/control_collection_shard1_replica_n1/ shard1 [junit4] 2> 597070 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary [junit4] 2> 597072 INFO (qtp1195114600-8102) [n:127.0.0.1:34913_ c:control_collection s:shard1 x:control_collection_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1319 [junit4] 2> 597075 INFO (qtp1195114600-8107) [n:127.0.0.1:34913_ ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas [junit4] 2> 597172 INFO (zkCallback-2150-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/control_collection/state.json] for collection [control_collection] has occurred - updating... (live nodes size: [1]) [junit4] 2> 597625 INFO (OverseerCollectionConfigSetProcessor-72105817116246020-127.0.0.1:34913_-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper [junit4] 2> 598075 INFO (qtp1195114600-8107) [n:127.0.0.1:34913_ ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:34913_&wt=javabin&version=2} status=0 QTime=2455 [junit4] 2> 598079 INFO (zkConnectionManagerCallback-2169-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 598080 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 598081 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38373/solr ready [junit4] 2> 598081 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection loss:false [junit4] 2> 598081 INFO (SocketProxy-Acceptor-34913) [ ] o.a.s.c.SocketProxy accepted Socket[addr=/127.0.0.1,port=59166,localport=34913], receiveBufferSize:531000 [junit4] 2> 598081 INFO (SocketProxy-Acceptor-34913) [ ] o.a.s.c.SocketProxy proxy connection Socket[addr=/127.0.0.1,port=44877,localport=52008], receiveBufferSize=530904 [junit4] 2> 598082 INFO (qtp1195114600-8103) [n:127.0.0.1:34913_ ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=1&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 598100 INFO (OverseerThreadFactory-3556-thread-2) [ ] o.a.s.c.a.c.CreateCollectionCmd Create collection collection1 [junit4] 2> 598100 WARN (OverseerThreadFactory-3556-thread-2) [ ] o.a.s.c.a.c.CreateCollectionCmd It is unusual to create a collection (collection1) without cores. [junit4] 2> 598303 INFO (qtp1195114600-8103) [n:127.0.0.1:34913_ ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas [junit4] 2> 598303 INFO (qtp1195114600-8103) [n:127.0.0.1:34913_ ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=&stateFormat=1&wt=javabin&version=2} status=0 QTime=221 [junit4] 2> 598412 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-1-001 of type TLOG [junit4] 2> 598413 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 10.0.1+10 [junit4] 2> 598413 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0 [junit4] 2> 598413 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.session No SessionScavenger set, using defaults [junit4] 2> 598413 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.session node0 Scavenging every 660000ms [junit4] 2> 598413 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@cfbb72c{/,null,AVAILABLE} [junit4] 2> 598414 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@3fdeee88{HTTP/1.1,[http/1.1]}{127.0.0.1:40717} [junit4] 2> 598414 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.Server Started @598450ms [junit4] 2> 598414 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/, solrconfig=solrconfig.xml, solr.data.dir=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/tempDir-001/jetty1, hostPort=32995, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-1-001/cores, replicaType=TLOG} [junit4] 2> 598414 ERROR (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete. [junit4] 2> 598414 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory [junit4] 2> 598414 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0 [junit4] 2> 598414 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null [junit4] 2> 598414 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null [junit4] 2> 598414 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-07-10T23:50:52.223826Z [junit4] 2> 598426 INFO (zkConnectionManagerCallback-2171-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 598427 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper) [junit4] 2> 598427 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-1-001/solr.xml [junit4] 2> 598430 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored [junit4] 2> 598430 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored [junit4] 2> 598430 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f, but no JMX reporters were configured - adding default JMX reporter. [junit4] 2> 598524 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:38373/solr [junit4] 2> 598529 INFO (zkConnectionManagerCallback-2175-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 598532 INFO (zkConnectionManagerCallback-2177-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 598539 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:32995_ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1) [junit4] 2> 598539 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:32995_ ] o.a.s.c.Overseer Overseer (id=null) closing [junit4] 2> 598540 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:32995_ ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores [junit4] 2> 598540 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:32995_ ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:32995_ [junit4] 2> 598542 INFO (zkCallback-2168-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 598542 INFO (zkCallback-2157-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 598542 INFO (zkCallback-2150-thread-2) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 598542 INFO (zkCallback-2176-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2) [junit4] 2> 598554 INFO (zkConnectionManagerCallback-2184-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 598555 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:32995_ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2) [junit4] 2> 598556 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:32995_ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38373/solr ready [junit4] 2> 598556 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:32995_ ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory. [junit4] 2> 598573 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:32995_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 598580 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:32995_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 598580 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:32995_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 598581 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:32995_ ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-1-001/cores [junit4] 2> 598597 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy accepted Socket[addr=/127.0.0.1,port=35160,localport=32995], receiveBufferSize:531000 [junit4] 2> 598598 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy proxy connection Socket[addr=/127.0.0.1,port=40717,localport=53566], receiveBufferSize=530904 [junit4] 2> 598602 INFO (qtp986294191-8171) [n:127.0.0.1:32995_ ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:32995_&action=ADDREPLICA&collection=collection1&shard=shard1&type=TLOG&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 598609 INFO (OverseerCollectionConfigSetProcessor-72105817116246020-127.0.0.1:34913_-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000002 doesn't exist. Requestor may have disconnected from ZooKeeper [junit4] 2> 598614 INFO (OverseerThreadFactory-3556-thread-3) [ c:collection1 s:shard1 ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:32995_ for creating new replica [junit4] 2> 598615 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy accepted Socket[addr=/127.0.0.1,port=35164,localport=32995], receiveBufferSize:531000 [junit4] 2> 598620 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy proxy connection Socket[addr=/127.0.0.1,port=40717,localport=53570], receiveBufferSize=530904 [junit4] 2> 598620 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ x:collection1_shard1_replica_t21] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_t21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=TLOG [junit4] 2> 599631 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0 [junit4] 2> 599641 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.s.IndexSchema [collection1_shard1_replica_t21] Schema name=test [junit4] 2> 599705 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.s.IndexSchema Loaded schema test/1.0 with uniqueid field id [junit4] 2> 599736 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard1_replica_t21' using configuration from collection collection1, trusted=true [junit4] 2> 599736 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard1.replica_t21' (registry 'solr.core.collection1.shard1.replica_t21') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 599736 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder [junit4] 2> 599736 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SolrCore [[collection1_shard1_replica_t21] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-1-001/cores/collection1_shard1_replica_t21], dataDir=[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-1-001/cores/collection1_shard1_replica_t21/data/] [junit4] 2> 599738 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy: minMergeSize=1000, mergeFactor=29, maxMergeSize=9223372036854775807, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.4020472703188249] [junit4] 2> 599739 WARN (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}} [junit4] 2> 599788 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog [junit4] 2> 599788 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=2 maxNumLogsToKeep=10 numVersionBuckets=65536 [junit4] 2> 599789 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.CommitTracker Hard AutoCommit: disabled [junit4] 2> 599789 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.CommitTracker Soft AutoCommit: disabled [junit4] 2> 599790 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=24, maxMergeAtOnceExplicit=29, maxMergedSegmentMB=34.2255859375, floorSegmentMB=0.5888671875, forceMergeDeletesPctAllowed=20.25650626245356, segmentsPerTier=30.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, reclaimDeletesWeight=0.38370259958152886 [junit4] 2> 599790 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.s.SolrIndexSearcher Opening [Searcher@4afef50b[collection1_shard1_replica_t21] main] [junit4] 2> 599792 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1 [junit4] 2> 599793 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1 [junit4] 2> 599793 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms. [junit4] 2> 599793 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1605649462567370752 [junit4] 2> 599794 INFO (searcherExecutor-3575-thread-1-processing-n:127.0.0.1:32995_ x:collection1_shard1_replica_t21 c:collection1 s:shard1) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SolrCore [collection1_shard1_replica_t21] Registered new searcher Searcher@4afef50b[collection1_shard1_replica_t21] main{ExitableDirectoryReader(UninvertingDirectoryReader())} [junit4] 2> 599805 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard1 to Terms{values={core_node22=0}, version=0} [junit4] 2> 599815 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue. [junit4] 2> 599815 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync [junit4] 2> 599815 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:32995/collection1_shard1_replica_t21/ [junit4] 2> 599816 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me [junit4] 2> 599816 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.SyncStrategy http://127.0.0.1:32995/collection1_shard1_replica_t21/ has no replicas [junit4] 2> 599816 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR [junit4] 2> 599816 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ZkController collection1_shard1_replica_t21 stopping background replication from leader [junit4] 2> 599818 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:32995/collection1_shard1_replica_t21/ shard1 [junit4] 2> 599968 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.c.ZkController I am the leader, no recovery necessary [junit4] 2> 599976 INFO (qtp986294191-8166) [n:127.0.0.1:32995_ c:collection1 s:shard1 x:collection1_shard1_replica_t21] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_t21&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=TLOG} status=0 QTime=1356 [junit4] 2> 599978 INFO (qtp986294191-8171) [n:127.0.0.1:32995_ c:collection1 ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={node=127.0.0.1:32995_&action=ADDREPLICA&collection=collection1&shard=shard1&type=TLOG&wt=javabin&version=2} status=0 QTime=1376 [junit4] 2> 600106 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-2-001 of type TLOG [junit4] 2> 600106 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 10.0.1+10 [junit4] 2> 600107 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0 [junit4] 2> 600107 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.session No SessionScavenger set, using defaults [junit4] 2> 600107 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.session node0 Scavenging every 660000ms [junit4] 2> 600107 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@6c299510{/,null,AVAILABLE} [junit4] 2> 600107 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.AbstractConnector Started ServerConnector@5c8ba052{HTTP/1.1,[http/1.1]}{127.0.0.1:38153} [junit4] 2> 600107 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.e.j.s.Server Started @600144ms [junit4] 2> 600107 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/, solrconfig=solrconfig.xml, solr.data.dir=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/tempDir-001/jetty2, hostPort=44299, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-2-001/cores, replicaType=TLOG} [junit4] 2> 600108 ERROR (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete. [junit4] 2> 600108 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory [junit4] 2> 600108 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version 7.5.0 [junit4] 2> 600108 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null [junit4] 2> 600108 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null [junit4] 2> 600108 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-07-10T23:50:53.917255Z [junit4] 2> 600109 INFO (zkConnectionManagerCallback-2187-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 600109 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper) [junit4] 2> 600110 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-2-001/solr.xml [junit4] 2> 600112 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored [junit4] 2> 600112 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored [junit4] 2> 600113 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f, but no JMX reporters were configured - adding default JMX reporter. [junit4] 2> 600187 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:38373/solr [junit4] 2> 600192 INFO (zkConnectionManagerCallback-2191-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 600202 INFO (zkConnectionManagerCallback-2193-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 600205 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:44299_ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2) [junit4] 2> 600206 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:44299_ ] o.a.s.c.Overseer Overseer (id=null) closing [junit4] 2> 600207 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:44299_ ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores [junit4] 2> 600207 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:44299_ ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:44299_ [junit4] 2> 600208 INFO (zkCallback-2157-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 600208 INFO (zkCallback-2150-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 600209 INFO (zkCallback-2192-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 600210 INFO (zkCallback-2168-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 600210 INFO (zkCallback-2176-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 600210 INFO (zkCallback-2183-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3) [junit4] 2> 600234 INFO (zkConnectionManagerCallback-2200-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected [junit4] 2> 600234 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:44299_ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3) [junit4] 2> 600246 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:44299_ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38373/solr ready [junit4] 2> 600249 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:44299_ ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory. [junit4] 2> 600264 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:44299_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 600279 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:44299_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 600279 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:44299_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 600281 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [n:127.0.0.1:44299_ ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-2-001/cores [junit4] 2> 600333 INFO (qtp986294191-8165) [n:127.0.0.1:32995_ ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:44299_&action=ADDREPLICA&collection=collection1&shard=shard1&type=TLOG&wt=javabin&version=2 and sendToOCPQueue=true [junit4] 2> 600334 INFO (OverseerCollectionConfigSetProcessor-72105817116246020-127.0.0.1:34913_-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000004 doesn't exist. Requestor may have disconnected from ZooKeeper [junit4] 2> 600335 INFO (OverseerThreadFactory-3556-thread-4) [ c:collection1 s:shard1 ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:44299_ for creating new replica [junit4] 2> 600336 INFO (SocketProxy-Acceptor-44299) [ ] o.a.s.c.SocketProxy accepted Socket[addr=/127.0.0.1,port=39384,localport=44299], receiveBufferSize:531000 [junit4] 2> 600338 INFO (SocketProxy-Acceptor-44299) [ ] o.a.s.c.SocketProxy proxy connection Socket[addr=/127.0.0.1,port=38153,localport=44206], receiveBufferSize=530904 [junit4] 2> 600339 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ x:collection1_shard1_replica_t23] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_t23&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=TLOG [junit4] 2> 601352 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0 [junit4] 2> 601396 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.s.IndexSchema [collection1_shard1_replica_t23] Schema name=test [junit4] 2> 601486 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.s.IndexSchema Loaded schema test/1.0 with uniqueid field id [junit4] 2> 601541 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard1_replica_t23' using configuration from collection collection1, trusted=true [junit4] 2> 601541 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard1.replica_t23' (registry 'solr.core.collection1.shard1.replica_t23') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@5039755f [junit4] 2> 601541 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder [junit4] 2> 601541 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.c.SolrCore [[collection1_shard1_replica_t23] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-2-001/cores/collection1_shard1_replica_t23], dataDir=[/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.RecoveryAfterSoftCommitTest_EF501846732A343-001/shard-2-001/cores/collection1_shard1_replica_t23/data/] [junit4] 2> 601544 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy: minMergeSize=1000, mergeFactor=29, maxMergeSize=9223372036854775807, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.4020472703188249] [junit4] 2> 601545 WARN (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}} [junit4] 2> 601610 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog [junit4] 2> 601610 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=2 maxNumLogsToKeep=10 numVersionBuckets=65536 [junit4] 2> 601611 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.u.CommitTracker Hard AutoCommit: disabled [junit4] 2> 601611 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.u.CommitTracker Soft AutoCommit: disabled [junit4] 2> 601612 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=24, maxMergeAtOnceExplicit=29, maxMergedSegmentMB=34.2255859375, floorSegmentMB=0.5888671875, forceMergeDeletesPctAllowed=20.25650626245356, segmentsPerTier=30.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, reclaimDeletesWeight=0.38370259958152886 [junit4] 2> 601612 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.s.SolrIndexSearcher Opening [Searcher@10e10505[collection1_shard1_replica_t23] main] [junit4] 2> 601613 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1 [junit4] 2> 601613 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1 [junit4] 2> 601614 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms. [junit4] 2> 601614 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1605649464476827648 [junit4] 2> 601616 INFO (searcherExecutor-3589-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.c.SolrCore [collection1_shard1_replica_t23] Registered new searcher Searcher@10e10505[collection1_shard1_replica_t23] main{ExitableDirectoryReader(UninvertingDirectoryReader())} [junit4] 2> 601639 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard1 to Terms{values={core_node24=0, core_node22=0}, version=1} [junit4] 2> 601642 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.c.ZkController Core needs to recover:collection1_shard1_replica_t23 [junit4] 2> 601654 INFO (updateExecutor-2188-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.u.DefaultSolrCoreState Running recovery [junit4] 2> 601654 INFO (qtp1905655413-8215) [n:127.0.0.1:44299_ c:collection1 s:shard1 x:collection1_shard1_replica_t23] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_t23&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=TLOG} status=0 QTime=1315 [junit4] 2> 601655 INFO (qtp986294191-8165) [n:127.0.0.1:32995_ c:collection1 ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={node=127.0.0.1:44299_&action=ADDREPLICA&collection=collection1&shard=shard1&type=TLOG&wt=javabin&version=2} status=0 QTime=1322 [junit4] 2> 601674 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.RecoveryStrategy Starting recovery process. recoveringAfterStartup=true [junit4] 2> 601680 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.RecoveryStrategy ###### startupVersions=[[]] [junit4] 2> 601680 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.ZkController collection1_shard1_replica_t23 stopping background replication from leader [junit4] 2> 601680 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy accepted Socket[addr=/127.0.0.1,port=35180,localport=32995], receiveBufferSize:531000 [junit4] 2> 601681 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.SolrTestCaseJ4 ###Starting test [junit4] 2> 601681 INFO (TEST-RecoveryAfterSoftCommitTest.test-seed#[EF501846732A343]) [ ] o.a.s.c.AbstractDistribZkTestBase Wait for recoveries to finish - collection: collection1 failOnTimeout:true timeout (sec):330 [junit4] 2> 601681 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy proxy connection Socket[addr=/127.0.0.1,port=40717,localport=53586], receiveBufferSize=530904 [junit4] 1> - [junit4] 1> replica:core_node22 rstate:active live:true [junit4] 1> replica:core_node24 rstate:down live:true [junit4] 2> 601682 INFO (qtp986294191-8167) [n:127.0.0.1:32995_ c:collection1 s:shard1 r:core_node22 x:collection1_shard1_replica_t21] o.a.s.c.S.Request [collection1_shard1_replica_t21] webapp= path=/admin/ping params={wt=javabin&version=2} hits=0 status=0 QTime=0 [junit4] 2> 601682 INFO (qtp986294191-8167) [n:127.0.0.1:32995_ c:collection1 s:shard1 r:core_node22 x:collection1_shard1_replica_t21] o.a.s.c.S.Request [collection1_shard1_replica_t21] webapp= path=/admin/ping params={wt=javabin&version=2} status=0 QTime=0 [junit4] 2> 601682 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.RecoveryStrategy Begin buffering updates. core=[collection1_shard1_replica_t23] [junit4] 2> 601682 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.u.UpdateLog Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null} [junit4] 2> 601682 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.RecoveryStrategy Publishing state of core [collection1_shard1_replica_t23] as recovering, leader is [http://127.0.0.1:32995/collection1_shard1_replica_t21/] and I am [http://127.0.0.1:44299/collection1_shard1_replica_t23/] [junit4] 2> 601683 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard1 to Terms{values={core_node24_recovering=0, core_node24=0, core_node22=0}, version=2} [junit4] 2> 601684 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.RecoveryStrategy Sending prep recovery command to [http://127.0.0.1:32995]; [WaitForState: action=PREPRECOVERY&core=collection1_shard1_replica_t21&nodeName=127.0.0.1:44299_&coreNodeName=core_node24&state=recovering&checkLive=true&onlyIfLeader=true&onlyIfLeaderActive=true] [junit4] 2> 601685 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy accepted Socket[addr=/127.0.0.1,port=35184,localport=32995], receiveBufferSize:531000 [junit4] 2> 601685 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy proxy connection Socket[addr=/127.0.0.1,port=40717,localport=53590], receiveBufferSize=530904 [junit4] 2> 601691 INFO (qtp986294191-8164) [n:127.0.0.1:32995_ x:collection1_shard1_replica_t21] o.a.s.h.a.PrepRecoveryOp Going to wait for coreNodeName: core_node24, state: recovering, checkLive: true, onlyIfLeader: true, onlyIfLeaderActive: true, maxTime: 183 s [junit4] 2> 601691 INFO (qtp986294191-8164) [n:127.0.0.1:32995_ x:collection1_shard1_replica_t21] o.a.s.h.a.PrepRecoveryOp In WaitForState(recovering): collection=collection1, shard=shard1, thisCore=collection1_shard1_replica_t21, leaderDoesNotNeedRecovery=false, isLeader? true, live=true, checkLive=true, currentState=down, localState=active, nodeName=127.0.0.1:44299_, coreNodeName=core_node24, onlyIfActiveCheckResult=false, nodeProps: core_node24:{"core":"collection1_shard1_replica_t23","base_url":"http://127.0.0.1:44299","node_name":"127.0.0.1:44299_","state":"down","type":"TLOG"} [junit4] 2> 602343 INFO (OverseerCollectionConfigSetProcessor-72105817116246020-127.0.0.1:34913_-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000006 doesn't exist. Requestor may have disconnected from ZooKeeper [junit4] 1> - [junit4] 1> replica:core_node22 rstate:active live:true [junit4] 1> replica:core_node24 rstate:recovering live:true [junit4] 2> 602691 INFO (qtp986294191-8164) [n:127.0.0.1:32995_ x:collection1_shard1_replica_t21] o.a.s.h.a.PrepRecoveryOp In WaitForState(recovering): collection=collection1, shard=shard1, thisCore=collection1_shard1_replica_t21, leaderDoesNotNeedRecovery=false, isLeader? true, live=true, checkLive=true, currentState=recovering, localState=active, nodeName=127.0.0.1:44299_, coreNodeName=core_node24, onlyIfActiveCheckResult=false, nodeProps: core_node24:{"core":"collection1_shard1_replica_t23","base_url":"http://127.0.0.1:44299","node_name":"127.0.0.1:44299_","state":"recovering","type":"TLOG"} [junit4] 2> 602691 INFO (qtp986294191-8164) [n:127.0.0.1:32995_ x:collection1_shard1_replica_t21] o.a.s.h.a.PrepRecoveryOp Waited coreNodeName: core_node24, state: recovering, checkLive: true, onlyIfLeader: true for: 1 seconds. [junit4] 2> 602691 INFO (qtp986294191-8164) [n:127.0.0.1:32995_ x:collection1_shard1_replica_t21] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={nodeName=127.0.0.1:44299_&onlyIfLeaderActive=true&core=collection1_shard1_replica_t21&coreNodeName=core_node24&action=PREPRECOVERY&checkLive=true&state=recovering&onlyIfLeader=true&wt=javabin&version=2} status=0 QTime=1000 [junit4] 2> 603192 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.RecoveryStrategy Starting Replication Recovery. [junit4] 2> 603192 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.RecoveryStrategy Attempting to replicate from [http://127.0.0.1:32995/collection1_shard1_replica_t21/]. [junit4] 2> 603193 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy accepted Socket[addr=/127.0.0.1,port=35216,localport=32995], receiveBufferSize:531000 [junit4] 2> 603193 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy proxy connection Socket[addr=/127.0.0.1,port=40717,localport=53622], receiveBufferSize=530904 [junit4] 2> 603194 INFO (qtp986294191-8200) [n:127.0.0.1:32995_ c:collection1 s:shard1 r:core_node22 x:collection1_shard1_replica_t21] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1605649466133577728,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false} [junit4] 2> 603194 INFO (qtp986294191-8200) [n:127.0.0.1:32995_ c:collection1 s:shard1 r:core_node22 x:collection1_shard1_replica_t21] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit. [junit4] 2> 603194 INFO (qtp986294191-8200) [n:127.0.0.1:32995_ c:collection1 s:shard1 r:core_node22 x:collection1_shard1_replica_t21] o.a.s.u.DirectUpdateHandler2 end_commit_flush [junit4] 2> 603194 INFO (qtp986294191-8200) [n:127.0.0.1:32995_ c:collection1 s:shard1 r:core_node22 x:collection1_shard1_replica_t21] o.a.s.u.p.LogUpdateProcessorFactory [collection1_shard1_replica_t21] webapp= path=/update params={waitSearcher=true&openSearcher=false&commit=true&softCommit=false&commit_end_point=true&wt=javabin&version=2}{commit=} 0 0 [junit4] 2> 603196 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy accepted Socket[addr=/127.0.0.1,port=35220,localport=32995], receiveBufferSize:531000 [junit4] 2> 603197 INFO (SocketProxy-Acceptor-32995) [ ] o.a.s.c.SocketProxy proxy connection Socket[addr=/127.0.0.1,port=40717,localport=53626], receiveBufferSize=530904 [junit4] 2> 603206 INFO (qtp986294191-8170) [n:127.0.0.1:32995_ c:collection1 s:shard1 r:core_node22 x:collection1_shard1_replica_t21] o.a.s.c.S.Request [collection1_shard1_replica_t21] webapp= path=/replication params={qt=/replication&wt=javabin&version=2&command=indexversion} status=0 QTime=0 [junit4] 2> 603206 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.h.IndexFetcher Master's generation: 1 [junit4] 2> 603206 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.h.IndexFetcher Master's version: 0 [junit4] 2> 603206 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.h.IndexFetcher Slave's generation: 1 [junit4] 2> 603206 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.h.IndexFetcher Slave's version: 0 [junit4] 2> 603206 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.h.IndexFetcher New index in Master. Deleting mine... [junit4] 2> 603207 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.s.SolrIndexSearcher Opening [Searcher@2e80c035[collection1_shard1_replica_t23] main] [junit4] 2> 603208 INFO (searcherExecutor-3589-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.SolrCore [collection1_shard1_replica_t23] Registered new searcher Searcher@2e80c035[collection1_shard1_replica_t23] main{ExitableDirectoryReader(UninvertingDirectoryReader())} [junit4] 2> 603209 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.RecoveryStrategy Replication Recovery was successful. [junit4] 2> 603209 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.RecoveryStrategy Registering as Active after recovery. [junit4] 2> 603209 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.ZkController collection1_shard1_replica_t23 starting background replication from leader [junit4] 2> 603209 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.c.ReplicateFromLeader Will start replication from leader with poll interval: 00:00:03 [junit4] 2> 603209 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.h.ReplicationHandler Poll scheduled at an interval of 3000ms [junit4] 2> 603210 INFO (recoveryExecutor-2189-thread-1-processing-n:127.0.0.1:44299_ x:collection1_shard1_replica_t23 c:collection1 s:shard1 r:core_node24) [n:127.0.0.1:44299_ c:collection1 s:shard1 r:core_node24 x:collection1_shard1_replica_t23] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms. [junit4] 2> 603210 INFO (indexFetcher-3595-thread-1) [ ] o.a.s.h.IndexFetcher Updated masterUrl to htt [...truncated too long message...] current data watches above limit: [junit4] 2> [junit4] 2> 19 /solr/clusterprops.json [junit4] 2> 19 /solr/clusterstate.json [junit4] 2> 11 /solr/collections/MoveReplicaHDFSTest_coll_true/state.json [junit4] 2> 10 /solr/aliases.json [junit4] 2> 7 /solr/collections/MoveReplicaHDFSTest_coll_false/state.json [junit4] 2> 4 /solr/collections/MoveReplicaHDFSTest_failed_coll_false/state.json [junit4] 2> 4 /solr/collections/MoveReplicaHDFSTest_failed_coll_true/state.json [junit4] 2> 3 /solr/collections/MoveReplicaHDFSTest_coll_false/leader_elect/shard1/election/72105821468360715-core_node5-n_0000000000 [junit4] 2> 3 /solr/overseer_elect/election/72105821468360719-127.0.0.1:38555_solr-n_0000000003 [junit4] 2> 3 /solr/collections/MoveReplicaHDFSTest_coll_true/leader_elect/shard2/election/72105821468360715-core_node8-n_0000000000 [junit4] 2> 2 /solr/autoscaling.json [junit4] 2> [junit4] 2> Maximum concurrent children watches above limit: [junit4] 2> [junit4] 2> 40 /solr/collections [junit4] 2> 28 /solr/live_nodes [junit4] 2> 2 /solr/overseer/queue [junit4] 2> 2 /solr/autoscaling/events/.scheduled_maintenance [junit4] 2> 2 /solr/autoscaling/events/.auto_add_replicas [junit4] 2> 2 /solr/overseer/queue-work [junit4] 2> 2 /solr/overseer/collection-queue-work [junit4] 2> [junit4] 2> 738250 WARN (SUITE-MoveReplicaHDFSTest-seed#[EF501846732A343]-worker) [ ] o.a.h.h.s.d.DirectoryScanner DirectoryScanner: shutdown has been called [junit4] 2> 739397 INFO (SUITE-MoveReplicaHDFSTest-seed#[EF501846732A343]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0 [junit4] 2> 739499 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_EF501846732A343-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_EF501846732A343-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:35053) [ ] o.a.h.h.s.d.IncrementalBlockReportManager IncrementalBlockReportManager interrupted [junit4] 2> 739499 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_EF501846732A343-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_EF501846732A343-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:35053) [ ] o.a.h.h.s.d.DataNode Ending block pool service for: Block pool BP-1857384683-88.99.242.108-1531266716089 (Datanode Uuid af4f0e7c-fc23-4e90-a705-3d74a8614092) service to localhost.localdomain/127.0.0.1:35053 [junit4] 2> 739499 WARN (SUITE-MoveReplicaHDFSTest-seed#[EF501846732A343]-worker) [ ] o.a.h.h.s.d.DirectoryScanner DirectoryScanner: shutdown has been called [junit4] 2> 739504 INFO (SUITE-MoveReplicaHDFSTest-seed#[EF501846732A343]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:0 [junit4] 2> 739606 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_EF501846732A343-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_EF501846732A343-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:35053) [ ] o.a.h.h.s.d.IncrementalBlockReportManager IncrementalBlockReportManager interrupted [junit4] 2> 739606 WARN (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_EF501846732A343-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_EF501846732A343-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:35053) [ ] o.a.h.h.s.d.DataNode Ending block pool service for: Block pool BP-1857384683-88.99.242.108-1531266716089 (Datanode Uuid 7bf7654d-41bb-4961-9016-0e0ce4037045) service to localhost.localdomain/127.0.0.1:35053 [junit4] 2> 739610 INFO (SUITE-MoveReplicaHDFSTest-seed#[EF501846732A343]-worker) [ ] o.m.log Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:0 [junit4] 2> 739721 INFO (SUITE-MoveReplicaHDFSTest-seed#[EF501846732A343]-worker) [ ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:44667 44667 [junit4] 2> NOTE: leaving temporary files on disk at: /home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_EF501846732A343-001 [junit4] 2> Jul 10, 2018 11:53:13 PM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks [junit4] 2> WARNING: Will linger awaiting termination of 66 leaked thread(s). [junit4] 2> NOTE: test params are: codec=Asserting(Lucene70): {id=Lucene50(blocksize=128)}, docValues:{_version_=DocValuesFormat(name=Memory)}, maxPointsInLeafNode=76, maxMBSortInHeap=5.213616764232611, sim=RandomSimilarity(queryNorm=true): {}, locale=en-AU, timezone=NET [junit4] 2> NOTE: Linux 4.15.0-24-generic amd64/Oracle Corporation 10.0.1 (64-bit)/cpus=8,threads=5,free=143611616,total=536870912 [junit4] 2> NOTE: All tests run in this JVM: [AtomicUpdatesTest, RequiredFieldsTest, HdfsRecoveryZkTest, TestLegacyNumericUtils, QueryEqualityTest, DOMUtilTest, TestBlendedInfixSuggestions, TestShardHandlerFactory, TestSystemIdResolver, TestJettySolrRunner, TestCloudSchemaless, TestTrackingShardHandlerFactory, BlobRepositoryCloudTest, BlockJoinFacetRandomTest, TestSolrDeletionPolicy1, TestLeaderElectionZkExpiry, TestLocalFSCloudBackupRestore, TolerantUpdateProcessorTest, TestNamedUpdateProcessors, DirectUpdateHandlerTest, TestConfigOverlay, LeaderFailoverAfterPartitionTest, TestSizeLimitedDistributedMap, DistanceFunctionTest, DistributedIntervalFacetingTest, TestDynamicFieldCollectionResource, SolrIndexSplitterTest, AddReplicaTest, TestMiniSolrCloudClusterSSL, HdfsThreadLeakTest, MultiThreadedOCPTest, TestClusterProperties, TestSubQueryTransformerDistrib, FileUtilsTest, TestJoin, ZkShardTermsTest, JSONWriterTest, ShardRoutingCustomTest, TestBadConfig, LoggingHandlerTest, ConnectionManagerTest, OverriddenZkACLAndCredentialsProvidersTest, TestLazyCores, TestSolrCloudWithHadoopAuthPlugin, DirectSolrConnectionTest, HLLSerializationTest, BadCopyFieldTest, DistribCursorPagingTest, DistributedSpellCheckComponentTest, TestCollectionAPI, TestSolrXml, ZkControllerTest, BasicAuthIntegrationTest, TestEmbeddedSolrServerAdminHandler, TestCSVLoader, OutputWriterTest, DocumentAnalysisRequestHandlerTest, ChaosMonkeySafeLeaderTest, AsyncCallRequestStatusResponseTest, TestSolrQueryParser, TestWordDelimiterFilterFactory, TestAuthorizationFramework, PreAnalyzedFieldManagedSchemaCloudTest, TestCloudPseudoReturnFields, TestLMDirichletSimilarityFactory, UnloadDistributedZkTest, RegexBoostProcessorTest, TestQueryTypes, DistributedFacetPivotLargeTest, TestCharFilters, UniqFieldsUpdateProcessorFactoryTest, ComputePlanActionTest, TestConfigSets, DocValuesMultiTest, RegexBytesRefFilterTest, RestoreTriggerStateTest, TestSolrCoreParser, RecoveryZkTest, TestGeoJSONResponseWriter, SpellCheckCollatorTest, TestBinaryField, DistributedFacetPivotSmallTest, TestFuzzyAnalyzedSuggestions, TestFastOutputStream, TestCloudJSONFacetJoinDomain, TestReqParamsAPI, TestGraphTermsQParserPlugin, HdfsChaosMonkeySafeLeaderTest, DocExpirationUpdateProcessorFactoryTest, TestXmlQParserPlugin, CoreAdminRequestStatusTest, SimpleMLTQParserTest, TestInPlaceUpdatesStandalone, TestDocumentBuilder, PluginInfoTest, UninvertDocValuesMergePolicyTest, HttpTriggerListenerTest, ResponseBuilderTest, TestSSLRandomization, TestChildDocTransformer, NodeLostTriggerIntegrationTest, TestScoreJoinQPScore, TestFieldCacheSort, PrimUtilsTest, MoveReplicaHDFSTest] [junit4] Completed [302/822 (2!)] on J2 in 86.53s, 4 tests, 1 error <<< FAILURES! [...truncated 43756 lines...] [repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/2294/consoleText [repro] Revision: 937ae2c7b6a88aa080cd843ad3663937be61ee2f [repro] Ant options: "-Dargs=-XX:+UseCompressedOops -XX:+UseG1GC" [repro] Repro line: ant test -Dtestcase=RecoveryAfterSoftCommitTest -Dtests.method=test -Dtests.seed=EF501846732A343 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=es-PH -Dtests.timezone=Africa/Mbabane -Dtests.asserts=true -Dtests.file.encoding=UTF-8 [repro] Repro line: ant test -Dtestcase=MoveReplicaHDFSTest -Dtests.method=testFailedMove -Dtests.seed=EF501846732A343 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=en-AU -Dtests.timezone=NET -Dtests.asserts=true -Dtests.file.encoding=UTF-8 [repro] ant clean [...truncated 6 lines...] [repro] Test suites by module: [repro] solr/core [repro] RecoveryAfterSoftCommitTest [repro] MoveReplicaHDFSTest [repro] ant compile-test [...truncated 3323 lines...] [repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=10 -Dtests.class="*.RecoveryAfterSoftCommitTest|*.MoveReplicaHDFSTest" -Dtests.showOutput=onerror "-Dargs=-XX:+UseCompressedOops -XX:+UseG1GC" -Dtests.seed=EF501846732A343 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=es-PH -Dtests.timezone=Africa/Mbabane -Dtests.asserts=true -Dtests.file.encoding=UTF-8 [...truncated 112 lines...] [repro] Failures: [repro] 0/5 failed: org.apache.solr.cloud.MoveReplicaHDFSTest [repro] 0/5 failed: org.apache.solr.cloud.RecoveryAfterSoftCommitTest [repro] Exiting with code 0 [...truncated 40 lines...]
--------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
