Build: https://builds.apache.org/job/Lucene-Solr-BadApples-Tests-7.x/176/
2 tests failed.
FAILED:
org.apache.solr.cloud.api.collections.ShardSplitTest.testSplitMixedReplicaTypesLink
Error Message:
unexpected shard state expected:<inactive> but was:<active>
Stack Trace:
java.lang.AssertionError: unexpected shard state expected:<inactive> but
was:<active>
at
__randomizedtesting.SeedInfo.seed([9223DC651078BC8A:AE4D153CB5A01713]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.failNotEquals(Assert.java:647)
at org.junit.Assert.assertEquals(Assert.java:128)
at
org.apache.solr.cloud.api.collections.ShardSplitTest.verifyShard(ShardSplitTest.java:374)
at
org.apache.solr.cloud.api.collections.ShardSplitTest.doSplitMixedReplicaTypes(ShardSplitTest.java:366)
at
org.apache.solr.cloud.api.collections.ShardSplitTest.testSplitMixedReplicaTypesLink(ShardSplitTest.java:338)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1742)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:935)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:971)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:985)
at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:1010)
at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:985)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:944)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:830)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:880)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:891)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
FAILED: org.apache.solr.cloud.api.collections.TestHdfsCloudBackupRestore.test
Error Message:
Error from server at http://127.0.0.1:45193/solr: create the collection time
out:180s
Stack Trace:
org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error
from server at http://127.0.0.1:45193/solr: create the collection time out:180s
at
__randomizedtesting.SeedInfo.seed([9223DC651078BC8A:1A77E3BFBE84D172]:0)
at
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:643)
at
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255)
at
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244)
at
org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:483)
at
org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:413)
at
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1107)
at
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:884)
at
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:817)
at
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194)
at
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:211)
at
org.apache.solr.cloud.api.collections.AbstractCloudBackupRestoreTestCase.test(AbstractCloudBackupRestoreTestCase.java:125)
at
org.apache.solr.cloud.api.collections.TestHdfsCloudBackupRestore.test(TestHdfsCloudBackupRestore.java:213)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1742)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:935)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:971)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:985)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:944)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:830)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:880)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:891)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
Build Log:
[...truncated 13413 lines...]
[junit4] Suite: org.apache.solr.cloud.api.collections.ShardSplitTest
[junit4] 2> Creating dataDir:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/init-core-data-001
[junit4] 2> 1063305 INFO
(SUITE-ShardSplitTest-seed#[9223DC651078BC8A]-worker) [ ]
o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /
[junit4] 2> 1063316 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 1063327 INFO (Thread-2238) [ ] o.a.s.c.ZkTestServer client
port:0.0.0.0/0.0.0.0:0
[junit4] 2> 1063327 INFO (Thread-2238) [ ] o.a.s.c.ZkTestServer
Starting server
[junit4] 2> 1063343 ERROR (Thread-2238) [ ] o.a.z.s.ZooKeeperServer
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action
on ERROR or SHUTDOWN server state changes
[junit4] 2> 1063427 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.ZkTestServer start zk server on port:43542
[junit4] 2> 1063429 INFO (zkConnectionManagerCallback-7185-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1063431 INFO (zkConnectionManagerCallback-7187-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1063436 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
to /configs/conf1/solrconfig.xml
[junit4] 2> 1063437 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/schema15.xml
to /configs/conf1/schema.xml
[junit4] 2> 1063438 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
[junit4] 2> 1063439 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
to /configs/conf1/stopwords.txt
[junit4] 2> 1063440 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/protwords.txt
to /configs/conf1/protwords.txt
[junit4] 2> 1063441 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/currency.xml
to /configs/conf1/currency.xml
[junit4] 2> 1063442 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml
to /configs/conf1/enumsConfig.xml
[junit4] 2> 1063443 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json
to /configs/conf1/open-exchange-rates.json
[junit4] 2> 1063444 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt
to /configs/conf1/mapping-ISOLatin1Accent.txt
[junit4] 2> 1063445 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt
to /configs/conf1/old_synonyms.txt
[junit4] 2> 1063446 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractZkTestCase put
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/synonyms.txt
to /configs/conf1/synonyms.txt
[junit4] 2> 1063448 INFO (zkConnectionManagerCallback-7191-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1063455 WARN (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [ ]
o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid
0x10162ca86e40002, likely client has closed socket
[junit4] 2> 1063455 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase Will use NRT replicas unless explicitly
asked otherwise
[junit4] 2> 1063577 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git:
d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
[junit4] 2> 1063578 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1063578 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1063578 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 1063578 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@380393b9{/,null,AVAILABLE}
[junit4] 2> 1063580 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.AbstractConnector Started ServerConnector@6866b11c{SSL,[ssl,
http/1.1]}{127.0.0.1:41386}
[junit4] 2> 1063580 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.Server Started @1063672ms
[junit4] 2> 1063580 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/tempDir-001/control/data,
hostContext=/, hostPort=41386,
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/control-001/cores}
[junit4] 2> 1063580 ERROR
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 1063580 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter Using logger factory
org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1063581 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.6.0
[junit4] 2> 1063581 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1063581 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1063581 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2018-10-04T14:34:51.423Z
[junit4] 2> 1063582 INFO (zkConnectionManagerCallback-7193-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1063583 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 1063583 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/control-001/solr.xml
[junit4] 2> 1063587 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay
is ignored
[junit4] 2> 1063587 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.SolrXmlConfig Configuration parameter
autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 1063588 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.SolrXmlConfig MBean server found:
com.sun.jmx.mbeanserver.JmxMBeanServer@32eae29b, but no JMX reporters were
configured - adding default JMX reporter.
[junit4] 2> 1063756 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:43542/solr
[junit4] 2> 1063757 INFO (zkConnectionManagerCallback-7197-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1063759 INFO (zkConnectionManagerCallback-7199-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1063837 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A])
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerElectionContext I am going to be the
leader 127.0.0.1:41386_
[junit4] 2> 1063838 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A])
[n:127.0.0.1:41386_ ] o.a.s.c.Overseer Overseer
(id=72447691000643589-127.0.0.1:41386_-n_0000000000) starting
[junit4] 2> 1063844 INFO (zkConnectionManagerCallback-7206-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1063846 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A])
[n:127.0.0.1:41386_ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at
127.0.0.1:43542/solr ready
[junit4] 2> 1063847 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Process current queue
of overseer operations
[junit4] 2> 1063847 INFO
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.Overseer Starting to work on the main queue :
127.0.0.1:41386_
[junit4] 2> 1063847 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A])
[n:127.0.0.1:41386_ ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:41386_
[junit4] 2> 1063849 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Cleaning up work-queue.
#Running tasks: 0
[junit4] 2> 1063849 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningTasks: []
[junit4] 2> 1063849 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
[junit4] 2> 1063849 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor CompletedTasks: []
[junit4] 2> 1063849 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningZKTasks: []
[junit4] 2> 1063851 INFO (zkCallback-7205-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1063864 INFO
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (1)
[junit4] 2> 1063876 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A])
[n:127.0.0.1:41386_ ] o.a.s.h.a.MetricsHistoryHandler No .system collection,
keeping metrics history in memory.
[junit4] 2> 1063895 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A])
[n:127.0.0.1:41386_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.node' (registry 'solr.node') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@32eae29b
[junit4] 2> 1063904 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A])
[n:127.0.0.1:41386_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jvm' (registry 'solr.jvm') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@32eae29b
[junit4] 2> 1063904 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A])
[n:127.0.0.1:41386_ ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jetty' (registry 'solr.jetty') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@32eae29b
[junit4] 2> 1063905 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A])
[n:127.0.0.1:41386_ ] o.a.s.c.CorePropertiesLocator Found 0 core definitions
underneath
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/control-001/cores
[junit4] 2> 1063926 INFO (zkConnectionManagerCallback-7212-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1063927 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1063928 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:43542/solr ready
[junit4] 2> 1063950 INFO (qtp1635669701-15079) [n:127.0.0.1:41386_ ]
o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params
collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:41386_&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 1063957 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Got 1 tasks from
work-queue : [[org.apache.solr.cloud.OverseerTaskQueue$QueueEvent@2e684ecf]]
[junit4] 2> 1063957 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Marked task
[/overseer/collection-queue-work/qn-0000000000] as running
[junit4] 2> 1063957 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Overseer Collection
Message Handler: Get the message
id:/overseer/collection-queue-work/qn-0000000000 message:{
[junit4] 2> "name":"control_collection",
[junit4] 2> "fromApi":"true",
[junit4] 2> "collection.configName":"conf1",
[junit4] 2> "numShards":"1",
[junit4] 2> "createNodeSet":"127.0.0.1:41386_",
[junit4] 2> "nrtReplicas":"1",
[junit4] 2> "stateFormat":"2",
[junit4] 2> "replicationFactor":"1",
[junit4] 2> "operation":"create"}
[junit4] 2> 1063958 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Runner processing
/overseer/collection-queue-work/qn-0000000000
[junit4] 2> 1063958 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.OverseerCollectionMessageHandler
OverseerCollectionMessageHandler.processMessage : create , {
[junit4] 2> "name":"control_collection",
[junit4] 2> "fromApi":"true",
[junit4] 2> "collection.configName":"conf1",
[junit4] 2> "numShards":"1",
[junit4] 2> "createNodeSet":"127.0.0.1:41386_",
[junit4] 2> "nrtReplicas":"1",
[junit4] 2> "stateFormat":"2",
[junit4] 2> "replicationFactor":"1",
[junit4] 2> "operation":"create"}
[junit4] 2> 1063958 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Cleaning up work-queue.
#Running tasks: 1
[junit4] 2> 1063958 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningTasks:
[/overseer/collection-queue-work/qn-0000000000]
[junit4] 2> 1063958 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
[junit4] 2> 1063958 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor CompletedTasks: []
[junit4] 2> 1063958 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningZKTasks:
[/overseer/collection-queue-work/qn-0000000000]
[junit4] 2> 1063958 INFO
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.CreateCollectionCmd Create collection
control_collection
[junit4] 2> 1063959 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.OverseerCollectionMessageHandler creating
collections conf node /collections/control_collection
[junit4] 2> 1063960 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.CreateCollectionCmd Check for collection
zkNode:control_collection
[junit4] 2> 1063960 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.CreateCollectionCmd Collection zkNode
exists
[junit4] 2> 1063961 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.Overseer processMessage: queueSize: 1, message
= {
[junit4] 2> "name":"control_collection",
[junit4] 2> "fromApi":"true",
[junit4] 2> "collection.configName":"conf1",
[junit4] 2> "numShards":"1",
[junit4] 2> "createNodeSet":"127.0.0.1:41386_",
[junit4] 2> "nrtReplicas":"1",
[junit4] 2> "stateFormat":"2",
[junit4] 2> "replicationFactor":"1",
[junit4] 2> "operation":"create"} current state version: 0
[junit4] 2> 1063962 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ClusterStateMutator building a new cName:
control_collection
[junit4] 2> 1063962 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ZkStateWriter going to create_collection
/collections/control_collection/state.json
[junit4] 2> 1064063 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.CreateCollectionCmd Creating SolrCores for
new collection control_collection, shardNames [shard1] , message : {
[junit4] 2> "name":"control_collection",
[junit4] 2> "fromApi":"true",
[junit4] 2> "collection.configName":"conf1",
[junit4] 2> "numShards":"1",
[junit4] 2> "createNodeSet":"127.0.0.1:41386_",
[junit4] 2> "nrtReplicas":"1",
[junit4] 2> "stateFormat":"2",
[junit4] 2> "replicationFactor":"1",
[junit4] 2> "operation":"create"}
[junit4] 2> 1064065 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.CreateCollectionCmd Creating core
control_collection_shard1_replica_n1 as part of shard shard1 of collection
control_collection on 127.0.0.1:41386_
[junit4] 2> 1064075 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
x:control_collection_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core
create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1064076 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
x:control_collection_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault
Allocating transient cache for 4 transient cores
[junit4] 2> 1064078 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.Overseer processMessage: queueSize: 1, message
= {
[junit4] 2> "core":"control_collection_shard1_replica_n1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"https://127.0.0.1:41386",
[junit4] 2> "node_name":"127.0.0.1:41386_",
[junit4] 2> "numShards":"1",
[junit4] 2> "state":"down",
[junit4] 2> "shard":"shard1",
[junit4] 2> "collection":"control_collection",
[junit4] 2> "type":"NRT",
[junit4] 2> "operation":"state"} current state version: 0
[junit4] 2> 1064078 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Update state numShards=1
message={
[junit4] 2> "core":"control_collection_shard1_replica_n1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"https://127.0.0.1:41386",
[junit4] 2> "node_name":"127.0.0.1:41386_",
[junit4] 2> "numShards":"1",
[junit4] 2> "state":"down",
[junit4] 2> "shard":"shard1",
[junit4] 2> "collection":"control_collection",
[junit4] 2> "type":"NRT",
[junit4] 2> "operation":"state"}
[junit4] 2> 1064080 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Will update state for
replica:
core_node2:{"core":"control_collection_shard1_replica_n1","base_url":"https://127.0.0.1:41386","node_name":"127.0.0.1:41386_","state":"down","type":"NRT"}
[junit4] 2> 1064080 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Collection is now:
DocCollection(control_collection//collections/control_collection/state.json/0)={
[junit4] 2> "pullReplicas":"0",
[junit4] 2> "replicationFactor":"1",
[junit4] 2> "router":{"name":"compositeId"},
[junit4] 2> "maxShardsPerNode":"1",
[junit4] 2> "autoAddReplicas":"false",
[junit4] 2> "nrtReplicas":"1",
[junit4] 2> "tlogReplicas":"0",
[junit4] 2> "shards":{"shard1":{
[junit4] 2> "range":"80000000-7fffffff",
[junit4] 2> "state":"active",
[junit4] 2> "replicas":{"core_node2":{
[junit4] 2> "core":"control_collection_shard1_replica_n1",
[junit4] 2> "base_url":"https://127.0.0.1:41386",
[junit4] 2> "node_name":"127.0.0.1:41386_",
[junit4] 2> "state":"down",
[junit4] 2> "type":"NRT"}}}}}
[junit4] 2> 1064180 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ZkStateWriter going to update_collection
/collections/control_collection/state.json version: 0
[junit4] 2> 1065090 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1065102 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test
[junit4] 2> 1065177 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
[junit4] 2> 1065189 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1'
using configuration from collection control_collection, trusted=true
[junit4] 2> 1065190 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.core.control_collection.shard1.replica_n1' (registry
'solr.core.control_collection.shard1.replica_n1') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@32eae29b
[junit4] 2> 1065190 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore
at
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/control-001/cores/control_collection_shard1_replica_n1],
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/control-001/cores/control_collection_shard1_replica_n1/data/]
[junit4] 2> 1065193 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=38, maxMergeAtOnceExplicit=31, maxMergedSegmentMB=82.66796875,
floorSegmentMB=0.6572265625, forceMergeDeletesPctAllowed=9.293648791015867,
segmentsPerTier=25.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.0,
deletesPctAllowed=24.691808265826015
[junit4] 2> 1065198 WARN (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type =
requestHandler,name = /dump,class = DumpRequestHandler,attributes =
{initParams=a, name=/dump, class=DumpRequestHandler},args =
{defaults={a=A,b=B}}}
[junit4] 2> 1065238 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.UpdateHandler Using UpdateLog implementation:
org.apache.solr.update.UpdateLog
[junit4] 2> 1065238 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH
numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1065239 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1065239 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1065240 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=44, maxMergeAtOnceExplicit=46, maxMergedSegmentMB=76.6865234375,
floorSegmentMB=0.3447265625, forceMergeDeletesPctAllowed=6.457592487277185,
segmentsPerTier=27.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.5517682385712898, deletesPctAllowed=44.30343685123432
[junit4] 2> 1065241 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.SolrIndexSearcher Opening
[Searcher@7bcb7d09[control_collection_shard1_replica_n1] main]
[junit4] 2> 1065247 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 1065247 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1065248 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1065248 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.u.UpdateLog Could not find max version in index or recent updates, using
new clock 1613405820645539840
[junit4] 2> 1065252 INFO
(searcherExecutor-4383-thread-1-processing-n:127.0.0.1:41386_
x:control_collection_shard1_replica_n1 c:control_collection s:shard1)
[n:127.0.0.1:41386_ c:control_collection s:shard1
x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore
[control_collection_shard1_replica_n1] Registered new searcher
Searcher@7bcb7d09[control_collection_shard1_replica_n1]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1065253 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ZkShardTerms Successful update of terms at
/collections/control_collection/terms/shard1 to Terms{values={core_node2=0},
version=0}
[junit4] 2> 1065256 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1065256 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1065256 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SyncStrategy Sync replicas to
https://127.0.0.1:41386/control_collection_shard1_replica_n1/
[junit4] 2> 1065256 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 1065256 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.SyncStrategy
https://127.0.0.1:41386/control_collection_shard1_replica_n1/ has no replicas
[junit4] 2> 1065256 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in
election, clear LIR
[junit4] 2> 1065259 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ShardLeaderElectionContext I am the new leader:
https://127.0.0.1:41386/control_collection_shard1_replica_n1/ shard1
[junit4] 2> 1065259 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.Overseer processMessage: queueSize: 1, message
= {
[junit4] 2> "operation":"leader",
[junit4] 2> "shard":"shard1",
[junit4] 2> "collection":"control_collection",
[junit4] 2> "base_url":"https://127.0.0.1:41386",
[junit4] 2> "core":"control_collection_shard1_replica_n1",
[junit4] 2> "state":"active"} current state version: 0
[junit4] 2> 1065360 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ZkStateWriter going to update_collection
/collections/control_collection/state.json version: 1
[junit4] 2> 1065411 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1065413 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.Overseer processMessage: queueSize: 1, message
= {
[junit4] 2> "core":"control_collection_shard1_replica_n1",
[junit4] 2> "core_node_name":"core_node2",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"https://127.0.0.1:41386",
[junit4] 2> "node_name":"127.0.0.1:41386_",
[junit4] 2> "numShards":"1",
[junit4] 2> "state":"active",
[junit4] 2> "shard":"shard1",
[junit4] 2> "collection":"control_collection",
[junit4] 2> "type":"NRT",
[junit4] 2> "operation":"state"} current state version: 0
[junit4] 2> 1065413 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Update state numShards=1
message={
[junit4] 2> "core":"control_collection_shard1_replica_n1",
[junit4] 2> "core_node_name":"core_node2",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"https://127.0.0.1:41386",
[junit4] 2> "node_name":"127.0.0.1:41386_",
[junit4] 2> "numShards":"1",
[junit4] 2> "state":"active",
[junit4] 2> "shard":"shard1",
[junit4] 2> "collection":"control_collection",
[junit4] 2> "type":"NRT",
[junit4] 2> "operation":"state"}
[junit4] 2> 1065413 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Will update state for
replica:
core_node2:{"core":"control_collection_shard1_replica_n1","base_url":"https://127.0.0.1:41386","node_name":"127.0.0.1:41386_","state":"active","type":"NRT","leader":"true"}
[junit4] 2> 1065413 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Collection is now:
DocCollection(control_collection//collections/control_collection/state.json/2)={
[junit4] 2> "pullReplicas":"0",
[junit4] 2> "replicationFactor":"1",
[junit4] 2> "router":{"name":"compositeId"},
[junit4] 2> "maxShardsPerNode":"1",
[junit4] 2> "autoAddReplicas":"false",
[junit4] 2> "nrtReplicas":"1",
[junit4] 2> "tlogReplicas":"0",
[junit4] 2> "shards":{"shard1":{
[junit4] 2> "range":"80000000-7fffffff",
[junit4] 2> "state":"active",
[junit4] 2> "replicas":{"core_node2":{
[junit4] 2> "core":"control_collection_shard1_replica_n1",
[junit4] 2> "base_url":"https://127.0.0.1:41386",
[junit4] 2> "node_name":"127.0.0.1:41386_",
[junit4] 2> "state":"active",
[junit4] 2> "type":"NRT",
[junit4] 2> "leader":"true"}}}}}
[junit4] 2> 1065414 INFO (qtp1635669701-15077) [n:127.0.0.1:41386_
c:control_collection s:shard1 x:control_collection_shard1_replica_n1]
o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores
params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT}
status=0 QTime=1339
[junit4] 2> 1065416 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.CreateCollectionCmd Finished create
command on all shards for collection: control_collection
[junit4] 2> 1065416 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Completed
task:[/overseer/collection-queue-work/qn-0000000000]
[junit4] 2> 1065417 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Marked task
[/overseer/collection-queue-work/qn-0000000000] as completed.
[junit4] 2> 1065417 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningTasks: []
[junit4] 2> 1065417 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
[junit4] 2> 1065417 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor CompletedTasks:
[/overseer/collection-queue-work/qn-0000000000]
[junit4] 2> 1065417 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningZKTasks:
[/overseer/collection-queue-work/qn-0000000000]
[junit4] 2> 1065417 DEBUG
(OverseerThreadFactory-4378-thread-1-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Overseer Collection
Message Handler: Message id:/overseer/collection-queue-work/qn-0000000000
complete,
response:{success={127.0.0.1:41386_={responseHeader={status=0,QTime=1339},core=control_collection_shard1_replica_n1}}}
[junit4] 2> 1065418 INFO (qtp1635669701-15079) [n:127.0.0.1:41386_ ]
o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most
30 seconds. Check all shard replicas
[junit4] 2> 1065516 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ZkStateWriter going to update_collection
/collections/control_collection/state.json version: 2
[junit4] 2> 1065517 INFO (zkCallback-7198-thread-1) [ ]
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent
state:SyncConnected type:NodeDataChanged
path:/collections/control_collection/state.json] for collection
[control_collection] has occurred - updating... (live nodes size: [1])
[junit4] 2> 1065959 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Got 0 tasks from
work-queue : [[]]
[junit4] 2> 1065959 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Cleaning up work-queue.
#Running tasks: 0
[junit4] 2> 1065959 INFO
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskQueue Response ZK path:
/overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may
have disconnected from ZooKeeper
[junit4] 2> 1065960 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningTasks: []
[junit4] 2> 1065960 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
[junit4] 2> 1065960 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor CompletedTasks: []
[junit4] 2> 1065960 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningZKTasks: []
[junit4] 2> 1066418 INFO (qtp1635669701-15079) [n:127.0.0.1:41386_ ]
o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections
params={collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:41386_&wt=javabin&version=2}
status=0 QTime=2468
[junit4] 2> 1066422 INFO (zkConnectionManagerCallback-7217-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066423 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1066424 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:43542/solr ready
[junit4] 2> 1066424 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection
loss:false
[junit4] 2> 1066426 INFO (qtp1635669701-15079) [n:127.0.0.1:41386_ ]
o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params
collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=2&createNodeSet=&stateFormat=2&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 1066427 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Got 1 tasks from
work-queue : [[org.apache.solr.cloud.OverseerTaskQueue$QueueEvent@2e684ed1]]
[junit4] 2> 1066428 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Marked task
[/overseer/collection-queue-work/qn-0000000002] as running
[junit4] 2> 1066428 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Overseer Collection
Message Handler: Get the message
id:/overseer/collection-queue-work/qn-0000000002 message:{
[junit4] 2> "name":"collection1",
[junit4] 2> "fromApi":"true",
[junit4] 2> "collection.configName":"conf1",
[junit4] 2> "numShards":"2",
[junit4] 2> "createNodeSet":"",
[junit4] 2> "stateFormat":"2",
[junit4] 2> "nrtReplicas":"1",
[junit4] 2> "replicationFactor":"1",
[junit4] 2> "operation":"create"}
[junit4] 2> 1066428 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Runner processing
/overseer/collection-queue-work/qn-0000000002
[junit4] 2> 1066428 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.OverseerCollectionMessageHandler
OverseerCollectionMessageHandler.processMessage : create , {
[junit4] 2> "name":"collection1",
[junit4] 2> "fromApi":"true",
[junit4] 2> "collection.configName":"conf1",
[junit4] 2> "numShards":"2",
[junit4] 2> "createNodeSet":"",
[junit4] 2> "stateFormat":"2",
[junit4] 2> "nrtReplicas":"1",
[junit4] 2> "replicationFactor":"1",
[junit4] 2> "operation":"create"}
[junit4] 2> 1066428 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Cleaning up work-queue.
#Running tasks: 1
[junit4] 2> 1066428 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningTasks:
[/overseer/collection-queue-work/qn-0000000002]
[junit4] 2> 1066428 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
[junit4] 2> 1066428 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor CompletedTasks: []
[junit4] 2> 1066428 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningZKTasks:
[/overseer/collection-queue-work/qn-0000000002]
[junit4] 2> 1066428 INFO
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.CreateCollectionCmd Create collection
collection1
[junit4] 2> 1066429 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.OverseerCollectionMessageHandler creating
collections conf node /collections/collection1
[junit4] 2> 1066429 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.CreateCollectionCmd Check for collection
zkNode:collection1
[junit4] 2> 1066430 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.CreateCollectionCmd Collection zkNode
exists
[junit4] 2> 1066431 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.Overseer processMessage: queueSize: 1, message
= {
[junit4] 2> "name":"collection1",
[junit4] 2> "fromApi":"true",
[junit4] 2> "collection.configName":"conf1",
[junit4] 2> "numShards":"2",
[junit4] 2> "createNodeSet":"",
[junit4] 2> "stateFormat":"2",
[junit4] 2> "nrtReplicas":"1",
[junit4] 2> "replicationFactor":"1",
[junit4] 2> "operation":"create"} current state version: 0
[junit4] 2> 1066431 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ClusterStateMutator building a new cName:
collection1
[junit4] 2> 1066532 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ZkStateWriter going to create_collection
/collections/collection1/state.json
[junit4] 2> 1066632 WARN
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.CreateCollectionCmd It is unusual to
create a collection (collection1) without cores.
[junit4] 2> 1066633 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.a.c.CreateCollectionCmd Finished create
command for collection: collection1
[junit4] 2> 1066633 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Completed
task:[/overseer/collection-queue-work/qn-0000000002]
[junit4] 2> 1066634 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Marked task
[/overseer/collection-queue-work/qn-0000000002] as completed.
[junit4] 2> 1066634 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningTasks: []
[junit4] 2> 1066634 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
[junit4] 2> 1066634 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor CompletedTasks:
[/overseer/collection-queue-work/qn-0000000002]
[junit4] 2> 1066634 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningZKTasks:
[/overseer/collection-queue-work/qn-0000000002]
[junit4] 2> 1066634 DEBUG
(OverseerThreadFactory-4378-thread-2-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Overseer Collection
Message Handler: Message id:/overseer/collection-queue-work/qn-0000000002
complete, response:{}
[junit4] 2> 1066634 INFO (qtp1635669701-15079) [n:127.0.0.1:41386_ ]
o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most
30 seconds. Check all shard replicas
[junit4] 2> 1066635 INFO (qtp1635669701-15079) [n:127.0.0.1:41386_ ]
o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections
params={collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=2&createNodeSet=&stateFormat=2&wt=javabin&version=2}
status=0 QTime=209
[junit4] 2> 1066736 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/shard-1-001
of type NRT
[junit4] 2> 1066737 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git:
d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
[junit4] 2> 1066738 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1066738 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1066738 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 1066739 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@38cb6036{/,null,AVAILABLE}
[junit4] 2> 1066741 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.AbstractConnector Started ServerConnector@332f1ef4{SSL,[ssl,
http/1.1]}{127.0.0.1:40685}
[junit4] 2> 1066741 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.e.j.s.Server Started @1066833ms
[junit4] 2> 1066741 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/tempDir-001/jetty1,
replicaType=NRT, solrconfig=solrconfig.xml, hostContext=/, hostPort=40685,
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/shard-1-001/cores}
[junit4] 2> 1066742 ERROR
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 1066742 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter Using logger factory
org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1066742 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.6.0
[junit4] 2> 1066742 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1066742 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1066742 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2018-10-04T14:34:54.584Z
[junit4] 2> 1066743 INFO (zkConnectionManagerCallback-7219-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066744 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 1066744 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/shard-1-001/solr.xml
[junit4] 2> 1066747 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay
is ignored
[junit4] 2> 1066747 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.SolrXmlConfig Configuration parameter
autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 1066749 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.SolrXmlConfig MBean server found:
com.sun.jmx.mbeanserver.JmxMBeanServer@32eae29b, but no JMX reporters were
configured - adding default JMX reporter.
[junit4] 2> 1066963 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:43542/solr
[junit4] 2> 1066964 INFO (zkConnectionManagerCallback-7223-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066965 INFO (zkConnectionManagerCallback-7225-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1066970 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1066974 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4
transient cores
[junit4] 2> 1066974 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:40685_
[junit4] 2> 1066974 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.Overseer processMessage: queueSize: 1, message
= {
[junit4] 2> "operation":"downnode",
[junit4] 2> "node_name":"127.0.0.1:40685_"} current state version: 0
[junit4] 2> 1066974 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.NodeMutator DownNode state invoked for node:
127.0.0.1:40685_
[junit4] 2> 1066975 INFO (zkCallback-7198-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1066975 INFO (zkCallback-7205-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1066984 INFO (zkCallback-7216-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1066985 INFO (zkCallback-7224-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1067002 INFO (zkConnectionManagerCallback-7232-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1067003 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1067003 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:43542/solr ready
[junit4] 2> 1067004 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history
in memory.
[junit4] 2> 1067020 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node')
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32eae29b
[junit4] 2> 1067029 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm')
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32eae29b
[junit4] 2> 1067029 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry
'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32eae29b
[junit4] 2> 1067030 INFO
(TEST-ShardSplitTest.testSplitMixedReplicaTypes-seed#[9223DC651078BC8A]) [ ]
o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/shard-1-001/cores
[junit4] 2> 1067072 INFO (qtp771908827-15135) [n:127.0.0.1:40685_ ]
o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params
node=127.0.0.1:40685_&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 1067073 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Got 1 tasks from
work-queue : [[org.apache.solr.cloud.OverseerTaskQueue$QueueEvent@2e684ed3]]
[junit4] 2> 1067074 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Marked task
[/overseer/collection-queue-work/qn-0000000004] as running
[junit4] 2> 1067074 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Overseer Collection
Message Handler: Get the message
id:/overseer/collection-queue-work/qn-0000000004 message:{
[junit4] 2> "collection":"collection1",
[junit4] 2> "node":"127.0.0.1:40685_",
[junit4] 2> "shard":"shard2",
[junit4] 2> "type":"NRT",
[junit4] 2> "operation":"addreplica"}
[junit4] 2> 1067074 DEBUG
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Runner processing
/overseer/collection-queue-work/qn-0000000004
[junit4] 2> 1067074 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor Cleaning up work-queue.
#Running tasks: 1
[junit4] 2> 1067074 DEBUG
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ c:collection1 s:shard2 ]
o.a.s.c.a.c.OverseerCollectionMessageHandler
OverseerCollectionMessageHandler.processMessage : addreplica , {
[junit4] 2> "collection":"collection1",
[junit4] 2> "node":"127.0.0.1:40685_",
[junit4] 2> "shard":"shard2",
[junit4] 2> "type":"NRT",
[junit4] 2> "operation":"addreplica"}
[junit4] 2> 1067074 DEBUG
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ c:collection1 s:shard2 ] o.a.s.c.a.c.AddReplicaCmd
addReplica() : {
[junit4] 2> "collection":"collection1",
[junit4] 2> "node":"127.0.0.1:40685_",
[junit4] 2> "shard":"shard2",
[junit4] 2> "type":"NRT",
[junit4] 2> "operation":"addreplica"}
[junit4] 2> 1067075 INFO
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskQueue Response ZK path:
/overseer/collection-queue-work/qnr-0000000002 doesn't exist. Requestor may
have disconnected from ZooKeeper
[junit4] 2> 1067075 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningTasks:
[/overseer/collection-queue-work/qn-0000000004]
[junit4] 2> 1067075 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
[junit4] 2> 1067075 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor CompletedTasks: []
[junit4] 2> 1067075 DEBUG
(OverseerCollectionConfigSetProcessor-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.OverseerTaskProcessor RunningZKTasks:
[/overseer/collection-queue-work/qn-0000000004]
[junit4] 2> 1067075 DEBUG
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ c:collection1 s:shard2 ] o.a.s.c.a.c.Assign
getNodesForNewReplicas() shard: shard2 , nrtReplicas : 1 , tlogReplicas: 0 ,
pullReplicas: 0 , createNodeSet 127.0.0.1:40685_
[junit4] 2> 1067076 INFO
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ c:collection1 s:shard2 ] o.a.s.c.a.c.AddReplicaCmd Node
Identified 127.0.0.1:40685_ for creating new replica of shard shard2 for
collection collection1
[junit4] 2> 1067080 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
x:collection1_shard2_replica_n41] o.a.s.h.a.CoreAdminOperation core create
command
qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n41&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1067084 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.Overseer processMessage: queueSize: 1, message
= {
[junit4] 2> "core":"collection1_shard2_replica_n41",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"https://127.0.0.1:40685",
[junit4] 2> "node_name":"127.0.0.1:40685_",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"down",
[junit4] 2> "shard":"shard2",
[junit4] 2> "collection":"collection1",
[junit4] 2> "type":"NRT",
[junit4] 2> "operation":"state"} current state version: 0
[junit4] 2> 1067084 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Update state numShards=2
message={
[junit4] 2> "core":"collection1_shard2_replica_n41",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"https://127.0.0.1:40685",
[junit4] 2> "node_name":"127.0.0.1:40685_",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"down",
[junit4] 2> "shard":"shard2",
[junit4] 2> "collection":"collection1",
[junit4] 2> "type":"NRT",
[junit4] 2> "operation":"state"}
[junit4] 2> 1067085 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Will update state for
replica:
core_node42:{"core":"collection1_shard2_replica_n41","base_url":"https://127.0.0.1:40685","node_name":"127.0.0.1:40685_","state":"down","type":"NRT"}
[junit4] 2> 1067085 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Collection is now:
DocCollection(collection1//collections/collection1/state.json/0)={
[junit4] 2> "pullReplicas":"0",
[junit4] 2> "replicationFactor":"1",
[junit4] 2> "router":{"name":"compositeId"},
[junit4] 2> "maxShardsPerNode":"1",
[junit4] 2> "autoAddReplicas":"false",
[junit4] 2> "nrtReplicas":"1",
[junit4] 2> "tlogReplicas":"0",
[junit4] 2> "shards":{
[junit4] 2> "shard1":{
[junit4] 2> "range":"80000000-ffffffff",
[junit4] 2> "state":"active",
[junit4] 2> "replicas":{}},
[junit4] 2> "shard2":{
[junit4] 2> "range":"0-7fffffff",
[junit4] 2> "state":"active",
[junit4] 2> "replicas":{"core_node42":{
[junit4] 2> "core":"collection1_shard2_replica_n41",
[junit4] 2> "base_url":"https://127.0.0.1:40685",
[junit4] 2> "node_name":"127.0.0.1:40685_",
[junit4] 2> "state":"down",
[junit4] 2> "type":"NRT"}}}}}
[junit4] 2> 1067186 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ZkStateWriter going to update_collection
/collections/collection1/state.json version: 0
[junit4] 2> 1068096 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SolrConfig
Using Lucene MatchVersion: 7.6.0
[junit4] 2> 1068118 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.s.IndexSchema
[collection1_shard2_replica_n41] Schema name=test
[junit4] 2> 1068179 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.s.IndexSchema
Loaded schema test/1.6 with uniqueid field id
[junit4] 2> 1068192 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.CoreContainer
Creating SolrCore 'collection1_shard2_replica_n41' using configuration from
collection collection1, trusted=true
[junit4] 2> 1068193 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.core.collection1.shard2.replica_n41' (registry
'solr.core.collection1.shard2.replica_n41') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@32eae29b
[junit4] 2> 1068193 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SolrCore
[[collection1_shard2_replica_n41] ] Opening new SolrCore at
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/shard-1-001/cores/collection1_shard2_replica_n41],
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_9223DC651078BC8A-001/shard-1-001/cores/collection1_shard2_replica_n41/data/]
[junit4] 2> 1068196 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=38, maxMergeAtOnceExplicit=31, maxMergedSegmentMB=82.66796875,
floorSegmentMB=0.6572265625, forceMergeDeletesPctAllowed=9.293648791015867,
segmentsPerTier=25.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.0,
deletesPctAllowed=24.691808265826015
[junit4] 2> 1068205 WARN (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type =
requestHandler,name = /dump,class = DumpRequestHandler,attributes =
{initParams=a, name=/dump, class=DumpRequestHandler},args =
{defaults={a=A,b=B}}}
[junit4] 2> 1068268 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.UpdateHandler
Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1068268 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.UpdateLog
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1068269 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.CommitTracker
Hard AutoCommit: disabled
[junit4] 2> 1068269 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.CommitTracker
Soft AutoCommit: disabled
[junit4] 2> 1068270 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=44, maxMergeAtOnceExplicit=46, maxMergedSegmentMB=76.6865234375,
floorSegmentMB=0.3447265625, forceMergeDeletesPctAllowed=6.457592487277185,
segmentsPerTier=27.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.5517682385712898, deletesPctAllowed=44.30343685123432
[junit4] 2> 1068271 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.s.SolrIndexSearcher Opening
[Searcher@5b8af9d[collection1_shard2_replica_n41] main]
[junit4] 2> 1068278 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 1068279 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1068279 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 1068279 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.u.UpdateLog
Could not find max version in index or recent updates, using new clock
1613405823823773696
[junit4] 2> 1068282 INFO
(searcherExecutor-4397-thread-1-processing-n:127.0.0.1:40685_
x:collection1_shard2_replica_n41 c:collection1 s:shard2) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SolrCore
[collection1_shard2_replica_n41] Registered new searcher
Searcher@5b8af9d[collection1_shard2_replica_n41]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1068290 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.ZkShardTerms
Successful update of terms at /collections/collection1/terms/shard2 to
Terms{values={core_node42=0}, version=0}
[junit4] 2> 1068293 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 1068293 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 1068293 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy
Sync replicas to https://127.0.0.1:40685/collection1_shard2_replica_n41/
[junit4] 2> 1068293 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy
Sync Success - now sync replicas to me
[junit4] 2> 1068293 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy
https://127.0.0.1:40685/collection1_shard2_replica_n41/ has no replicas
[junit4] 2> 1068293 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in
election, clear LIR
[junit4] 2> 1068296 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41]
o.a.s.c.ShardLeaderElectionContext I am the new leader:
https://127.0.0.1:40685/collection1_shard2_replica_n41/ shard2
[junit4] 2> 1068297 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.Overseer processMessage: queueSize: 1, message
= {
[junit4] 2> "operation":"leader",
[junit4] 2> "shard":"shard2",
[junit4] 2> "collection":"collection1",
[junit4] 2> "base_url":"https://127.0.0.1:40685",
[junit4] 2> "core":"collection1_shard2_replica_n41",
[junit4] 2> "state":"active"} current state version: 0
[junit4] 2> 1068398 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ZkStateWriter going to update_collection
/collections/collection1/state.json version: 1
[junit4] 2> 1068399 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.c.ZkController
I am the leader, no recovery necessary
[junit4] 2> 1068401 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.Overseer processMessage: queueSize: 1, message
= {
[junit4] 2> "core":"collection1_shard2_replica_n41",
[junit4] 2> "core_node_name":"core_node42",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"https://127.0.0.1:40685",
[junit4] 2> "node_name":"127.0.0.1:40685_",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"active",
[junit4] 2> "shard":"shard2",
[junit4] 2> "collection":"collection1",
[junit4] 2> "type":"NRT",
[junit4] 2> "operation":"state"} current state version: 0
[junit4] 2> 1068401 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Update state numShards=2
message={
[junit4] 2> "core":"collection1_shard2_replica_n41",
[junit4] 2> "core_node_name":"core_node42",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"https://127.0.0.1:40685",
[junit4] 2> "node_name":"127.0.0.1:40685_",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"active",
[junit4] 2> "shard":"shard2",
[junit4] 2> "collection":"collection1",
[junit4] 2> "type":"NRT",
[junit4] 2> "operation":"state"}
[junit4] 2> 1068401 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Will update state for
replica:
core_node42:{"core":"collection1_shard2_replica_n41","base_url":"https://127.0.0.1:40685","node_name":"127.0.0.1:40685_","state":"active","type":"NRT","leader":"true"}
[junit4] 2> 1068401 DEBUG
(OverseerStateUpdate-72447691000643589-127.0.0.1:41386_-n_0000000000)
[n:127.0.0.1:41386_ ] o.a.s.c.o.ReplicaMutator Collection is now:
DocCollection(collection1//collections/collection1/state.json/2)={
[junit4] 2> "pullReplicas":"0",
[junit4] 2> "replicationFactor":"1",
[junit4] 2> "router":{"name":"compositeId"},
[junit4] 2> "maxShardsPerNode":"1",
[junit4] 2> "autoAddReplicas":"false",
[junit4] 2> "nrtReplicas":"1",
[junit4] 2> "tlogReplicas":"0",
[junit4] 2> "shards":{
[junit4] 2> "shard1":{
[junit4] 2> "range":"80000000-ffffffff",
[junit4] 2> "state":"active",
[junit4] 2> "replicas":{}},
[junit4] 2> "shard2":{
[junit4] 2> "range":"0-7fffffff",
[junit4] 2> "state":"active",
[junit4] 2> "replicas":{"core_node42":{
[junit4] 2> "core":"collection1_shard2_replica_n41",
[junit4] 2> "base_url":"https://127.0.0.1:40685",
[junit4] 2> "node_name":"127.0.0.1:40685_",
[junit4] 2> "state":"active",
[junit4] 2> "type":"NRT",
[junit4] 2> "leader":"true"}}}}}
[junit4] 2> 1068403 INFO (qtp771908827-15134) [n:127.0.0.1:40685_
c:collection1 s:shard2 x:collection1_shard2_replica_n41] o.a.s.s.HttpSolrCall
[admin] webapp=null path=/admin/cores
params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n41&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT}
status=0 QTime=1323
[junit4] 2> 1068404 DEBUG
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ c:collection1 s:shard2 ] o.a.s.c.OverseerTaskProcessor
Completed task:[/overseer/collection-queue-work/qn-0000000004]
[junit4] 2> 1068405 DEBUG
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ c:collection1 s:shard2 ] o.a.s.c.OverseerTaskProcessor
Marked task [/overseer/collection-queue-work/qn-0000000004] as completed.
[junit4] 2> 1068405 DEBUG
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ c:collection1 s:shard2 ] o.a.s.c.OverseerTaskProcessor
RunningTasks: []
[junit4] 2> 1068405 DEBUG
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ c:collection1 s:shard2 ] o.a.s.c.OverseerTaskProcessor
BlockedTasks: []
[junit4] 2> 1068405 DEBUG
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ c:collection1 s:shard2 ] o.a.s.c.OverseerTaskProcessor
CompletedTasks: [/overseer/collection-queue-work/qn-0000000004]
[junit4] 2> 1068405 DEBUG
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ c:collection1 s:shard2 ] o.a.s.c.OverseerTaskProcessor
RunningZKTasks: [/overseer/collection-queue-work/qn-0000000004]
[junit4] 2> 1068405 DEBUG
(OverseerThreadFactory-4378-thread-3-processing-n:127.0.0.1:41386_)
[n:127.0.0.1:41386_ c:collection1 s:shard2 ] o.a.s.c.OverseerTaskProcessor
Overseer Collection Message Handler: Message
id:/overseer/collection-queue-work/qn-0000000004 complete,
response:{success={127.0.0.1:40685_={responseHeader={status=0,QTime=1323},core=collection1_shard2_replica_n41}}}
[junit4] 2> 1068406 INFO (qtp771908827-15135) [n:127
[...truncated too long message...]
ins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/search/TestSearcherManager.java
(at line 313)
[ecj-lint] SearcherManager sm = new SearcherManager(writer, false, false,
new SearcherFactory());
[ecj-lint] ^^
[ecj-lint] Resource leak: 'sm' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 71. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/search/TestTermQuery.java
(at line 52)
[ecj-lint] new TermQuery(new Term("foo", "bar"), TermContext.build(new
MultiReader().getContext(), new Term("foo", "bar"))));
[ecj-lint]
^^^^^^^^^^^^^^^^^
[ecj-lint] Resource leak: '<unassigned Closeable value>' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 72. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/store/TestBufferedIndexInput.java
(at line 50)
[ecj-lint] MyBufferedIndexInput input = new MyBufferedIndexInput();
[ecj-lint] ^^^^^
[ecj-lint] Resource leak: 'input' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 73. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/store/TestHugeRamFile.java
(at line 88)
[ecj-lint] RAMInputStream in = new RAMInputStream("testcase", f);
[ecj-lint] ^^
[ecj-lint] Resource leak: 'in' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 74. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/store/TestRAMDirectory.java
(at line 81)
[ecj-lint] RAMDirectory ramDir = new RAMDirectory(fsDir,
newIOContext(random()));
[ecj-lint] ^^^^^^
[ecj-lint] Resource leak: 'ramDir' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 75. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/store/TestTrackingDirectoryWrapper.java
(at line 32)
[ecj-lint] TrackingDirectoryWrapper dir = new TrackingDirectoryWrapper(new
RAMDirectory());
[ecj-lint] ^^^
[ecj-lint] Resource leak: 'dir' is never closed
[ecj-lint] ----------
[ecj-lint] 76. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/store/TestTrackingDirectoryWrapper.java
(at line 37)
[ecj-lint] TrackingDirectoryWrapper dir = new TrackingDirectoryWrapper(new
RAMDirectory());
[ecj-lint] ^^^
[ecj-lint] Resource leak: 'dir' is never closed
[ecj-lint] ----------
[ecj-lint] 77. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/store/TestTrackingDirectoryWrapper.java
(at line 43)
[ecj-lint] TrackingDirectoryWrapper dir = new TrackingDirectoryWrapper(new
RAMDirectory());
[ecj-lint] ^^^
[ecj-lint] Resource leak: 'dir' is never closed
[ecj-lint] ----------
[ecj-lint] 78. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/store/TestTrackingDirectoryWrapper.java
(at line 51)
[ecj-lint] TrackingDirectoryWrapper dir = new TrackingDirectoryWrapper(new
RAMDirectory());
[ecj-lint] ^^^
[ecj-lint] Resource leak: 'dir' is never closed
[ecj-lint] ----------
[ecj-lint] 79. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/store/TestTrackingDirectoryWrapper.java
(at line 60)
[ecj-lint] TrackingDirectoryWrapper dest = new
TrackingDirectoryWrapper(new RAMDirectory());
[ecj-lint] ^^^^
[ecj-lint] Resource leak: 'dest' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 80. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/TestCloseableThreadLocal.java
(at line 31)
[ecj-lint] CloseableThreadLocal<Object> ctl = new CloseableThreadLocal<>();
[ecj-lint] ^^^
[ecj-lint] Resource leak: 'ctl' is never closed
[ecj-lint] ----------
[ecj-lint] 81. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/TestCloseableThreadLocal.java
(at line 39)
[ecj-lint] CloseableThreadLocal<Object> ctl = new CloseableThreadLocal<>();
[ecj-lint] ^^^
[ecj-lint] Resource leak: 'ctl' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 82. ERROR in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/TestQueryBuilder.java
(at line 21)
[ecj-lint] import java.io.Reader;
[ecj-lint] ^^^^^^^^^^^^^^
[ecj-lint] The import java.io.Reader is never used
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 83. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/bkd/Test2BBKDPoints.java
(at line 44)
[ecj-lint] BKDWriter w = new BKDWriter(numDocs, dir, "_0", 1, Long.BYTES,
[ecj-lint] ^
[ecj-lint] Resource leak: 'w' is never closed
[ecj-lint] ----------
[ecj-lint] 84. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/bkd/Test2BBKDPoints.java
(at line 81)
[ecj-lint] BKDWriter w = new BKDWriter(numDocs, dir, "_0", 2, Long.BYTES,
[ecj-lint] ^
[ecj-lint] Resource leak: 'w' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 85. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/bkd/TestBKD.java
(at line 51)
[ecj-lint] BKDWriter w = new BKDWriter(100, dir, "tmp", 1, 4, 2, 1.0f,
100, true);
[ecj-lint] ^
[ecj-lint] Resource leak: 'w' is never closed
[ecj-lint] ----------
[ecj-lint] 86. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/bkd/TestBKD.java
(at line 126)
[ecj-lint] BKDWriter w = new BKDWriter(numDocs, dir, "tmp", numDims, 4,
maxPointsInLeafNode, maxMB, numDocs, true);
[ecj-lint] ^
[ecj-lint] Resource leak: 'w' is never closed
[ecj-lint] ----------
[ecj-lint] 87. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/bkd/TestBKD.java
(at line 267)
[ecj-lint] BKDWriter w = new BKDWriter(numDocs, dir, "tmp", numDims,
numBytesPerDim, maxPointsInLeafNode, maxMB, numDocs, true);
[ecj-lint] ^
[ecj-lint] Resource leak: 'w' is never closed
[ecj-lint] ----------
[ecj-lint] 88. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/bkd/TestBKD.java
(at line 715)
[ecj-lint] w = new BKDWriter(numValues, dir, "_" + seg, numDims,
numBytesPerDim, maxPointsInLeafNode, maxMB, docValues.length, false);
[ecj-lint]
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[ecj-lint] Resource leak: 'w' is not closed at this location
[ecj-lint] ----------
[ecj-lint] 89. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/bkd/TestBKD.java
(at line 974)
[ecj-lint] BKDWriter w = new BKDWriter(numDocs+1, dir, "tmp", 1,
Integer.BYTES, 2, 0.01f, numDocs, true);
[ecj-lint] ^
[ecj-lint] Resource leak: 'w' is never closed
[ecj-lint] ----------
[ecj-lint] 90. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/bkd/TestBKD.java
(at line 1015)
[ecj-lint] BKDWriter w = new BKDWriter(numDocs+1, dir, "tmp", 2,
Integer.BYTES, 2, 0.01f, numDocs,
[ecj-lint] ^
[ecj-lint] Resource leak: 'w' is never closed
[ecj-lint] ----------
[ecj-lint] 91. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/bkd/TestBKD.java
(at line 1066)
[ecj-lint] BKDWriter w = new BKDWriter(numDocs+1, dir, "tmp", numDims,
bytesPerDim, 32, 1f, numDocs, true);
[ecj-lint] ^
[ecj-lint] Resource leak: 'w' is never closed
[ecj-lint] ----------
[ecj-lint] ----------
[ecj-lint] 92. WARNING in
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/core/src/test/org/apache/lucene/util/fst/TestFSTs.java
(at line 311)
[ecj-lint] final LineFileDocs docs = new LineFileDocs(random());
[ecj-lint] ^^^^
[ecj-lint] Resource leak: 'docs' is never closed
[ecj-lint] ----------
[ecj-lint] 92 problems (1 error, 91 warnings)
BUILD FAILED
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/build.xml:642:
The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/build.xml:101:
The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/build.xml:202:
The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/common-build.xml:2092:
The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-BadApples-Tests-7.x/lucene/common-build.xml:2125:
Compile failed; see the compiler error output for details.
Total time: 114 minutes 7 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]