Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/19080/
Java: 64bit/jdk1.8.0_121 -XX:+UseCompressedOops -XX:+UseParallelGC
1 tests failed.
FAILED: org.apache.solr.cloud.LeaderFailoverAfterPartitionTest.test
Error Message:
Expected 2 of 3 replicas to be active but only found 1;
[core_node3:{"core":"c8n_1x3_lf_shard1_replica2","base_url":"http://127.0.0.1:43964/sj_a","node_name":"127.0.0.1:43964_sj_a","state":"active","leader":"true"}];
clusterState: DocCollection(c8n_1x3_lf//clusterstate.json/33)={
"replicationFactor":"3", "shards":{"shard1":{
"range":"80000000-7fffffff", "state":"active", "replicas":{
"core_node1":{ "core":"c8n_1x3_lf_shard1_replica1",
"base_url":"http://127.0.0.1:33369/sj_a",
"node_name":"127.0.0.1:33369_sj_a", "state":"down"},
"core_node2":{ "state":"down",
"base_url":"http://127.0.0.1:36651/sj_a",
"core":"c8n_1x3_lf_shard1_replica3",
"node_name":"127.0.0.1:36651_sj_a"}, "core_node3":{
"core":"c8n_1x3_lf_shard1_replica2",
"base_url":"http://127.0.0.1:43964/sj_a",
"node_name":"127.0.0.1:43964_sj_a", "state":"active",
"leader":"true"}}}}, "router":{"name":"compositeId"},
"maxShardsPerNode":"1", "autoAddReplicas":"false"}
Stack Trace:
java.lang.AssertionError: Expected 2 of 3 replicas to be active but only found
1;
[core_node3:{"core":"c8n_1x3_lf_shard1_replica2","base_url":"http://127.0.0.1:43964/sj_a","node_name":"127.0.0.1:43964_sj_a","state":"active","leader":"true"}];
clusterState: DocCollection(c8n_1x3_lf//clusterstate.json/33)={
"replicationFactor":"3",
"shards":{"shard1":{
"range":"80000000-7fffffff",
"state":"active",
"replicas":{
"core_node1":{
"core":"c8n_1x3_lf_shard1_replica1",
"base_url":"http://127.0.0.1:33369/sj_a",
"node_name":"127.0.0.1:33369_sj_a",
"state":"down"},
"core_node2":{
"state":"down",
"base_url":"http://127.0.0.1:36651/sj_a",
"core":"c8n_1x3_lf_shard1_replica3",
"node_name":"127.0.0.1:36651_sj_a"},
"core_node3":{
"core":"c8n_1x3_lf_shard1_replica2",
"base_url":"http://127.0.0.1:43964/sj_a",
"node_name":"127.0.0.1:43964_sj_a",
"state":"active",
"leader":"true"}}}},
"router":{"name":"compositeId"},
"maxShardsPerNode":"1",
"autoAddReplicas":"false"}
at
__randomizedtesting.SeedInfo.seed([6F8A169181351238:E7DE294B2FC97FC0]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.assertTrue(Assert.java:43)
at
org.apache.solr.cloud.LeaderFailoverAfterPartitionTest.testRf3WithLeaderFailover(LeaderFailoverAfterPartitionTest.java:170)
at
org.apache.solr.cloud.LeaderFailoverAfterPartitionTest.test(LeaderFailoverAfterPartitionTest.java:57)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1713)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:907)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:943)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:957)
at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:985)
at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:960)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:916)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:802)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:852)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:745)
Build Log:
[...truncated 12478 lines...]
[junit4] Suite: org.apache.solr.cloud.LeaderFailoverAfterPartitionTest
[junit4] 2> Creating dataDir:
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/init-core-data-001
[junit4] 2> 1363168 INFO
(SUITE-LeaderFailoverAfterPartitionTest-seed#[6F8A169181351238]-worker) [ ]
o.a.s.SolrTestCaseJ4 Using PointFields
[junit4] 2> 1363170 INFO
(SUITE-LeaderFailoverAfterPartitionTest-seed#[6F8A169181351238]-worker) [ ]
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via:
@org.apache.solr.SolrTestCaseJ4$SuppressSSL(bugUrl=https://issues.apache.org/jira/browse/SOLR-5776)
[junit4] 2> 1363170 INFO
(SUITE-LeaderFailoverAfterPartitionTest-seed#[6F8A169181351238]-worker) [ ]
o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /sj_a/
[junit4] 2> 1363171 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 1363171 INFO (Thread-2451) [ ] o.a.s.c.ZkTestServer client
port:0.0.0.0/0.0.0.0:0
[junit4] 2> 1363171 INFO (Thread-2451) [ ] o.a.s.c.ZkTestServer
Starting server
[junit4] 2> 1363271 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.ZkTestServer start zk server on port:40029
[junit4] 2> 1363276 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
to /configs/conf1/solrconfig.xml
[junit4] 2> 1363277 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/schema.xml
to /configs/conf1/schema.xml
[junit4] 2> 1363278 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
[junit4] 2> 1363278 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
to /configs/conf1/stopwords.txt
[junit4] 2> 1363279 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/protwords.txt
to /configs/conf1/protwords.txt
[junit4] 2> 1363279 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/currency.xml
to /configs/conf1/currency.xml
[junit4] 2> 1363280 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml
to /configs/conf1/enumsConfig.xml
[junit4] 2> 1363280 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json
to /configs/conf1/open-exchange-rates.json
[junit4] 2> 1363281 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt
to /configs/conf1/mapping-ISOLatin1Accent.txt
[junit4] 2> 1363281 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt
to /configs/conf1/old_synonyms.txt
[junit4] 2> 1363282 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractZkTestCase put
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/synonyms.txt
to /configs/conf1/synonyms.txt
[junit4] 2> 1363338 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.SolrTestCaseJ4 Writing core.properties file to
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/control-001/cores/collection1
[junit4] 2> 1363340 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.Server jetty-9.3.14.v20161028
[junit4] 2> 1363340 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@1c2b1139{/sj_a,null,AVAILABLE}
[junit4] 2> 1363341 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@1407e9e2{HTTP/1.1,[http/1.1]}{127.0.0.1:38103}
[junit4] 2> 1363341 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.Server Started @1365348ms
[junit4] 2> 1363341 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/tempDir-001/control/data,
hostContext=/sj_a, hostPort=35368,
coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/control-001/cores}
[junit4] 2> 1363342 ERROR
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 1363342 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.0.0
[junit4] 2> 1363342 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1363342 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1363342 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2017-03-01T15:18:47.576Z
[junit4] 2> 1363343 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 1363344 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/control-001/solr.xml
[junit4] 2> 1363348 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:40029/solr
[junit4] 2> 1363356 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:35368_sj_a ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1363357 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:35368_sj_a ] o.a.s.c.OverseerElectionContext I am going to be
the leader 127.0.0.1:35368_sj_a
[junit4] 2> 1363357 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:35368_sj_a ] o.a.s.c.Overseer Overseer
(id=97542571780079620-127.0.0.1:35368_sj_a-n_0000000000) starting
[junit4] 2> 1363359 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:35368_sj_a ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:35368_sj_a
[junit4] 2> 1363360 INFO
(zkCallback-2055-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (1)
[junit4] 2> 1363396 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:35368_sj_a ] o.a.s.c.CorePropertiesLocator Found 1 core
definitions underneath
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/control-001/cores
[junit4] 2> 1363396 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:35368_sj_a ] o.a.s.c.CorePropertiesLocator Cores are:
[collection1]
[junit4] 2> 1363397 INFO
(OverseerStateUpdate-97542571780079620-127.0.0.1:35368_sj_a-n_0000000000)
[n:127.0.0.1:35368_sj_a ] o.a.s.c.o.ReplicaMutator Assigning new node to
shard shard=shard1
[junit4] 2> 1364405 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.0.0
[junit4] 2> 1364414 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.s.IndexSchema [collection1] Schema name=test
[junit4] 2> 1364481 WARN
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.s.IndexSchema [collection1] default search field in schema is text.
WARNING: Deprecated, please use 'df' on request instead.
[junit4] 2> 1364483 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.s.IndexSchema Loaded schema test/1.0 with uniqueid field id
[junit4] 2> 1364488 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.c.CoreContainer Creating SolrCore 'collection1' using configuration from
collection control_collection
[junit4] 2> 1364489 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1] o.a.s.c.SolrCore
[[collection1] ] Opening new SolrCore at
[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/control-001/cores/collection1],
dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/control-001/cores/collection1/data/]
[junit4] 2> 1364489 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.c.JmxMonitoredMap JMX monitoring is enabled. Adding Solr mbeans to JMX
Server: com.sun.jmx.mbeanserver.JmxMBeanServer@1047d1f9
[junit4] 2> 1364490 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=12, maxMergeAtOnceExplicit=38, maxMergedSegmentMB=2.697265625,
floorSegmentMB=1.9716796875, forceMergeDeletesPctAllowed=20.872616676584496,
segmentsPerTier=35.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.504853127495482
[junit4] 2> 1364493 WARN
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type =
requestHandler,name = /dump,class = DumpRequestHandler,attributes =
{initParams=a, name=/dump, class=DumpRequestHandler},args =
{defaults={a=A,b=B}}}
[junit4] 2> 1364516 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.u.UpdateHandler Using UpdateLog implementation:
org.apache.solr.update.UpdateLog
[junit4] 2> 1364516 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1] o.a.s.u.UpdateLog
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1364516 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 1364516 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 1364517 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy:
minMergeSize=1000, mergeFactor=46, maxMergeSize=9223372036854775807,
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false,
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.7629180363599304]
[junit4] 2> 1364517 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.s.SolrIndexSearcher Opening [Searcher@433dca2[collection1] main]
[junit4] 2> 1364518 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 1364518 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1364518 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000
[junit4] 2> 1364519 INFO
(searcherExecutor-5318-thread-1-processing-n:127.0.0.1:35368_sj_a x:collection1
c:control_collection) [n:127.0.0.1:35368_sj_a c:control_collection
x:collection1] o.a.s.c.SolrCore [collection1] Registered new searcher
Searcher@433dca2[collection1]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1364519 INFO
(coreLoadExecutor-5317-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a c:control_collection x:collection1] o.a.s.u.UpdateLog
Could not find max version in index or recent updates, using new clock
1560681149893705728
[junit4] 2> 1364522 INFO
(coreZkRegister-5310-thread-1-processing-n:127.0.0.1:35368_sj_a x:collection1
c:control_collection) [n:127.0.0.1:35368_sj_a c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext Enough replicas
found to continue.
[junit4] 2> 1364522 INFO
(coreZkRegister-5310-thread-1-processing-n:127.0.0.1:35368_sj_a x:collection1
c:control_collection) [n:127.0.0.1:35368_sj_a c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext I may be the new
leader - try and sync
[junit4] 2> 1364523 INFO
(coreZkRegister-5310-thread-1-processing-n:127.0.0.1:35368_sj_a x:collection1
c:control_collection) [n:127.0.0.1:35368_sj_a c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.SyncStrategy Sync replicas to
http://127.0.0.1:35368/sj_a/collection1/
[junit4] 2> 1364523 INFO
(coreZkRegister-5310-thread-1-processing-n:127.0.0.1:35368_sj_a x:collection1
c:control_collection) [n:127.0.0.1:35368_sj_a c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.SyncStrategy Sync Success - now sync
replicas to me
[junit4] 2> 1364523 INFO
(coreZkRegister-5310-thread-1-processing-n:127.0.0.1:35368_sj_a x:collection1
c:control_collection) [n:127.0.0.1:35368_sj_a c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.SyncStrategy
http://127.0.0.1:35368/sj_a/collection1/ has no replicas
[junit4] 2> 1364523 INFO
(coreZkRegister-5310-thread-1-processing-n:127.0.0.1:35368_sj_a x:collection1
c:control_collection) [n:127.0.0.1:35368_sj_a c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext Found all
replicas participating in election, clear LIR
[junit4] 2> 1364524 INFO
(coreZkRegister-5310-thread-1-processing-n:127.0.0.1:35368_sj_a x:collection1
c:control_collection) [n:127.0.0.1:35368_sj_a c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext I am the new
leader: http://127.0.0.1:35368/sj_a/collection1/ shard1
[junit4] 2> 1364674 INFO
(coreZkRegister-5310-thread-1-processing-n:127.0.0.1:35368_sj_a x:collection1
c:control_collection) [n:127.0.0.1:35368_sj_a c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.ZkController I am the leader, no recovery
necessary
[junit4] 2> 1364843 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1364844 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:40029/solr ready
[junit4] 2> 1364844 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection
loss:false
[junit4] 2> 1364899 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.SolrTestCaseJ4 Writing core.properties file to
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-1-001/cores/collection1
[junit4] 2> 1364899 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-1-001
[junit4] 2> 1364900 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.Server jetty-9.3.14.v20161028
[junit4] 2> 1364900 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@2049d6a4{/sj_a,null,AVAILABLE}
[junit4] 2> 1364901 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@4b8dd8dd{HTTP/1.1,[http/1.1]}{127.0.0.1:38434}
[junit4] 2> 1364901 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.Server Started @1366907ms
[junit4] 2> 1364901 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/tempDir-001/jetty1,
solrconfig=solrconfig.xml, hostContext=/sj_a, hostPort=33369,
coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-1-001/cores}
[junit4] 2> 1364901 ERROR
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 1364901 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.0.0
[junit4] 2> 1364902 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1364902 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1364902 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2017-03-01T15:18:49.136Z
[junit4] 2> 1364903 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 1364903 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-1-001/solr.xml
[junit4] 2> 1364908 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:40029/solr
[junit4] 2> 1364913 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:33369_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (1)
[junit4] 2> 1364914 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:33369_sj_a ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1364915 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:33369_sj_a ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:33369_sj_a
[junit4] 2> 1364915 INFO (zkCallback-2059-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1364915 INFO
(zkCallback-2055-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (1) -> (2)
[junit4] 2> 1364915 INFO
(zkCallback-2064-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (1) -> (2)
[junit4] 2> 1364967 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:33369_sj_a ] o.a.s.c.CorePropertiesLocator Found 1 core
definitions underneath
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-1-001/cores
[junit4] 2> 1364967 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:33369_sj_a ] o.a.s.c.CorePropertiesLocator Cores are:
[collection1]
[junit4] 2> 1364968 INFO
(OverseerStateUpdate-97542571780079620-127.0.0.1:35368_sj_a-n_0000000000)
[n:127.0.0.1:35368_sj_a ] o.a.s.c.o.ReplicaMutator Assigning new node to
shard shard=shard2
[junit4] 2> 1365990 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.c.SolrConfig Using
Lucene MatchVersion: 7.0.0
[junit4] 2> 1365999 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.s.IndexSchema
[collection1] Schema name=test
[junit4] 2> 1366058 WARN
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.s.IndexSchema
[collection1] default search field in schema is text. WARNING: Deprecated,
please use 'df' on request instead.
[junit4] 2> 1366059 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.s.IndexSchema
Loaded schema test/1.0 with uniqueid field id
[junit4] 2> 1366065 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.c.CoreContainer
Creating SolrCore 'collection1' using configuration from collection collection1
[junit4] 2> 1366066 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.c.SolrCore
[[collection1] ] Opening new SolrCore at
[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-1-001/cores/collection1],
dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-1-001/cores/collection1/data/]
[junit4] 2> 1366066 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.c.JmxMonitoredMap
JMX monitoring is enabled. Adding Solr mbeans to JMX Server:
com.sun.jmx.mbeanserver.JmxMBeanServer@1047d1f9
[junit4] 2> 1366066 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=12, maxMergeAtOnceExplicit=38, maxMergedSegmentMB=2.697265625,
floorSegmentMB=1.9716796875, forceMergeDeletesPctAllowed=20.872616676584496,
segmentsPerTier=35.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.504853127495482
[junit4] 2> 1366070 WARN
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.c.RequestHandlers
INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class
= DumpRequestHandler,attributes = {initParams=a, name=/dump,
class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
[junit4] 2> 1366093 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.u.UpdateHandler
Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1366093 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.u.UpdateLog
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1366094 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.u.CommitTracker
Hard AutoCommit: disabled
[junit4] 2> 1366094 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.u.CommitTracker
Soft AutoCommit: disabled
[junit4] 2> 1366094 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy:
minMergeSize=1000, mergeFactor=46, maxMergeSize=9223372036854775807,
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false,
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.7629180363599304]
[junit4] 2> 1366094 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1]
o.a.s.s.SolrIndexSearcher Opening [Searcher@334a2c38[collection1] main]
[junit4] 2> 1366095 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 1366095 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1366095 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000
[junit4] 2> 1366096 INFO
(searcherExecutor-5329-thread-1-processing-n:127.0.0.1:33369_sj_a x:collection1
c:collection1) [n:127.0.0.1:33369_sj_a c:collection1 x:collection1]
o.a.s.c.SolrCore [collection1] Registered new searcher
Searcher@334a2c38[collection1]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1366096 INFO
(coreLoadExecutor-5328-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a c:collection1 x:collection1] o.a.s.u.UpdateLog Could
not find max version in index or recent updates, using new clock
1560681151547310080
[junit4] 2> 1366100 INFO
(coreZkRegister-5323-thread-1-processing-n:127.0.0.1:33369_sj_a x:collection1
c:collection1) [n:127.0.0.1:33369_sj_a c:collection1 s:shard2 r:core_node1
x:collection1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to
continue.
[junit4] 2> 1366100 INFO
(coreZkRegister-5323-thread-1-processing-n:127.0.0.1:33369_sj_a x:collection1
c:collection1) [n:127.0.0.1:33369_sj_a c:collection1 s:shard2 r:core_node1
x:collection1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try
and sync
[junit4] 2> 1366100 INFO
(coreZkRegister-5323-thread-1-processing-n:127.0.0.1:33369_sj_a x:collection1
c:collection1) [n:127.0.0.1:33369_sj_a c:collection1 s:shard2 r:core_node1
x:collection1] o.a.s.c.SyncStrategy Sync replicas to
http://127.0.0.1:33369/sj_a/collection1/
[junit4] 2> 1366101 INFO
(coreZkRegister-5323-thread-1-processing-n:127.0.0.1:33369_sj_a x:collection1
c:collection1) [n:127.0.0.1:33369_sj_a c:collection1 s:shard2 r:core_node1
x:collection1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 1366101 INFO
(coreZkRegister-5323-thread-1-processing-n:127.0.0.1:33369_sj_a x:collection1
c:collection1) [n:127.0.0.1:33369_sj_a c:collection1 s:shard2 r:core_node1
x:collection1] o.a.s.c.SyncStrategy http://127.0.0.1:33369/sj_a/collection1/
has no replicas
[junit4] 2> 1366101 INFO
(coreZkRegister-5323-thread-1-processing-n:127.0.0.1:33369_sj_a x:collection1
c:collection1) [n:127.0.0.1:33369_sj_a c:collection1 s:shard2 r:core_node1
x:collection1] o.a.s.c.ShardLeaderElectionContext Found all replicas
participating in election, clear LIR
[junit4] 2> 1366102 INFO
(coreZkRegister-5323-thread-1-processing-n:127.0.0.1:33369_sj_a x:collection1
c:collection1) [n:127.0.0.1:33369_sj_a c:collection1 s:shard2 r:core_node1
x:collection1] o.a.s.c.ShardLeaderElectionContext I am the new leader:
http://127.0.0.1:33369/sj_a/collection1/ shard2
[junit4] 2> 1366253 INFO
(coreZkRegister-5323-thread-1-processing-n:127.0.0.1:33369_sj_a x:collection1
c:collection1) [n:127.0.0.1:33369_sj_a c:collection1 s:shard2 r:core_node1
x:collection1] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1366387 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.SolrTestCaseJ4 Writing core.properties file to
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-2-001/cores/collection1
[junit4] 2> 1366387 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-2-001
[junit4] 2> 1366388 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.Server jetty-9.3.14.v20161028
[junit4] 2> 1366388 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@59351286{/sj_a,null,AVAILABLE}
[junit4] 2> 1366389 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@32b40b3b{HTTP/1.1,[http/1.1]}{127.0.0.1:46170}
[junit4] 2> 1366389 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.Server Started @1368395ms
[junit4] 2> 1366389 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/tempDir-001/jetty2,
solrconfig=solrconfig.xml, hostContext=/sj_a, hostPort=36651,
coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-2-001/cores}
[junit4] 2> 1366389 ERROR
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 1366389 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.0.0
[junit4] 2> 1366389 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1366389 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1366389 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2017-03-01T15:18:50.623Z
[junit4] 2> 1366391 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 1366391 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-2-001/solr.xml
[junit4] 2> 1366396 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:40029/solr
[junit4] 2> 1366400 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:36651_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (2)
[junit4] 2> 1366401 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:36651_sj_a ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1366402 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:36651_sj_a ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:36651_sj_a
[junit4] 2> 1366403 INFO (zkCallback-2059-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1366403 INFO
(zkCallback-2064-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (2) -> (3)
[junit4] 2> 1366403 INFO
(zkCallback-2055-thread-2-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (2) -> (3)
[junit4] 2> 1366403 INFO
(zkCallback-2070-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (2) -> (3)
[junit4] 2> 1366460 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:36651_sj_a ] o.a.s.c.CorePropertiesLocator Found 1 core
definitions underneath
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-2-001/cores
[junit4] 2> 1366460 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:36651_sj_a ] o.a.s.c.CorePropertiesLocator Cores are:
[collection1]
[junit4] 2> 1366461 INFO
(OverseerStateUpdate-97542571780079620-127.0.0.1:35368_sj_a-n_0000000000)
[n:127.0.0.1:35368_sj_a ] o.a.s.c.o.ReplicaMutator Assigning new node to
shard shard=shard1
[junit4] 2> 1367472 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.c.SolrConfig Using
Lucene MatchVersion: 7.0.0
[junit4] 2> 1367486 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.s.IndexSchema
[collection1] Schema name=test
[junit4] 2> 1368197 ERROR (Finalizer) [ ] o.a.s.u.ConcurrentLFUCache
ConcurrentLFUCache was not destroyed prior to finalize(), indicates a bug --
POSSIBLE RESOURCE LEAK!!!
[junit4] 2> 1368262 WARN
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.s.IndexSchema
[collection1] default search field in schema is text. WARNING: Deprecated,
please use 'df' on request instead.
[junit4] 2> 1368265 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.s.IndexSchema
Loaded schema test/1.0 with uniqueid field id
[junit4] 2> 1368273 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.c.CoreContainer
Creating SolrCore 'collection1' using configuration from collection collection1
[junit4] 2> 1368274 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.c.SolrCore
[[collection1] ] Opening new SolrCore at
[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-2-001/cores/collection1],
dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-2-001/cores/collection1/data/]
[junit4] 2> 1368274 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.c.JmxMonitoredMap
JMX monitoring is enabled. Adding Solr mbeans to JMX Server:
com.sun.jmx.mbeanserver.JmxMBeanServer@1047d1f9
[junit4] 2> 1368275 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=12, maxMergeAtOnceExplicit=38, maxMergedSegmentMB=2.697265625,
floorSegmentMB=1.9716796875, forceMergeDeletesPctAllowed=20.872616676584496,
segmentsPerTier=35.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.504853127495482
[junit4] 2> 1368282 WARN
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.c.RequestHandlers
INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class
= DumpRequestHandler,attributes = {initParams=a, name=/dump,
class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
[junit4] 2> 1368323 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.u.UpdateHandler
Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1368323 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.u.UpdateLog
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1368324 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.u.CommitTracker
Hard AutoCommit: disabled
[junit4] 2> 1368324 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.u.CommitTracker
Soft AutoCommit: disabled
[junit4] 2> 1368324 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy:
minMergeSize=1000, mergeFactor=46, maxMergeSize=9223372036854775807,
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false,
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.7629180363599304]
[junit4] 2> 1368325 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1]
o.a.s.s.SolrIndexSearcher Opening [Searcher@25b4bf9e[collection1] main]
[junit4] 2> 1368326 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 1368326 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1368326 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000
[junit4] 2> 1368327 INFO
(searcherExecutor-5340-thread-1-processing-n:127.0.0.1:36651_sj_a x:collection1
c:collection1) [n:127.0.0.1:36651_sj_a c:collection1 x:collection1]
o.a.s.c.SolrCore [collection1] Registered new searcher
Searcher@25b4bf9e[collection1]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1368328 INFO
(coreLoadExecutor-5339-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a c:collection1 x:collection1] o.a.s.u.UpdateLog Could
not find max version in index or recent updates, using new clock
1560681153887731712
[junit4] 2> 1368332 INFO
(coreZkRegister-5334-thread-1-processing-n:127.0.0.1:36651_sj_a x:collection1
c:collection1) [n:127.0.0.1:36651_sj_a c:collection1 s:shard1 r:core_node2
x:collection1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to
continue.
[junit4] 2> 1368332 INFO
(coreZkRegister-5334-thread-1-processing-n:127.0.0.1:36651_sj_a x:collection1
c:collection1) [n:127.0.0.1:36651_sj_a c:collection1 s:shard1 r:core_node2
x:collection1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try
and sync
[junit4] 2> 1368332 INFO
(coreZkRegister-5334-thread-1-processing-n:127.0.0.1:36651_sj_a x:collection1
c:collection1) [n:127.0.0.1:36651_sj_a c:collection1 s:shard1 r:core_node2
x:collection1] o.a.s.c.SyncStrategy Sync replicas to
http://127.0.0.1:36651/sj_a/collection1/
[junit4] 2> 1368332 INFO
(coreZkRegister-5334-thread-1-processing-n:127.0.0.1:36651_sj_a x:collection1
c:collection1) [n:127.0.0.1:36651_sj_a c:collection1 s:shard1 r:core_node2
x:collection1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 1368332 INFO
(coreZkRegister-5334-thread-1-processing-n:127.0.0.1:36651_sj_a x:collection1
c:collection1) [n:127.0.0.1:36651_sj_a c:collection1 s:shard1 r:core_node2
x:collection1] o.a.s.c.SyncStrategy http://127.0.0.1:36651/sj_a/collection1/
has no replicas
[junit4] 2> 1368332 INFO
(coreZkRegister-5334-thread-1-processing-n:127.0.0.1:36651_sj_a x:collection1
c:collection1) [n:127.0.0.1:36651_sj_a c:collection1 s:shard1 r:core_node2
x:collection1] o.a.s.c.ShardLeaderElectionContext Found all replicas
participating in election, clear LIR
[junit4] 2> 1368334 INFO
(coreZkRegister-5334-thread-1-processing-n:127.0.0.1:36651_sj_a x:collection1
c:collection1) [n:127.0.0.1:36651_sj_a c:collection1 s:shard1 r:core_node2
x:collection1] o.a.s.c.ShardLeaderElectionContext I am the new leader:
http://127.0.0.1:36651/sj_a/collection1/ shard1
[junit4] 2> 1368485 INFO
(coreZkRegister-5334-thread-1-processing-n:127.0.0.1:36651_sj_a x:collection1
c:collection1) [n:127.0.0.1:36651_sj_a c:collection1 s:shard1 r:core_node2
x:collection1] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 1368751 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.SolrTestCaseJ4 Writing core.properties file to
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-3-001/cores/collection1
[junit4] 2> 1368751 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase create jetty 3 in directory
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-3-001
[junit4] 2> 1368752 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.Server jetty-9.3.14.v20161028
[junit4] 2> 1368753 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@3d2b36bd{/sj_a,null,AVAILABLE}
[junit4] 2> 1368753 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@548a67b4{HTTP/1.1,[http/1.1]}{127.0.0.1:41350}
[junit4] 2> 1368754 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.Server Started @1370760ms
[junit4] 2> 1368754 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/tempDir-001/jetty3,
solrconfig=solrconfig.xml, hostContext=/sj_a, hostPort=43964,
coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-3-001/cores}
[junit4] 2> 1368754 ERROR
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 1368754 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
7.0.0
[junit4] 2> 1368754 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1368754 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1368754 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2017-03-01T15:18:52.988Z
[junit4] 2> 1368756 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 1368756 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.SolrXmlConfig Loading container configuration from
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-3-001/solr.xml
[junit4] 2> 1368761 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:40029/solr
[junit4] 2> 1368765 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:43964_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (3)
[junit4] 2> 1368766 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:43964_sj_a ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1368767 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:43964_sj_a ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:43964_sj_a
[junit4] 2> 1368768 INFO (zkCallback-2059-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
[junit4] 2> 1368768 INFO
(zkCallback-2055-thread-2-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (3) -> (4)
[junit4] 2> 1368768 INFO
(zkCallback-2064-thread-1-processing-n:127.0.0.1:33369_sj_a)
[n:127.0.0.1:33369_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (3) -> (4)
[junit4] 2> 1368768 INFO
(zkCallback-2070-thread-1-processing-n:127.0.0.1:36651_sj_a)
[n:127.0.0.1:36651_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (3) -> (4)
[junit4] 2> 1368768 INFO
(zkCallback-2076-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (3) -> (4)
[junit4] 2> 1368822 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:43964_sj_a ] o.a.s.c.CorePropertiesLocator Found 1 core
definitions underneath
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-3-001/cores
[junit4] 2> 1368822 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238])
[n:127.0.0.1:43964_sj_a ] o.a.s.c.CorePropertiesLocator Cores are:
[collection1]
[junit4] 2> 1368823 INFO
(OverseerStateUpdate-97542571780079620-127.0.0.1:35368_sj_a-n_0000000000)
[n:127.0.0.1:35368_sj_a ] o.a.s.c.o.ReplicaMutator Assigning new node to
shard shard=shard2
[junit4] 2> 1369838 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.c.SolrConfig Using
Lucene MatchVersion: 7.0.0
[junit4] 2> 1369848 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.s.IndexSchema
[collection1] Schema name=test
[junit4] 2> 1369908 WARN
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.s.IndexSchema
[collection1] default search field in schema is text. WARNING: Deprecated,
please use 'df' on request instead.
[junit4] 2> 1369910 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.s.IndexSchema
Loaded schema test/1.0 with uniqueid field id
[junit4] 2> 1369916 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.c.CoreContainer
Creating SolrCore 'collection1' using configuration from collection collection1
[junit4] 2> 1369916 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.c.SolrCore
[[collection1] ] Opening new SolrCore at
[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-3-001/cores/collection1],
dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-3-001/cores/collection1/data/]
[junit4] 2> 1369916 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.c.JmxMonitoredMap
JMX monitoring is enabled. Adding Solr mbeans to JMX Server:
com.sun.jmx.mbeanserver.JmxMBeanServer@1047d1f9
[junit4] 2> 1369917 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=12, maxMergeAtOnceExplicit=38, maxMergedSegmentMB=2.697265625,
floorSegmentMB=1.9716796875, forceMergeDeletesPctAllowed=20.872616676584496,
segmentsPerTier=35.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.504853127495482
[junit4] 2> 1369920 WARN
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.c.RequestHandlers
INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class
= DumpRequestHandler,attributes = {initParams=a, name=/dump,
class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
[junit4] 2> 1369944 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.u.UpdateHandler
Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 1369944 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.u.UpdateLog
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 1369945 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.u.CommitTracker
Hard AutoCommit: disabled
[junit4] 2> 1369945 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.u.CommitTracker
Soft AutoCommit: disabled
[junit4] 2> 1369945 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy:
minMergeSize=1000, mergeFactor=46, maxMergeSize=9223372036854775807,
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false,
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.7629180363599304]
[junit4] 2> 1369945 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1]
o.a.s.s.SolrIndexSearcher Opening [Searcher@5d6088da[collection1] main]
[junit4] 2> 1369946 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 1369946 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 1369946 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000
[junit4] 2> 1369947 INFO
(searcherExecutor-5351-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
c:collection1) [n:127.0.0.1:43964_sj_a c:collection1 x:collection1]
o.a.s.c.SolrCore [collection1] Registered new searcher
Searcher@5d6088da[collection1]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 1369947 INFO
(coreLoadExecutor-5350-thread-1-processing-n:127.0.0.1:43964_sj_a)
[n:127.0.0.1:43964_sj_a c:collection1 x:collection1] o.a.s.u.UpdateLog Could
not find max version in index or recent updates, using new clock
1560681155585376256
[junit4] 2> 1369950 INFO
(coreZkRegister-5345-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
c:collection1) [n:127.0.0.1:43964_sj_a c:collection1 s:shard2 r:core_node3
x:collection1] o.a.s.c.ZkController Core needs to recover:collection1
[junit4] 2> 1369951 INFO
(updateExecutor-2073-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.u.DefaultSolrCoreState Running
recovery
[junit4] 2> 1369951 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.c.RecoveryStrategy Starting recovery
process. recoveringAfterStartup=true
[junit4] 2> 1369951 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.c.RecoveryStrategy ######
startupVersions=[[]]
[junit4] 2> 1369951 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.c.RecoveryStrategy Begin buffering
updates. core=[collection1]
[junit4] 2> 1369951 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.u.UpdateLog Starting to buffer
updates. FSUpdateLog{state=ACTIVE, tlog=null}
[junit4] 2> 1369951 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.c.RecoveryStrategy Publishing state
of core [collection1] as recovering, leader is
[http://127.0.0.1:33369/sj_a/collection1/] and I am
[http://127.0.0.1:43964/sj_a/collection1/]
[junit4] 2> 1369952 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.c.RecoveryStrategy Sending prep
recovery command to [http://127.0.0.1:33369/sj_a]; [WaitForState:
action=PREPRECOVERY&core=collection1&nodeName=127.0.0.1:43964_sj_a&coreNodeName=core_node3&state=recovering&checkLive=true&onlyIfLeader=true&onlyIfLeaderActive=true]
[junit4] 2> 1369953 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=49178,localport=33369], receiveBufferSize:531000
[junit4] 2> 1369956 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=38434,localport=53866], receiveBufferSize=530904
[junit4] 2> 1369957 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
] o.a.s.h.a.PrepRecoveryOp Going to wait for coreNodeName: core_node3, state:
recovering, checkLive: true, onlyIfLeader: true, onlyIfLeaderActive: true
[junit4] 2> 1369957 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
] o.a.s.h.a.PrepRecoveryOp Will wait a max of 183 seconds to see collection1
(shard2 of collection1) have state: recovering
[junit4] 2> 1369957 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
] o.a.s.h.a.PrepRecoveryOp In WaitForState(recovering): collection=collection1,
shard=shard2, thisCore=collection1, leaderDoesNotNeedRecovery=false, isLeader?
true, live=true, checkLive=true, currentState=down, localState=active,
nodeName=127.0.0.1:43964_sj_a, coreNodeName=core_node3,
onlyIfActiveCheckResult=false, nodeProps:
core_node3:{"core":"collection1","base_url":"http://127.0.0.1:43964/sj_a","node_name":"127.0.0.1:43964_sj_a","state":"down"}
[junit4] 2> 1370187 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.SolrTestCaseJ4 ###Starting test
[junit4] 2> 1370187 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase Wait for recoveries to finish - wait
30000 for each attempt
[junit4] 2> 1370187 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractDistribZkTestBase Wait for recoveries to finish - collection:
collection1 failOnTimeout:true timeout (sec):30000
[junit4] 2> 1370957 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
] o.a.s.h.a.PrepRecoveryOp In WaitForState(recovering): collection=collection1,
shard=shard2, thisCore=collection1, leaderDoesNotNeedRecovery=false, isLeader?
true, live=true, checkLive=true, currentState=recovering, localState=active,
nodeName=127.0.0.1:43964_sj_a, coreNodeName=core_node3,
onlyIfActiveCheckResult=false, nodeProps:
core_node3:{"core":"collection1","base_url":"http://127.0.0.1:43964/sj_a","node_name":"127.0.0.1:43964_sj_a","state":"recovering"}
[junit4] 2> 1370957 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
] o.a.s.h.a.PrepRecoveryOp Waited coreNodeName: core_node3, state: recovering,
checkLive: true, onlyIfLeader: true for: 1 seconds.
[junit4] 2> 1370958 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores
params={nodeName=127.0.0.1:43964_sj_a&onlyIfLeaderActive=true&core=collection1&coreNodeName=core_node3&action=PREPRECOVERY&checkLive=true&state=recovering&onlyIfLeader=true&wt=javabin&version=2}
status=0 QTime=1000
[junit4] 2> 1371458 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.c.RecoveryStrategy Attempting to
PeerSync from [http://127.0.0.1:33369/sj_a/collection1/] -
recoveringAfterStartup=[true]
[junit4] 2> 1371459 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.u.PeerSync PeerSync:
core=collection1 url=http://127.0.0.1:43964/sj_a START
replicas=[http://127.0.0.1:33369/sj_a/collection1/] nUpdates=100
[junit4] 2> 1371459 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=49274,localport=33369], receiveBufferSize:531000
[junit4] 2> 1371459 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=38434,localport=53964], receiveBufferSize=530904
[junit4] 2> 1371463 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:collection1 s:shard2 r:core_node1 x:collection1] o.a.s.u.IndexFingerprint
IndexFingerprint millis:0.0 result:{maxVersionSpecified=9223372036854775807,
maxVersionEncountered=0, maxInHash=0, versionsHash=0, numVersions=0, numDocs=0,
maxDoc=0}
[junit4] 2> 1371464 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:collection1 s:shard2 r:core_node1 x:collection1] o.a.s.c.S.Request
[collection1] webapp=/sj_a path=/get
params={distrib=false&qt=/get&getFingerprint=9223372036854775807&wt=javabin&version=2}
status=0 QTime=1
[junit4] 2> 1371464 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.u.IndexFingerprint IndexFingerprint
millis:0.0 result:{maxVersionSpecified=9223372036854775807,
maxVersionEncountered=0, maxInHash=0, versionsHash=0, numVersions=0, numDocs=0,
maxDoc=0}
[junit4] 2> 1371464 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.u.PeerSync We are already in sync.
No need to do a PeerSync
[junit4] 2> 1371464 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.u.DirectUpdateHandler2 start
commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1371464 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.u.DirectUpdateHandler2 No
uncommitted changes. Skipping IW.commit.
[junit4] 2> 1371466 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.u.DirectUpdateHandler2
end_commit_flush
[junit4] 2> 1371466 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.c.RecoveryStrategy PeerSync stage of
recovery was successful.
[junit4] 2> 1371466 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.c.RecoveryStrategy Replaying updates
buffered during PeerSync.
[junit4] 2> 1371466 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.c.RecoveryStrategy No replay needed.
[junit4] 2> 1371466 INFO
(recoveryExecutor-2074-thread-1-processing-n:127.0.0.1:43964_sj_a x:collection1
s:shard2 c:collection1 r:core_node3) [n:127.0.0.1:43964_sj_a c:collection1
s:shard2 r:core_node3 x:collection1] o.a.s.c.RecoveryStrategy Registering as
Active after recovery.
[junit4] 2> 1372187 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractDistribZkTestBase Recoveries finished - collection: collection1
[junit4] 2> 1372188 INFO (SocketProxy-Acceptor-35368) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=53924,localport=35368], receiveBufferSize:531000
[junit4] 2> 1372189 INFO (SocketProxy-Acceptor-35368) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=38103,localport=46076], receiveBufferSize=530904
[junit4] 2> 1372190 INFO (qtp1974396158-11209) [n:127.0.0.1:35368_sj_a
c:control_collection s:shard1 r:core_node1 x:collection1]
o.a.s.u.DirectUpdateHandler2 start
commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1372190 INFO (qtp1974396158-11209) [n:127.0.0.1:35368_sj_a
c:control_collection s:shard1 r:core_node1 x:collection1]
o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
[junit4] 2> 1372190 INFO (qtp1974396158-11209) [n:127.0.0.1:35368_sj_a
c:control_collection s:shard1 r:core_node1 x:collection1]
o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 1372191 INFO (qtp1974396158-11209) [n:127.0.0.1:35368_sj_a
c:control_collection s:shard1 r:core_node1 x:collection1]
o.a.s.u.p.LogUpdateProcessorFactory [collection1] webapp=/sj_a path=/update
params={waitSearcher=true&commit=true&softCommit=false&wt=javabin&version=2}{commit=}
0 0
[junit4] 2> 1372191 INFO (SocketProxy-Acceptor-36651) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=56738,localport=36651], receiveBufferSize:531000
[junit4] 2> 1372192 INFO (SocketProxy-Acceptor-36651) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=46170,localport=58378], receiveBufferSize=530904
[junit4] 2> 1372193 INFO (SocketProxy-Acceptor-36651) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=56742,localport=36651], receiveBufferSize:531000
[junit4] 2> 1372193 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=49294,localport=33369], receiveBufferSize:531000
[junit4] 2> 1372193 INFO (SocketProxy-Acceptor-36651) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=46170,localport=58384], receiveBufferSize=530904
[junit4] 2> 1372194 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=38434,localport=53984], receiveBufferSize=530904
[junit4] 2> 1372194 INFO (SocketProxy-Acceptor-43964) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=36930,localport=43964], receiveBufferSize:531000
[junit4] 2> 1372199 INFO (qtp930611237-11277) [n:127.0.0.1:36651_sj_a
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.u.DirectUpdateHandler2
start
commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1372199 INFO (qtp930611237-11277) [n:127.0.0.1:36651_sj_a
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.u.DirectUpdateHandler2
No uncommitted changes. Skipping IW.commit.
[junit4] 2> 1372199 INFO (qtp753653889-11241) [n:127.0.0.1:33369_sj_a
c:collection1 s:shard2 r:core_node1 x:collection1] o.a.s.u.DirectUpdateHandler2
start
commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1372200 INFO (SocketProxy-Acceptor-43964) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=41350,localport=49538], receiveBufferSize=530904
[junit4] 2> 1372200 INFO (qtp753653889-11241) [n:127.0.0.1:33369_sj_a
c:collection1 s:shard2 r:core_node1 x:collection1] o.a.s.u.DirectUpdateHandler2
No uncommitted changes. Skipping IW.commit.
[junit4] 2> 1372200 INFO (qtp930611237-11277) [n:127.0.0.1:36651_sj_a
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.u.DirectUpdateHandler2
end_commit_flush
[junit4] 2> 1372201 INFO (qtp930611237-11277) [n:127.0.0.1:36651_sj_a
c:collection1 s:shard1 r:core_node2 x:collection1]
o.a.s.u.p.LogUpdateProcessorFactory [collection1] webapp=/sj_a path=/update
params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:36651/sj_a/collection1/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=}
0 6
[junit4] 2> 1372201 INFO (qtp753653889-11241) [n:127.0.0.1:33369_sj_a
c:collection1 s:shard2 r:core_node1 x:collection1] o.a.s.u.DirectUpdateHandler2
end_commit_flush
[junit4] 2> 1372201 INFO (qtp753653889-11241) [n:127.0.0.1:33369_sj_a
c:collection1 s:shard2 r:core_node1 x:collection1]
o.a.s.u.p.LogUpdateProcessorFactory [collection1] webapp=/sj_a path=/update
params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:36651/sj_a/collection1/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=}
0 1
[junit4] 2> 1372201 INFO (qtp1238309599-11304) [n:127.0.0.1:43964_sj_a
c:collection1 s:shard2 r:core_node3 x:collection1] o.a.s.u.DirectUpdateHandler2
start
commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 1372201 INFO (qtp1238309599-11304) [n:127.0.0.1:43964_sj_a
c:collection1 s:shard2 r:core_node3 x:collection1] o.a.s.u.DirectUpdateHandler2
No uncommitted changes. Skipping IW.commit.
[junit4] 2> 1372201 INFO (qtp1238309599-11304) [n:127.0.0.1:43964_sj_a
c:collection1 s:shard2 r:core_node3 x:collection1] o.a.s.u.DirectUpdateHandler2
end_commit_flush
[junit4] 2> 1372201 INFO (qtp1238309599-11304) [n:127.0.0.1:43964_sj_a
c:collection1 s:shard2 r:core_node3 x:collection1]
o.a.s.u.p.LogUpdateProcessorFactory [collection1] webapp=/sj_a path=/update
params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:36651/sj_a/collection1/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=}
0 0
[junit4] 2> 1372201 INFO (qtp930611237-11275) [n:127.0.0.1:36651_sj_a
c:collection1 s:shard1 r:core_node2 x:collection1]
o.a.s.u.p.LogUpdateProcessorFactory [collection1] webapp=/sj_a path=/update
params={waitSearcher=true&commit=true&softCommit=false&wt=javabin&version=2}{commit=}
0 8
[junit4] 2> 1372202 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=49304,localport=33369], receiveBufferSize:531000
[junit4] 2> 1372202 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=38434,localport=53992], receiveBufferSize=530904
[junit4] 2> 1372203 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:collection1 s:shard2 r:core_node1 x:collection1] o.a.s.c.S.Request
[collection1] webapp=/sj_a path=/select
params={q=*:*&distrib=false&tests=checkShardConsistency&rows=0&wt=javabin&version=2}
hits=0 status=0 QTime=0
[junit4] 2> 1372203 INFO (SocketProxy-Acceptor-43964) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=36938,localport=43964], receiveBufferSize:531000
[junit4] 2> 1372204 INFO (SocketProxy-Acceptor-43964) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=41350,localport=49546], receiveBufferSize=530904
[junit4] 2> 1372205 INFO (qtp1238309599-11306) [n:127.0.0.1:43964_sj_a
c:collection1 s:shard2 r:core_node3 x:collection1] o.a.s.c.S.Request
[collection1] webapp=/sj_a path=/select
params={q=*:*&distrib=false&tests=checkShardConsistency&rows=0&wt=javabin&version=2}
hits=0 status=0 QTime=0
[junit4] 2> 1372206 INFO (SocketProxy-Acceptor-36651) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=56762,localport=36651], receiveBufferSize:531000
[junit4] 2> 1372207 INFO (SocketProxy-Acceptor-36651) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=46170,localport=58402], receiveBufferSize=530904
[junit4] 2> 1372207 INFO (qtp930611237-11271) [n:127.0.0.1:36651_sj_a
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.c.S.Request
[collection1] webapp=/sj_a path=/select
params={q=*:*&distrib=false&tests=checkShardConsistency&rows=0&wt=javabin&version=2}
hits=0 status=0 QTime=0
[junit4] 2> 1374208 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.AbstractFullDistribZkTestBase Creating collection with stateFormat=1:
c8n_1x3_lf
[junit4] 2> 1374209 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=49332,localport=33369], receiveBufferSize:531000
[junit4] 2> 1374209 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=38434,localport=54020], receiveBufferSize=530904
[junit4] 2> 1374210 INFO (qtp753653889-11241) [n:127.0.0.1:33369_sj_a
] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params
replicationFactor=3&maxShardsPerNode=1&name=c8n_1x3_lf&action=CREATE&numShards=1&stateFormat=1&wt=javabin&version=2
and sendToOCPQueue=true
[junit4] 2> 1374211 INFO
(OverseerThreadFactory-5315-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a ] o.a.s.c.CreateCollectionCmd Create collection
c8n_1x3_lf
[junit4] 2> 1374211 INFO
(OverseerThreadFactory-5315-thread-1-processing-n:127.0.0.1:35368_sj_a)
[n:127.0.0.1:35368_sj_a ] o.a.s.c.CreateCollectionCmd Only one config set
found in zk - using it:conf1
[junit4] 2> 1374313 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=49336,localport=33369], receiveBufferSize:531000
[junit4] 2> 1374313 INFO (SocketProxy-Acceptor-36651) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=56788,localport=36651], receiveBufferSize:531000
[junit4] 2> 1374313 INFO (SocketProxy-Acceptor-43964) [ ]
o.a.s.c.SocketProxy accepted
Socket[addr=/127.0.0.1,port=36966,localport=43964], receiveBufferSize:531000
[junit4] 2> 1374314 INFO (SocketProxy-Acceptor-33369) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=38434,localport=54026], receiveBufferSize=530904
[junit4] 2> 1374315 INFO (SocketProxy-Acceptor-36651) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=46170,localport=58430], receiveBufferSize=530904
[junit4] 2> 1374315 INFO (SocketProxy-Acceptor-43964) [ ]
o.a.s.c.SocketProxy proxy connection
Socket[addr=/127.0.0.1,port=41350,localport=49580], receiveBufferSize=530904
[junit4] 2> 1374315 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
] o.a.s.h.a.CoreAdminOperation core create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&name=c8n_1x3_lf_shard1_replica1&action=CREATE&numShards=1&collection=c8n_1x3_lf&shard=shard1&wt=javabin&version=2
[junit4] 2> 1374316 INFO (qtp1238309599-11304) [n:127.0.0.1:43964_sj_a
] o.a.s.h.a.CoreAdminOperation core create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&name=c8n_1x3_lf_shard1_replica2&action=CREATE&numShards=1&collection=c8n_1x3_lf&shard=shard1&wt=javabin&version=2
[junit4] 2> 1374315 INFO (qtp930611237-11275) [n:127.0.0.1:36651_sj_a
] o.a.s.h.a.CoreAdminOperation core create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&name=c8n_1x3_lf_shard1_replica3&action=CREATE&numShards=1&collection=c8n_1x3_lf&shard=shard1&wt=javabin&version=2
[junit4] 2> 1375324 INFO (qtp930611237-11275) [n:127.0.0.1:36651_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica3] o.a.s.c.SolrConfig Using
Lucene MatchVersion: 7.0.0
[junit4] 2> 1375327 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica1] o.a.s.c.SolrConfig Using
Lucene MatchVersion: 7.0.0
[junit4] 2> 1375327 INFO (qtp1238309599-11304) [n:127.0.0.1:43964_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica2] o.a.s.c.SolrConfig Using
Lucene MatchVersion: 7.0.0
[junit4] 2> 1375334 INFO (qtp930611237-11275) [n:127.0.0.1:36651_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica3] o.a.s.s.IndexSchema
[c8n_1x3_lf_shard1_replica3] Schema name=test
[junit4] 2> 1375353 INFO (qtp1238309599-11304) [n:127.0.0.1:43964_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica2] o.a.s.s.IndexSchema
[c8n_1x3_lf_shard1_replica2] Schema name=test
[junit4] 2> 1375355 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica1] o.a.s.s.IndexSchema
[c8n_1x3_lf_shard1_replica1] Schema name=test
[junit4] 2> 1375422 WARN (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica1] o.a.s.s.IndexSchema
[c8n_1x3_lf_shard1_replica1] default search field in schema is text. WARNING:
Deprecated, please use 'df' on request instead.
[junit4] 2> 1375424 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica1] o.a.s.s.IndexSchema Loaded
schema test/1.0 with uniqueid field id
[junit4] 2> 1375431 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica1] o.a.s.c.CoreContainer
Creating SolrCore 'c8n_1x3_lf_shard1_replica1' using configuration from
collection c8n_1x3_lf
[junit4] 2> 1375431 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica1] o.a.s.c.SolrCore
[[c8n_1x3_lf_shard1_replica1] ] Opening new SolrCore at
[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-1-001/cores/c8n_1x3_lf_shard1_replica1],
dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001/shard-1-001/cores/c8n_1x3_lf_shard1_replica1/data/]
[junit4] 2> 1375431 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica1] o.a.s.c.JmxMonitoredMap
JMX monitoring is enabled. Adding Solr mbeans to JMX Server:
com.sun.jmx.mbeanserver.JmxMBeanServer@1047d1f9
[junit4] 2> 1375432 INFO (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica1] o.a.s.u.RandomMergePolicy
RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy:
[TieredMergePolicy: maxMergeAtOnce=12, maxMergeAtOnceExplicit=38,
maxMergedSegmentMB=2.697265625, floorSegmentMB=1.9716796875,
forceMergeDeletesPctAllowed=20.872616676584496, segmentsPerTier=35.0,
maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.504853127495482
[junit4] 2> 1375436 WARN (qtp753653889-11246) [n:127.0.0.1:33369_sj_a
c:c8n_1x3_lf s:shard1 x:c8n_1x3_lf_shard1_replica1] o.a.s.c.RequestHandlers
INVALID paramSet a in requestHandler {type = requestHandler,n
[...truncated too long message...]
o.a.s.m.SolrMetricManager Closing metric reporters for: solr.node
[junit4] 2> 1491359 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.AbstractConnector Stopped
ServerConnector@548a67b4{HTTP/1.1,[http/1.1]}{127.0.0.1:0}
[junit4] 2> 1491359 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.e.j.s.h.ContextHandler Stopped
o.e.j.s.ServletContextHandler@3d2b36bd{/sj_a,null,UNAVAILABLE}
[junit4] 2> 1491360 INFO
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.ZkTestServer connecting to 127.0.0.1:40029 40029
[junit4] 2> 1491381 INFO (Thread-2451) [ ] o.a.s.c.ZkTestServer
connecting to 127.0.0.1:40029 40029
[junit4] 2> 1491382 WARN (Thread-2451) [ ] o.a.s.c.ZkTestServer Watch
limit violations:
[junit4] 2> Maximum concurrent create/delete watches above limit:
[junit4] 2>
[junit4] 2> 5 /solr/aliases.json
[junit4] 2> 5 /solr/clusterprops.json
[junit4] 2> 4 /solr/security.json
[junit4] 2> 4 /solr/configs/conf1
[junit4] 2> 3 /solr/collections/c8n_1x3_lf/state.json
[junit4] 2> 3 /solr/collections/collection1/state.json
[junit4] 2>
[junit4] 2> Maximum concurrent data watches above limit:
[junit4] 2>
[junit4] 2> 5 /solr/clusterstate.json
[junit4] 2> 2
/solr/overseer_elect/election/97542571780079620-127.0.0.1:35368_sj_a-n_0000000000
[junit4] 2>
[junit4] 2> Maximum concurrent children watches above limit:
[junit4] 2>
[junit4] 2> 125 /solr/overseer/collection-queue-work
[junit4] 2> 39 /solr/overseer/queue
[junit4] 2> 5 /solr/live_nodes
[junit4] 2> 5 /solr/collections
[junit4] 2> 4 /solr/overseer/queue-work
[junit4] 2>
[junit4] 2> 1491382 WARN
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.SocketProxy Closing 1 connections to: http://127.0.0.1:35368/sj_a,
target: http://127.0.0.1:38103/sj_a
[junit4] 2> 1491382 WARN
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.SocketProxy Closing 0 connections to: http://127.0.0.1:33369/sj_a,
target: http://127.0.0.1:38434/sj_a
[junit4] 2> 1491382 WARN
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.SocketProxy Closing 15 connections to: http://127.0.0.1:43964/sj_a,
target: http://127.0.0.1:41350/sj_a
[junit4] 2> 1491382 WARN
(TEST-LeaderFailoverAfterPartitionTest.test-seed#[6F8A169181351238]) [ ]
o.a.s.c.SocketProxy Closing 1 connections to: http://127.0.0.1:36651/sj_a,
target: http://127.0.0.1:46170/sj_a
[junit4] 2> NOTE: reproduce with: ant test
-Dtestcase=LeaderFailoverAfterPartitionTest -Dtests.method=test
-Dtests.seed=6F8A169181351238 -Dtests.multiplier=3 -Dtests.slow=true
-Dtests.locale=fi-FI -Dtests.timezone=Europe/Vatican -Dtests.asserts=true
-Dtests.file.encoding=UTF-8
[junit4] FAILURE 128s J2 | LeaderFailoverAfterPartitionTest.test <<<
[junit4] > Throwable #1: java.lang.AssertionError: Expected 2 of 3
replicas to be active but only found 1;
[core_node3:{"core":"c8n_1x3_lf_shard1_replica2","base_url":"http://127.0.0.1:43964/sj_a","node_name":"127.0.0.1:43964_sj_a","state":"active","leader":"true"}];
clusterState: DocCollection(c8n_1x3_lf//clusterstate.json/33)={
[junit4] > "replicationFactor":"3",
[junit4] > "shards":{"shard1":{
[junit4] > "range":"80000000-7fffffff",
[junit4] > "state":"active",
[junit4] > "replicas":{
[junit4] > "core_node1":{
[junit4] > "core":"c8n_1x3_lf_shard1_replica1",
[junit4] > "base_url":"http://127.0.0.1:33369/sj_a",
[junit4] > "node_name":"127.0.0.1:33369_sj_a",
[junit4] > "state":"down"},
[junit4] > "core_node2":{
[junit4] > "state":"down",
[junit4] > "base_url":"http://127.0.0.1:36651/sj_a",
[junit4] > "core":"c8n_1x3_lf_shard1_replica3",
[junit4] > "node_name":"127.0.0.1:36651_sj_a"},
[junit4] > "core_node3":{
[junit4] > "core":"c8n_1x3_lf_shard1_replica2",
[junit4] > "base_url":"http://127.0.0.1:43964/sj_a",
[junit4] > "node_name":"127.0.0.1:43964_sj_a",
[junit4] > "state":"active",
[junit4] > "leader":"true"}}}},
[junit4] > "router":{"name":"compositeId"},
[junit4] > "maxShardsPerNode":"1",
[junit4] > "autoAddReplicas":"false"}
[junit4] > at
__randomizedtesting.SeedInfo.seed([6F8A169181351238:E7DE294B2FC97FC0]:0)
[junit4] > at
org.apache.solr.cloud.LeaderFailoverAfterPartitionTest.testRf3WithLeaderFailover(LeaderFailoverAfterPartitionTest.java:170)
[junit4] > at
org.apache.solr.cloud.LeaderFailoverAfterPartitionTest.test(LeaderFailoverAfterPartitionTest.java:57)
[junit4] > at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:985)
[junit4] > at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:960)
[junit4] > at java.lang.Thread.run(Thread.java:745)
[junit4] 2> 1491385 INFO
(SUITE-LeaderFailoverAfterPartitionTest-seed#[6F8A169181351238]-worker) [ ]
o.a.s.SolrTestCaseJ4 ###deleteCore
[junit4] 2> NOTE: leaving temporary files on disk at:
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.LeaderFailoverAfterPartitionTest_6F8A169181351238-001
[junit4] 2> Mar 01, 2017 3:20:55 PM
com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
[junit4] 2> WARNING: Will linger awaiting termination of 1 leaked
thread(s).
[junit4] 2> NOTE: test params are: codec=Lucene70,
sim=RandomSimilarity(queryNorm=false): {}, locale=fi-FI, timezone=Europe/Vatican
[junit4] 2> NOTE: Linux 4.4.0-53-generic amd64/Oracle Corporation
1.8.0_121 (64-bit)/cpus=12,threads=1,free=220362080,total=526385152
[junit4] 2> NOTE: All tests run in this JVM: [TestZkChroot,
StatsComponentTest, TestCSVResponseWriter, TestIBSimilarityFactory,
TestClassicSimilarityFactory, TestSort, GraphQueryTest, RequestHandlersTest,
BitVectorTest, SortSpecParsingTest, TestRTimerTree, LeaderElectionTest,
DeleteLastCustomShardedReplicaTest, AutoCommitTest,
SolrCoreCheckLockOnStartupTest, PeerSyncWithIndexFingerprintCachingTest,
NoCacheHeaderTest, TestSolrCloudWithKerberosAlt,
CloudExitableDirectoryReaderTest, TestTolerantUpdateProcessorRandomCloud,
DateFieldTest, UnloadDistributedZkTest, ReplicaListTransformerTest,
TestSchemaManager, TestRequestStatusCollectionAPI, TestDistribDocBasedVersion,
SuggesterWFSTTest, TestJsonFacetRefinement, SolrPluginUtilsTest,
StatelessScriptUpdateProcessorFactoryTest, TestReloadDeadlock,
TestPerFieldSimilarityWithDefaultOverride, ZkStateWriterTest,
DistributedVersionInfoTest, HdfsNNFailoverTest, TestSolrFieldCacheMBean,
TestStressInPlaceUpdates, TestMergePolicyConfig, HardAutoCommitTest,
TestQuerySenderNoQuery, CachingDirectoryFactoryTest, TestDistribIDF,
TestRuleBasedAuthorizationPlugin, TestConfigSetsAPIZkFailure,
HighlighterConfigTest, TestJsonRequest, XmlUpdateRequestHandlerTest,
SolrSlf4jReporterTest, TestReversedWildcardFilterFactory,
CdcrReplicationDistributedZkTest, CircularListTest, TestCoreDiscovery,
TestSmileRequest, PrimUtilsTest, URLClassifyProcessorTest,
SaslZkACLProviderTest, HdfsChaosMonkeyNothingIsSafeTest,
PrimitiveFieldTypeTest, DocValuesMultiTest, TestBulkSchemaConcurrent,
TestReloadAndDeleteDocs, TestMiniSolrCloudCluster,
ConcurrentDeleteAndCreateCollectionTest, DateRangeFieldTest,
ExitableDirectoryReaderTest, TestSerializedLuceneMatchVersion,
ClassificationUpdateProcessorIntegrationTest,
SuggestComponentContextFilterQueryTest, BinaryUpdateRequestHandlerTest,
DirectUpdateHandlerOptimizeTest, DistanceUnitsTest, TestCloudDeleteByQuery,
SharedFSAutoReplicaFailoverUtilsTest, VMParamsZkACLAndCredentialsProvidersTest,
SuggesterTest, TestDynamicFieldResource, TestMaxScoreQueryParser,
BadCopyFieldTest, ShowFileRequestHandlerTest, TestCopyFieldCollectionResource,
DistributedSuggestComponentTest, MigrateRouteKeyTest, TestFieldSortValues,
TermVectorComponentDistributedTest, PKIAuthenticationIntegrationTest,
CreateCollectionCleanupTest, CollectionTooManyReplicasTest,
DistributedFacetPivotLongTailTest, FieldAnalysisRequestHandlerTest,
AddSchemaFieldsUpdateProcessorFactoryTest, TestCustomStream,
TestPHPSerializedResponseWriter, TestExtendedDismaxParser, JavabinLoaderTest,
DirectUpdateHandlerTest, TestBackupRepositoryFactory,
DistributedFacetPivotWhiteBoxTest, AnalysisErrorHandlingTest,
TestSystemCollAutoCreate, AsyncCallRequestStatusResponseTest, TestReload,
CoreAdminHandlerTest, JSONWriterTest, TestGraphMLResponseWriter,
EchoParamsTest, TestNumericTerms64, TestConfigsApi, TestSQLHandlerNonCloud,
ActionThrottleTest, OutOfBoxZkACLAndCredentialsProvidersTest,
TestRawResponseWriter, RAMDirectoryFactoryTest,
TestImpersonationWithHadoopAuth, ExternalFileFieldSortTest, OverseerTest,
LeaderElectionIntegrationTest, RecoveryZkTest, FullSolrCloudDistribCmdsTest,
ClusterStateUpdateTest, TestRandomDVFaceting, TestDistributedSearch,
ZkControllerTest, TestStressVersions, SimpleFacetsTest,
SpellCheckComponentTest, SolrCmdDistributorTest, PeerSyncTest, TestLazyCores,
TestBadConfig, TestIndexSearcher, SoftAutoCommitTest,
PathHierarchyTokenizerFactoryTest, RequiredFieldsTest,
FastVectorHighlighterTest, TestCollationField, QueryParsingTest,
TestComponentsName, TestLFUCache, CleanupOldIndexTest, ConfigSetsAPITest,
DeleteInactiveReplicaTest, DeleteNodeTest,
DistribDocExpirationUpdateProcessorTest, LeaderElectionContextKeyTest,
LeaderFailoverAfterPartitionTest]
[junit4] Completed [504/693 (1!)] on J2 in 128.73s, 1 test, 1 failure <<<
FAILURES!
[...truncated 63230 lines...]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]