Build: https://jenkins.thetaphi.de/job/Lucene-Solr-6.x-Solaris/946/
Java: 64bit/jdk1.8.0 -XX:+UseCompressedOops -XX:+UseG1GC
3 tests failed.
FAILED: org.apache.solr.index.hdfs.CheckHdfsIndexTest.testChecksumsOnly
Error Message:
java.util.concurrent.TimeoutException: Could not connect to ZooKeeper
127.0.0.1:58165 within 45000 ms
Stack Trace:
org.apache.solr.common.SolrException: java.util.concurrent.TimeoutException:
Could not connect to ZooKeeper 127.0.0.1:58165 within 45000 ms
at
__randomizedtesting.SeedInfo.seed([7C4344571678932E:DEA68D74F4B76BEE]:0)
at
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:183)
at
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:117)
at
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:107)
at
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:86)
at
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:80)
at
org.apache.solr.cloud.AbstractDistribZkTestBase.distribSetUp(AbstractDistribZkTestBase.java:80)
at
org.apache.solr.cloud.AbstractFullDistribZkTestBase.distribSetUp(AbstractFullDistribZkTestBase.java:218)
at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:962)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:916)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:802)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:852)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.util.concurrent.TimeoutException: Could not connect to
ZooKeeper 127.0.0.1:58165 within 45000 ms
at
org.apache.solr.common.cloud.ConnectionManager.waitForConnected(ConnectionManager.java:233)
at
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:175)
... 38 more
FAILED: junit.framework.TestSuite.org.apache.solr.index.hdfs.CheckHdfsIndexTest
Error Message:
5 threads leaked from SUITE scope at
org.apache.solr.index.hdfs.CheckHdfsIndexTest: 1) Thread[id=7778,
name=SessionTracker, state=TIMED_WAITING, group=TGRP-CheckHdfsIndexTest]
at java.lang.Object.wait(Native Method) at
org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:146)
2) Thread[id=7780, name=ProcessThread(sid:0 cport:58165):, state=WAITING,
group=TGRP-CheckHdfsIndexTest] at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:122)
3) Thread[id=7777, name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0,
state=RUNNABLE, group=TGRP-CheckHdfsIndexTest] at
sun.nio.ch.DevPollArrayWrapper.poll0(Native Method) at
sun.nio.ch.DevPollArrayWrapper.poll(DevPollArrayWrapper.java:223) at
sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:98) at
sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) at
sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) at
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
at java.lang.Thread.run(Thread.java:748) 4) Thread[id=7779,
name=SyncThread:0, state=WAITING, group=TGRP-CheckHdfsIndexTest] at
sun.misc.Unsafe.park(Native Method) at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:127)
5) Thread[id=7776, name=Thread-2981, state=WAITING,
group=TGRP-CheckHdfsIndexTest] at java.lang.Object.wait(Native Method)
at java.lang.Thread.join(Thread.java:1252) at
java.lang.Thread.join(Thread.java:1326) at
org.apache.zookeeper.server.NIOServerCnxnFactory.join(NIOServerCnxnFactory.java:297)
at
org.apache.solr.cloud.ZkTestServer$ZKServerMain.runFromConfig(ZkTestServer.java:309)
at org.apache.solr.cloud.ZkTestServer$2.run(ZkTestServer.java:490)
Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: 5 threads leaked from SUITE
scope at org.apache.solr.index.hdfs.CheckHdfsIndexTest:
1) Thread[id=7778, name=SessionTracker, state=TIMED_WAITING,
group=TGRP-CheckHdfsIndexTest]
at java.lang.Object.wait(Native Method)
at
org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:146)
2) Thread[id=7780, name=ProcessThread(sid:0 cport:58165):, state=WAITING,
group=TGRP-CheckHdfsIndexTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:122)
3) Thread[id=7777, name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0,
state=RUNNABLE, group=TGRP-CheckHdfsIndexTest]
at sun.nio.ch.DevPollArrayWrapper.poll0(Native Method)
at sun.nio.ch.DevPollArrayWrapper.poll(DevPollArrayWrapper.java:223)
at sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:98)
at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
at
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
at java.lang.Thread.run(Thread.java:748)
4) Thread[id=7779, name=SyncThread:0, state=WAITING,
group=TGRP-CheckHdfsIndexTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at
org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:127)
5) Thread[id=7776, name=Thread-2981, state=WAITING,
group=TGRP-CheckHdfsIndexTest]
at java.lang.Object.wait(Native Method)
at java.lang.Thread.join(Thread.java:1252)
at java.lang.Thread.join(Thread.java:1326)
at
org.apache.zookeeper.server.NIOServerCnxnFactory.join(NIOServerCnxnFactory.java:297)
at
org.apache.solr.cloud.ZkTestServer$ZKServerMain.runFromConfig(ZkTestServer.java:309)
at org.apache.solr.cloud.ZkTestServer$2.run(ZkTestServer.java:490)
at __randomizedtesting.SeedInfo.seed([7C4344571678932E]:0)
FAILED: junit.framework.TestSuite.org.apache.solr.index.hdfs.CheckHdfsIndexTest
Error Message:
There are still zombie threads that couldn't be terminated: 1)
Thread[id=7777, name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0, state=RUNNABLE,
group=TGRP-CheckHdfsIndexTest] at sun.nio.ch.IOUtil.drain(Native
Method) at
sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:109)
at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) at
sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) at
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
at java.lang.Thread.run(Thread.java:748)
Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie
threads that couldn't be terminated:
1) Thread[id=7777, name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0,
state=RUNNABLE, group=TGRP-CheckHdfsIndexTest]
at sun.nio.ch.IOUtil.drain(Native Method)
at sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:109)
at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
at
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
at java.lang.Thread.run(Thread.java:748)
at __randomizedtesting.SeedInfo.seed([7C4344571678932E]:0)
Build Log:
[...truncated 11338 lines...]
[junit4] Suite: org.apache.solr.index.hdfs.CheckHdfsIndexTest
[junit4] 2> Creating dataDir:
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/init-core-data-001
[junit4] 2> 752884 WARN
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ]
o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=1 numCloses=1
[junit4] 2> 752884 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ]
o.a.s.SolrTestCaseJ4 Using PointFields
[junit4] 2> 752885 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ]
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via:
@org.apache.solr.util.RandomizeSSL(reason=, value=NaN, ssl=NaN, clientAuth=NaN)
[junit4] 2> 752885 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ]
o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /
[junit4] 1> Formatting using clusterid: testClusterID
[junit4] 2> 752907 WARN
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ]
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[junit4] 2> 752913 WARN
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 752915 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ] o.m.log
jetty-6.1.26
[junit4] 2> 752925 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ] o.m.log
Extract
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.2-tests.jar!/webapps/hdfs
to ./temp/Jetty_serv1.vm3_33624_hdfs____tg9mmx/webapp
[junit4] 2> 753434 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ] o.m.log
Started HttpServer2$SelectChannelConnectorWithSafeStartup@serv1-vm3:33624
[junit4] 2> 753547 WARN
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 753548 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ] o.m.log
jetty-6.1.26
[junit4] 2> 753612 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ] o.m.log
Extract
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.2-tests.jar!/webapps/datanode
to ./temp/Jetty_localhost_36641_datanode____.jmw2zt/webapp
[junit4] 2> 754087 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ] o.m.log
Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:36641
[junit4] 2> 754162 WARN
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 754177 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ] o.m.log
jetty-6.1.26
[junit4] 2> 754195 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ] o.m.log
Extract
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.2-tests.jar!/webapps/datanode
to ./temp/Jetty_localhost_60939_datanode____g07nw/webapp
[junit4] 2> 754240 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* processReport: from storage
DS-705520ae-634a-486d-a7b9-524dede0ad8d node
DatanodeRegistration(127.0.0.1:34809,
datanodeUuid=2febcede-2f41-4aaa-9139-438e267ad018, infoPort=58165,
infoSecurePort=0, ipcPort=36779,
storageInfo=lv=-56;cid=testClusterID;nsid=1990931026;c=0), blocks: 0,
hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 754240 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* processReport: from storage
DS-8b47dc82-aa25-425b-a22e-33227a176a14 node
DatanodeRegistration(127.0.0.1:34809,
datanodeUuid=2febcede-2f41-4aaa-9139-438e267ad018, infoPort=58165,
infoSecurePort=0, ipcPort=36779,
storageInfo=lv=-56;cid=testClusterID;nsid=1990931026;c=0), blocks: 0,
hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 754716 INFO
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [ ] o.m.log
Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:60939
[junit4] 2> 754819 INFO (IPC Server handler 7 on 62193) [ ]
BlockStateChange BLOCK* processReport: from storage
DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5 node
DatanodeRegistration(127.0.0.1:48010,
datanodeUuid=142b4244-02b0-4ab4-9bac-7983fe2b175d, infoPort=58917,
infoSecurePort=0, ipcPort=58493,
storageInfo=lv=-56;cid=testClusterID;nsid=1990931026;c=0), blocks: 0,
hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 754819 INFO (IPC Server handler 7 on 62193) [ ]
BlockStateChange BLOCK* processReport: from storage
DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065 node
DatanodeRegistration(127.0.0.1:48010,
datanodeUuid=142b4244-02b0-4ab4-9bac-7983fe2b175d, infoPort=58917,
infoSecurePort=0, ipcPort=58493,
storageInfo=lv=-56;cid=testClusterID;nsid=1990931026;c=0), blocks: 0,
hasStaleStorage: false, processing time: 1 msecs
[junit4] 2> 754909 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnly-seed#[7C4344571678932E]) [ ]
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 754910 INFO (Thread-2981) [ ] o.a.s.c.ZkTestServer client
port:0.0.0.0/0.0.0.0:0
[junit4] 2> 754910 INFO (Thread-2981) [ ] o.a.s.c.ZkTestServer
Starting server
[junit4] 2> 754913 ERROR (Thread-2981) [ ] o.a.z.s.ZooKeeperServer
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action
on ERROR or SHUTDOWN server state changes
[junit4] 2> 755010 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnly-seed#[7C4344571678932E]) [ ]
o.a.s.c.ZkTestServer start zk server on port:58165
[junit4] 2> 800020 WARN
(TEST-CheckHdfsIndexTest.testChecksumsOnly-seed#[7C4344571678932E]-SendThread(127.0.0.1:58165))
[ ] o.a.z.ClientCnxn Client session timed out, have not heard from server
in 45009ms for sessionid 0x0
[junit4] 2> NOTE: reproduce with: ant test -Dtestcase=CheckHdfsIndexTest
-Dtests.method=testChecksumsOnly -Dtests.seed=7C4344571678932E
-Dtests.slow=true -Dtests.locale=ga -Dtests.timezone=Europe/Samara
-Dtests.asserts=true -Dtests.file.encoding=UTF-8
[junit4] ERROR 45.5s J0 | CheckHdfsIndexTest.testChecksumsOnly <<<
[junit4] > Throwable #1: org.apache.solr.common.SolrException:
java.util.concurrent.TimeoutException: Could not connect to ZooKeeper
127.0.0.1:58165 within 45000 ms
[junit4] > at
__randomizedtesting.SeedInfo.seed([7C4344571678932E:DEA68D74F4B76BEE]:0)
[junit4] > at
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:183)
[junit4] > at
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:117)
[junit4] > at
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:107)
[junit4] > at
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:86)
[junit4] > at
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:80)
[junit4] > at
org.apache.solr.cloud.AbstractDistribZkTestBase.distribSetUp(AbstractDistribZkTestBase.java:80)
[junit4] > at
org.apache.solr.cloud.AbstractFullDistribZkTestBase.distribSetUp(AbstractFullDistribZkTestBase.java:218)
[junit4] > at
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:962)
[junit4] > at java.lang.Thread.run(Thread.java:748)
[junit4] > Caused by: java.util.concurrent.TimeoutException: Could not
connect to ZooKeeper 127.0.0.1:58165 within 45000 ms
[junit4] > at
org.apache.solr.common.cloud.ConnectionManager.waitForConnected(ConnectionManager.java:233)
[junit4] > at
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:175)
[junit4] > ... 38 more
[junit4] 2> 800425 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 800425 INFO (Thread-2983) [ ] o.a.s.c.ZkTestServer client
port:0.0.0.0/0.0.0.0:0
[junit4] 2> 800425 INFO (Thread-2983) [ ] o.a.s.c.ZkTestServer
Starting server
[junit4] 2> 800427 ERROR (Thread-2983) [ ] o.a.z.s.ZooKeeperServer
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action
on ERROR or SHUTDOWN server state changes
[junit4] 2> 800527 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.ZkTestServer start zk server on port:49657
[junit4] 2> 800550 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
to /configs/conf1/solrconfig.xml
[junit4] 2> 800553 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/schema.xml
to /configs/conf1/schema.xml
[junit4] 2> 800556 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
[junit4] 2> 800559 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
to /configs/conf1/stopwords.txt
[junit4] 2> 800564 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/protwords.txt
to /configs/conf1/protwords.txt
[junit4] 2> 800567 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/currency.xml
to /configs/conf1/currency.xml
[junit4] 2> 800572 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml
to /configs/conf1/enumsConfig.xml
[junit4] 2> 800575 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json
to /configs/conf1/open-exchange-rates.json
[junit4] 2> 800578 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt
to /configs/conf1/mapping-ISOLatin1Accent.txt
[junit4] 2> 800581 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt
to /configs/conf1/old_synonyms.txt
[junit4] 2> 800583 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/synonyms.txt
to /configs/conf1/synonyms.txt
[junit4] 2> 800758 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.SolrTestCaseJ4 Writing core.properties file to
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/control-001/cores/collection1
[junit4] 2> 800759 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.Server jetty-9.3.14.v20161028
[junit4] 2> 800761 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@6eff6b56{/,null,AVAILABLE}
[junit4] 2> 800761 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.AbstractConnector Started
ServerConnector@4811b32f{HTTP/1.1,[http/1.1]}{127.0.0.1:58678}
[junit4] 2> 800761 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.Server Started @803144ms
[junit4] 2> 800761 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=hdfs://serv1-vm3:62193/hdfs__serv1-vm3_62193__export_home_jenkins_workspace_Lucene-Solr-6.x-Solaris_solr_build_solr-core_test_J0_temp_solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001_tempDir-003_control_data,
hostContext=/, hostPort=58678,
coreRootDirectory=/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/../../../../../../../../../../export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/control-001/cores}
[junit4] 2> 800762 ERROR
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 800762 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
6.7.0
[junit4] 2> 800762 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port
null
[junit4] 2> 800762 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 800762 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2017-07-03T09:54:15.678Z
[junit4] 2> 800770 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 800770 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.SolrXmlConfig Loading container configuration from
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/control-001/solr.xml
[junit4] 2> 800785 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.u.UpdateShardHandler Creating UpdateShardHandler HTTP client with
params: socketTimeout=340000&connTimeout=45000&retry=true
[junit4] 2> 800786 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:49657/solr
[junit4] 2> 800840 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E])
[n:127.0.0.1:58678_ ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 800841 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E])
[n:127.0.0.1:58678_ ] o.a.s.c.OverseerElectionContext I am going to be the
leader 127.0.0.1:58678_
[junit4] 2> 800843 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E])
[n:127.0.0.1:58678_ ] o.a.s.c.Overseer Overseer
(id=98243422148493316-127.0.0.1:58678_-n_0000000000) starting
[junit4] 2> 800856 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E])
[n:127.0.0.1:58678_ ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:58678_
[junit4] 2> 800859 INFO
(zkCallback-1148-thread-1-processing-n:127.0.0.1:58678_) [n:127.0.0.1:58678_
] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 800991 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E])
[n:127.0.0.1:58678_ ] o.a.s.c.CorePropertiesLocator Found 1 core definitions
underneath
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/../../../../../../../../../../export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/control-001/cores
[junit4] 2> 800991 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E])
[n:127.0.0.1:58678_ ] o.a.s.c.CorePropertiesLocator Cores are: [collection1]
[junit4] 2> 800995 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4
transient cores
[junit4] 2> 800999 INFO
(OverseerStateUpdate-98243422148493316-127.0.0.1:58678_-n_0000000000)
[n:127.0.0.1:58678_ ] o.a.s.c.o.ReplicaMutator Assigning new node to shard
shard=shard1
[junit4] 2> 801019 WARN
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.c.Config
Beginning with Solr 5.5, <mergePolicy> is deprecated, use <mergePolicyFactory>
instead.
[junit4] 2> 801020 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.c.SolrConfig
Using Lucene MatchVersion: 6.7.0
[junit4] 2> 801060 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.s.IndexSchema
[collection1] Schema name=test
[junit4] 2> 801163 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.s.IndexSchema
Loaded schema test/1.0 with uniqueid field id
[junit4] 2> 801191 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.c.CoreContainer
Creating SolrCore 'collection1' using configuration from collection
control_collection, trusted=true
[junit4] 2> 801191 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.HdfsDirectoryFactory
solr.hdfs.home=hdfs://serv1-vm3:62193/solr_hdfs_home
[junit4] 2> 801191 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
[junit4] 2> 801191 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.c.SolrCore
solr.RecoveryStrategy.Builder
[junit4] 2> 801191 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.c.SolrCore
[[collection1] ] Opening new SolrCore at
[/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/control-001/cores/collection1],
dataDir=[hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data/]
[junit4] 2> 801191 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.JmxMonitoredMap JMX monitoring is enabled. Adding Solr mbeans to JMX
Server: com.sun.jmx.mbeanserver.JmxMBeanServer@16fda84b
[junit4] 2> 801192 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.HdfsDirectoryFactory creating directory factory for path
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data/snapshot_metadata
[junit4] 2> 801201 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct
memory allocation set to [true]
[junit4] 2> 801201 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of
[8388608] will allocate [1] slabs and use ~[8388608] bytes
[junit4] 2> 801208 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.s.b.BlockDirectory Block cache on write is disabled
[junit4] 2> 801210 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.HdfsDirectoryFactory creating directory factory for path
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data
[junit4] 2> 801230 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.HdfsDirectoryFactory creating directory factory for path
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data/index
[junit4] 2> 801237 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct
memory allocation set to [true]
[junit4] 2> 801237 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of
[8388608] will allocate [1] slabs and use ~[8388608] bytes
[junit4] 2> 801242 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.s.b.BlockDirectory Block cache on write is disabled
[junit4] 2> 801243 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy:
maxMergeAtOnce=18, maxMergeAtOnceExplicit=44, maxMergedSegmentMB=13.416015625,
floorSegmentMB=1.5537109375, forceMergeDeletesPctAllowed=24.94718922912415,
segmentsPerTier=15.0, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.2954474155087866
[junit4] 2> 801266 INFO (IPC Server handler 8 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741825_1001{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 801267 INFO (IPC Server handler 4 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741825_1001{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED],
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
size 0
[junit4] 2> 801292 WARN
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type =
requestHandler,name = /dump,class = DumpRequestHandler,attributes =
{initParams=a, name=/dump, class=DumpRequestHandler},args =
{defaults={a=A,b=B}}}
[junit4] 2> 801348 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.u.UpdateHandler
Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 801348 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.u.UpdateLog
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 801348 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.u.HdfsUpdateLog
Initializing HdfsUpdateLog: tlogDfsReplication=2
[junit4] 2> 801361 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.u.CommitTracker
Hard AutoCommit: disabled
[junit4] 2> 801361 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.u.CommitTracker
Soft AutoCommit: disabled
[junit4] 2> 801362 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class
org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy:
minMergeSize=1000, mergeFactor=44, maxMergeSize=9223372036854775807,
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false,
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.8665622297044102]
[junit4] 2> 801375 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.s.SolrIndexSearcher Opening [Searcher@753f972f[collection1] main]
[junit4] 2> 801378 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 801378 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 801378 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.h.ReplicationHandler Commits will be reserved for 10000
[junit4] 2> 801382 INFO
(searcherExecutor-3028-thread-1-processing-n:127.0.0.1:58678_ x:collection1
c:control_collection) [n:127.0.0.1:58678_ c:control_collection x:collection1]
o.a.s.c.SolrCore [collection1] Registered new searcher
Searcher@753f972f[collection1]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 801385 INFO
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_)
[n:127.0.0.1:58678_ c:control_collection x:collection1] o.a.s.u.UpdateLog
Could not find max version in index or recent updates, using new clock
1571894755381477376
[junit4] 2> 801397 INFO
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext Enough replicas
found to continue.
[junit4] 2> 801397 INFO
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext I may be the new
leader - try and sync
[junit4] 2> 801398 INFO
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.SyncStrategy Sync replicas to
http://127.0.0.1:58678/collection1/
[junit4] 2> 801398 INFO
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.SyncStrategy Sync Success - now sync
replicas to me
[junit4] 2> 801398 INFO
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.SyncStrategy
http://127.0.0.1:58678/collection1/ has no replicas
[junit4] 2> 801398 INFO
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext Found all
replicas participating in election, clear LIR
[junit4] 2> 801406 INFO
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext I am the new
leader: http://127.0.0.1:58678/collection1/ shard1
[junit4] 2> 801511 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 801514 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:49657/solr ready
[junit4] 2> 801514 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection
loss:false
[junit4] 2> 801559 INFO
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1
r:core_node1 x:collection1] o.a.s.c.ZkController I am the leader, no recovery
necessary
[junit4] 2> 801662 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.SolrTestCaseJ4 Writing core.properties file to
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001/cores/collection1
[junit4] 2> 801663 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001
[junit4] 2> 801664 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.Server jetty-9.3.14.v20161028
[junit4] 2> 801675 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@5e7d866{/,null,AVAILABLE}
[junit4] 2> 801676 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.AbstractConnector Started
ServerConnector@2127e18a{HTTP/1.1,[http/1.1]}{127.0.0.1:40887}
[junit4] 2> 801684 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.Server Started @804066ms
[junit4] 2> 801684 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.s.e.JettySolrRunner Jetty properties:
{solr.data.dir=hdfs://serv1-vm3:62193/hdfs__serv1-vm3_62193__export_home_jenkins_workspace_Lucene-Solr-6.x-Solaris_solr_build_solr-core_test_J0_temp_solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001_tempDir-003_jetty1,
solrconfig=solrconfig.xml, hostContext=/, hostPort=40887,
coreRootDirectory=/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001/cores}
[junit4] 2> 801684 ERROR
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 801685 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr™ version
6.7.0
[junit4] 2> 801685 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port
null
[junit4] 2> 801685 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 801685 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2017-07-03T09:54:16.601Z
[junit4] 2> 801695 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in
ZooKeeper)
[junit4] 2> 801695 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.SolrXmlConfig Loading container configuration from
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001/solr.xml
[junit4] 2> 801702 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.u.UpdateShardHandler Creating UpdateShardHandler HTTP client with
params: socketTimeout=340000&connTimeout=45000&retry=true
[junit4] 2> 801703 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:49657/solr
[junit4] 2> 801733 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E])
[n:127.0.0.1:40887_ ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (1)
[junit4] 2> 801738 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E])
[n:127.0.0.1:40887_ ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 801742 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E])
[n:127.0.0.1:40887_ ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:40887_
[junit4] 2> 801745 INFO
(zkCallback-1148-thread-1-processing-n:127.0.0.1:58678_) [n:127.0.0.1:58678_
] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 801745 INFO (zkCallback-1152-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 801746 INFO
(zkCallback-1157-thread-1-processing-n:127.0.0.1:40887_) [n:127.0.0.1:40887_
] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 801816 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E])
[n:127.0.0.1:40887_ ] o.a.s.c.CorePropertiesLocator Found 1 core definitions
underneath
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001/cores
[junit4] 2> 801816 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E])
[n:127.0.0.1:40887_ ] o.a.s.c.CorePropertiesLocator Cores are: [collection1]
[junit4] 2> 801817 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1]
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4
transient cores
[junit4] 2> 801819 INFO
(OverseerStateUpdate-98243422148493316-127.0.0.1:58678_-n_0000000000)
[n:127.0.0.1:58678_ ] o.a.s.c.o.ReplicaMutator Assigning new node to shard
shard=shard1
[junit4] 2> 802844 WARN
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.Config Beginning
with Solr 5.5, <mergePolicy> is deprecated, use <mergePolicyFactory> instead.
[junit4] 2> 802845 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.SolrConfig Using
Lucene MatchVersion: 6.7.0
[junit4] 2> 802858 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.s.IndexSchema
[collection1] Schema name=test
[junit4] 2> 802962 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.s.IndexSchema Loaded
schema test/1.0 with uniqueid field id
[junit4] 2> 802988 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.CoreContainer
Creating SolrCore 'collection1' using configuration from collection
collection1, trusted=true
[junit4] 2> 802988 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.HdfsDirectoryFactory
solr.hdfs.home=hdfs://serv1-vm3:62193/solr_hdfs_home
[junit4] 2> 802988 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.HdfsDirectoryFactory
Solr Kerberos Authentication disabled
[junit4] 2> 802988 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.SolrCore
solr.RecoveryStrategy.Builder
[junit4] 2> 802988 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.SolrCore
[[collection1] ] Opening new SolrCore at
[/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001/cores/collection1],
dataDir=[hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data/]
[junit4] 2> 802989 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.JmxMonitoredMap JMX
monitoring is enabled. Adding Solr mbeans to JMX Server:
com.sun.jmx.mbeanserver.JmxMBeanServer@16fda84b
[junit4] 2> 802989 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.HdfsDirectoryFactory
creating directory factory for path
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data/snapshot_metadata
[junit4] 2> 802999 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.HdfsDirectoryFactory
Number of slabs of block cache [1] with direct memory allocation set to [true]
[junit4] 2> 802999 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.HdfsDirectoryFactory
Block cache target memory usage, slab size of [8388608] will allocate [1] slabs
and use ~[8388608] bytes
[junit4] 2> 803003 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.s.b.BlockDirectory
Block cache on write is disabled
[junit4] 2> 803004 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.HdfsDirectoryFactory
creating directory factory for path
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data
[junit4] 2> 803024 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.HdfsDirectoryFactory
creating directory factory for path
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data/index
[junit4] 2> 803033 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.HdfsDirectoryFactory
Number of slabs of block cache [1] with direct memory allocation set to [true]
[junit4] 2> 803033 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.HdfsDirectoryFactory
Block cache target memory usage, slab size of [8388608] will allocate [1] slabs
and use ~[8388608] bytes
[junit4] 2> 803037 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.s.b.BlockDirectory
Block cache on write is disabled
[junit4] 2> 803038 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.u.RandomMergePolicy
RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy:
[TieredMergePolicy: maxMergeAtOnce=18, maxMergeAtOnceExplicit=44,
maxMergedSegmentMB=13.416015625, floorSegmentMB=1.5537109375,
forceMergeDeletesPctAllowed=24.94718922912415, segmentsPerTier=15.0,
maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.2954474155087866
[junit4] 2> 803058 INFO (IPC Server handler 5 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741826_1002{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 803061 INFO (IPC Server handler 0 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741826_1002{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 803071 WARN
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.c.RequestHandlers
INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class
= DumpRequestHandler,attributes = {initParams=a, name=/dump,
class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
[junit4] 2> 803127 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.u.UpdateHandler Using
UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
[junit4] 2> 803127 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.u.UpdateLog
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 803127 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.u.HdfsUpdateLog
Initializing HdfsUpdateLog: tlogDfsReplication=2
[junit4] 2> 803142 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.u.CommitTracker Hard
AutoCommit: disabled
[junit4] 2> 803142 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.u.CommitTracker Soft
AutoCommit: disabled
[junit4] 2> 803144 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.u.RandomMergePolicy
RandomMergePolicy wrapping class org.apache.lucene.index.LogDocMergePolicy:
[LogDocMergePolicy: minMergeSize=1000, mergeFactor=44,
maxMergeSize=9223372036854775807,
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false,
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12,
noCFSRatio=0.8665622297044102]
[junit4] 2> 803153 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.s.SolrIndexSearcher
Opening [Searcher@471dd8eb[collection1] main]
[junit4] 2> 803155 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1]
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase:
/configs/conf1
[junit4] 2> 803156 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1]
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 803156 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.h.ReplicationHandler
Commits will be reserved for 10000
[junit4] 2> 803163 INFO
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_)
[n:127.0.0.1:40887_ c:collection1 x:collection1] o.a.s.u.UpdateLog Could not
find max version in index or recent updates, using new clock 1571894757245845504
[junit4] 2> 803176 INFO
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1
x:collection1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to
continue.
[junit4] 2> 803176 INFO
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1
x:collection1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try
and sync
[junit4] 2> 803176 INFO
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1
x:collection1] o.a.s.c.SyncStrategy Sync replicas to
http://127.0.0.1:40887/collection1/
[junit4] 2> 803176 INFO
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1
x:collection1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 803176 INFO
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1
x:collection1] o.a.s.c.SyncStrategy http://127.0.0.1:40887/collection1/ has no
replicas
[junit4] 2> 803176 INFO
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1
x:collection1] o.a.s.c.ShardLeaderElectionContext Found all replicas
participating in election, clear LIR
[junit4] 2> 803180 INFO
(searcherExecutor-3039-thread-1-processing-n:127.0.0.1:40887_ x:collection1
c:collection1) [n:127.0.0.1:40887_ c:collection1 x:collection1]
o.a.s.c.SolrCore [collection1] Registered new searcher
Searcher@471dd8eb[collection1]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 803181 INFO
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1
x:collection1] o.a.s.c.ShardLeaderElectionContext I am the new leader:
http://127.0.0.1:40887/collection1/ shard1
[junit4] 2> 803333 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.SolrTestCaseJ4 ###Starting testChecksumsOnlyVerbose
[junit4] 2> 803336 INFO
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1
x:collection1] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 803425 INFO (IPC Server handler 4 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741827_1003{UCState=COMMITTED, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW]]}
size 261
[junit4] 2> 803426 INFO (IPC Server handler 1 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741827_1003 size 261
[junit4] 2> 803834 INFO (IPC Server handler 7 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741828_1004{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW]]}
size 0
[junit4] 2> 803834 INFO (IPC Server handler 3 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741828_1004{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|FINALIZED]]}
size 0
[junit4] 2> 803846 INFO (IPC Server handler 1 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741829_1005{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
size 0
[junit4] 2> 803846 INFO (IPC Server handler 4 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741829_1005{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
size 0
[junit4] 2> 803857 INFO (IPC Server handler 7 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741830_1006{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW]]}
size 0
[junit4] 2> 803858 INFO (IPC Server handler 3 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741830_1006{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW]]}
size 0
[junit4] 2> 803868 INFO (IPC Server handler 1 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741831_1007{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
size 0
[junit4] 2> 803869 INFO (IPC Server handler 4 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741831_1007{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED],
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 803880 INFO (IPC Server handler 7 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741832_1008{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW]]}
size 0
[junit4] 2> 803880 INFO (IPC Server handler 3 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741832_1008{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW]]}
size 0
[junit4] 2> 803889 INFO (IPC Server handler 0 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741833_1009{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 803889 INFO (IPC Server handler 8 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741833_1009{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 803895 INFO (IPC Server handler 2 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741834_1010{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 803899 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741834_1010{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED],
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|FINALIZED]]}
size 0
[junit4] 2> 803907 INFO (IPC Server handler 9 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741835_1011{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
size 0
[junit4] 2> 803907 INFO (IPC Server handler 9 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741835_1011{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
size 0
[junit4] 2> 803928 INFO (IPC Server handler 1 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741836_1012{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 803928 INFO (IPC Server handler 4 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741836_1012{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED],
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|FINALIZED]]}
size 0
[junit4] 2> 803996 INFO (IPC Server handler 0 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741837_1013{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 803996 INFO (IPC Server handler 8 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741837_1013{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED],
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
size 0
[junit4] 2> 804002 INFO (IPC Server handler 2 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741838_1014{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|FINALIZED]]}
size 0
[junit4] 2> 804003 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741838_1014{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|FINALIZED],
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 804010 INFO (IPC Server handler 9 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741839_1015{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW]]}
size 0
[junit4] 2> 804011 INFO (IPC Server handler 5 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741839_1015 size 5711
[junit4] 2> 804024 INFO (IPC Server handler 1 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741840_1016{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 804025 INFO (IPC Server handler 4 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741840_1016{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 804032 INFO (IPC Server handler 3 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741841_1017{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW]]}
size 0
[junit4] 2> 804032 INFO (IPC Server handler 9 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741841_1017{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW]]}
size 0
[junit4] 2> 804048 INFO (IPC Server handler 3 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741842_1018{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW]]}
size 0
[junit4] 2> 804048 INFO (IPC Server handler 9 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741842_1018{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW]]}
size 0
[junit4] 2> 804055 INFO (IPC Server handler 1 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is
added to blk_1073741843_1019{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW]]}
size 0
[junit4] 2> 804056 INFO (IPC Server handler 2 on 62193) [ ]
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is
added to blk_1073741843_1019{UCState=UNDER_CONSTRUCTION, truncateBlock=null,
primaryNodeIndex=-1,
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
size 0
[junit4] 2> 804517 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.s.h.HdfsDirectory Closing hdfs directory hdfs://serv1-vm3:62193/solr
[junit4] 2> 804518 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741831_1007 127.0.0.1:34809
127.0.0.1:48010
[junit4] 2> 804518 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741832_1008 127.0.0.1:48010
127.0.0.1:34809
[junit4] 2> 804518 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741833_1009 127.0.0.1:48010
127.0.0.1:34809
[junit4] 2> 804518 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741834_1010 127.0.0.1:48010
127.0.0.1:34809
[junit4] 2> 804518 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741841_1017 127.0.0.1:34809
127.0.0.1:48010
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741827_1003 127.0.0.1:48010
127.0.0.1:34809
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741828_1004 127.0.0.1:48010
127.0.0.1:34809
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741842_1018 127.0.0.1:34809
127.0.0.1:48010
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741835_1011 127.0.0.1:34809
127.0.0.1:48010
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741836_1012 127.0.0.1:48010
127.0.0.1:34809
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741839_1015 127.0.0.1:48010
127.0.0.1:34809
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741840_1016 127.0.0.1:48010
127.0.0.1:34809
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741837_1013 127.0.0.1:48010
127.0.0.1:34809
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741838_1014 127.0.0.1:34809
127.0.0.1:48010
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741829_1005 127.0.0.1:34809
127.0.0.1:48010
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741830_1006 127.0.0.1:34809
127.0.0.1:48010
[junit4] 2> 804519 INFO (IPC Server handler 6 on 62193) [ ]
BlockStateChange BLOCK* addToInvalidates: blk_1073741843_1019 127.0.0.1:34809
127.0.0.1:48010
[junit4] 2> 804519 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.SolrTestCaseJ4 ###Ending testChecksumsOnlyVerbose
[junit4] 2> 804519 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.ChaosMonkey monkey: stop shard! 58678
[junit4] 2> 804519 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.CoreContainer Shutting down CoreContainer instance=1495132002
[junit4] 2> 804519 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.node
[junit4] 2> 804519 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.jvm
[junit4] 2> 804519 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.jetty
[junit4] 2> 804525 INFO (coreCloseExecutor-3044-thread-1)
[n:127.0.0.1:58678_ c:control_collection s:shard1 r:core_node1 x:collection1]
o.a.s.c.SolrCore [collection1] CLOSING SolrCore
org.apache.solr.core.SolrCore@375d0a45
[junit4] 2> 804548 INFO
(org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@4cc02650)
[ ] BlockStateChange BLOCK* BlockManager: ask 127.0.0.1:34809 to delete
[blk_1073741827_1003, blk_1073741828_1004, blk_1073741829_1005,
blk_1073741830_1006, blk_1073741831_1007, blk_1073741832_1008,
blk_1073741833_1009, blk_1073741834_1010, blk_1073741835_1011,
blk_1073741836_1012, blk_1073741837_1013, blk_1073741838_1014,
blk_1073741839_1015, blk_1073741840_1016, blk_1073741841_1017,
blk_1073741842_1018, blk_1073741843_1019]
[junit4] 2> 804590 INFO (coreCloseExecutor-3044-thread-1)
[n:127.0.0.1:58678_ c:control_collection s:shard1 r:core_node1 x:collection1]
o.a.s.m.SolrMetricManager Closing metric reporters for: solr.core.collection1
[junit4] 2> 804594 INFO (coreCloseExecutor-3044-thread-1)
[n:127.0.0.1:58678_ c:control_collection s:shard1 r:core_node1 x:collection1]
o.a.s.s.h.HdfsDirectory Closing hdfs directory
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data/snapshot_metadata
[junit4] 2> 804595 INFO (coreCloseExecutor-3044-thread-1)
[n:127.0.0.1:58678_ c:control_collection s:shard1 r:core_node1 x:collection1]
o.a.s.s.h.HdfsDirectory Closing hdfs directory
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data
[junit4] 2> 804596 INFO (coreCloseExecutor-3044-thread-1)
[n:127.0.0.1:58678_ c:control_collection s:shard1 r:core_node1 x:collection1]
o.a.s.s.h.HdfsDirectory Closing hdfs directory
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data/index
[junit4] 2> 804596 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.Overseer Overseer
(id=98243422148493316-127.0.0.1:58678_-n_0000000000) closing
[junit4] 2> 804597 INFO
(OverseerStateUpdate-98243422148493316-127.0.0.1:58678_-n_0000000000)
[n:127.0.0.1:58678_ ] o.a.s.c.Overseer Overseer Loop exiting :
127.0.0.1:58678_
[junit4] 2> 804601 WARN
(zkCallback-1148-thread-2-processing-n:127.0.0.1:58678_) [n:127.0.0.1:58678_
] o.a.s.c.c.ZkStateReader ZooKeeper watch triggered, but Solr cannot talk to
ZK: [KeeperErrorCode = Session expired for /live_nodes]
[junit4] 2> 804601 INFO (zkCallback-1152-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (1)
[junit4] 2> 804601 INFO
(zkCallback-1157-thread-2-processing-n:127.0.0.1:40887_) [n:127.0.0.1:40887_
] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (1)
[junit4] 2> 804601 INFO
(zkCallback-1157-thread-1-processing-n:127.0.0.1:40887_) [n:127.0.0.1:40887_
] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:40887_
[junit4] 2> 804602 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.AbstractConnector Stopped
ServerConnector@4811b32f{HTTP/1.1,[http/1.1]}{127.0.0.1:0}
[junit4] 2> 804602 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.h.ContextHandler Stopped
o.e.j.s.ServletContextHandler@6eff6b56{/,null,UNAVAILABLE}
[junit4] 2> 804603 INFO
(zkCallback-1157-thread-1-processing-n:127.0.0.1:40887_) [n:127.0.0.1:40887_
] o.a.s.c.Overseer Overseer
(id=98243422148493320-127.0.0.1:40887_-n_0000000001) starting
[junit4] 2> 804606 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.ChaosMonkey monkey: stop shard! 40887
[junit4] 2> 804606 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.CoreContainer Shutting down CoreContainer instance=1504621779
[junit4] 2> 804607 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.node
[junit4] 2> 804607 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.jvm
[junit4] 2> 804607 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.jetty
[junit4] 2> 804611 INFO (coreCloseExecutor-3046-thread-1)
[n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 x:collection1]
o.a.s.c.SolrCore [collection1] CLOSING SolrCore
org.apache.solr.core.SolrCore@22e16c08
[junit4] 2> 804664 INFO (coreCloseExecutor-3046-thread-1)
[n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 x:collection1]
o.a.s.m.SolrMetricManager Closing metric reporters for: solr.core.collection1
[junit4] 2> 804664 INFO (coreCloseExecutor-3046-thread-1)
[n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 x:collection1]
o.a.s.s.h.HdfsDirectory Closing hdfs directory
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data
[junit4] 2> 804665 INFO (coreCloseExecutor-3046-thread-1)
[n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 x:collection1]
o.a.s.s.h.HdfsDirectory Closing hdfs directory
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data/index
[junit4] 2> 804666 INFO (coreCloseExecutor-3046-thread-1)
[n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 x:collection1]
o.a.s.s.h.HdfsDirectory Closing hdfs directory
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data/snapshot_metadata
[junit4] 2> 804666 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.Overseer Overseer
(id=98243422148493320-127.0.0.1:40887_-n_0000000001) closing
[junit4] 2> 804666 INFO
(OverseerStateUpdate-98243422148493320-127.0.0.1:40887_-n_0000000001)
[n:127.0.0.1:40887_ ] o.a.s.c.Overseer Overseer Loop exiting :
127.0.0.1:40887_
[junit4] 2> 804669 INFO (zkCallback-1152-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (0)
[junit4] 2> 804671 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.AbstractConnector Stopped
ServerConnector@2127e18a{HTTP/1.1,[http/1.1]}{127.0.0.1:0}
[junit4] 2> 804671 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.e.j.s.h.ContextHandler Stopped
o.e.j.s.ServletContextHandler@5e7d866{/,null,UNAVAILABLE}
[junit4] 2> 804676 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.ChaosMonkey monkey: stop shard! 58678
[junit4] 2> 804676 ERROR
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper
server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 804677 INFO
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [
] o.a.s.c.ZkTestServer connecting to 127.0.0.1:49657 49657
[junit4] 2> 804744 INFO (Thread-2983) [ ] o.a.s.c.ZkTestServer
connecting to 127.0.0.1:49657 49657
[junit4] 2> 804746 WARN (Thread-2983) [ ] o.a.s.c.ZkTestServer Watch
limit violations:
[junit4] 2> Maximum concurrent create/delete watches above limit:
[junit4] 2>
[junit4] 2> 3 /solr/aliases.json
[junit4] 2> 3 /solr/clusterprops.json
[junit4] 2> 2 /solr/security.json
[junit4] 2> 2 /solr/configs/conf1
[junit4] 2>
[junit4] 2> Maximum concurrent data watches above limit:
[junit4] 2>
[junit4] 2> 3 /solr/clusterstate.json
[junit4] 2>
[junit4] 2> Maximum concurrent children watches above limit:
[junit4] 2>
[junit4] 2> 3 /solr/live_nodes
[junit4] 2> 3 /solr/collections
[junit4] 2> 2 /solr/overseer/queue
[junit4] 2> 2 /solr/overseer/collection-queue-work
[junit4] 2>
[junit4] 2> 804781 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 804782 INFO (Thread-3124) [ ] o.a.s.c.ZkTestServer client
port:0.0.0.0/0.0.0.0:0
[junit4] 2> 804782 INFO (Thread-3124) [ ] o.a.s.c.ZkTestServer
Starting server
[junit4] 2> 804783 ERROR (Thread-3124) [ ] o.a.z.s.ZooKeeperServer
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action
on ERROR or SHUTDOWN server state changes
[junit4] 2> 804784 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.ZkTestServer start zk server on port:62052
[junit4] 2> 804796 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
to /configs/conf1/solrconfig.xml
[junit4] 2> 804797 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/schema.xml
to /configs/conf1/schema.xml
[junit4] 2> 804799 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
[junit4] 2> 804801 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
to /configs/conf1/stopwords.txt
[junit4] 2> 804803 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/protwords.txt
to /configs/conf1/protwords.txt
[junit4] 2> 804806 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/currency.xml
to /configs/conf1/currency.xml
[junit4] 2> 804807 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml
to /configs/conf1/enumsConfig.xml
[junit4] 2> 804809 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json
to /configs/conf1/open-exchange-rates.json
[junit4] 2> 804810 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt
to /configs/conf1/mapping-ISOLatin1Accent.txt
[junit4] 2> 804812 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [ ]
o.a.s.c.AbstractZkTestCase put
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt
to /configs/conf1/old_synonyms.txt
[junit4] 2> 804813 INFO
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E])
[...truncated too long message...]
edSynchronizer.java:2048)
[junit4] 2> at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
[junit4] 2> at
org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:122)
[junit4] 2> 849734 ERROR (ProcessThread(sid:0 cport:58165):) [ ]
o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper
server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 849734 ERROR (SessionTracker) [ ]
o.a.z.s.ZooKeeperCriticalThread Severe unrecoverable error, from thread :
SessionTracker
[junit4] 2> java.lang.InterruptedException
[junit4] 2> at java.lang.Object.wait(Native Method)
[junit4] 2> at
org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:146)
[junit4] 2> 849736 ERROR (SessionTracker) [ ] o.a.z.s.ZooKeeperServer
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action
on ERROR or SHUTDOWN server state changes
[junit4] 2> 849735 WARN (Thread-2981) [ ] o.a.s.c.ZkTestServer Server
interrupted
[junit4] 2> java.lang.InterruptedException
[junit4] 2> at java.lang.Object.wait(Native Method)
[junit4] 2> at java.lang.Thread.join(Thread.java:1252)
[junit4] 2> at java.lang.Thread.join(Thread.java:1326)
[junit4] 2> at
org.apache.zookeeper.server.NIOServerCnxnFactory.join(NIOServerCnxnFactory.java:297)
[junit4] 2> at
org.apache.solr.cloud.ZkTestServer$ZKServerMain.runFromConfig(ZkTestServer.java:309)
[junit4] 2> at
org.apache.solr.cloud.ZkTestServer$2.run(ZkTestServer.java:490)
[junit4] 2> 849735 ERROR (SyncThread:0) [ ]
o.a.z.s.ZooKeeperCriticalThread Severe unrecoverable error, from thread :
SyncThread:0
[junit4] 2> java.lang.InterruptedException
[junit4] 2> at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.java:2014)
[junit4] 2> at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2048)
[junit4] 2> at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
[junit4] 2> at
org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:127)
[junit4] 2> 849738 ERROR (SyncThread:0) [ ] o.a.z.s.ZooKeeperServer
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action
on ERROR or SHUTDOWN server state changes
[junit4] 2> Jul 03, 2017 9:55:07 AM
com.carrotsearch.randomizedtesting.ThreadLeakControl tryToInterruptAll
[junit4] 2> SEVERE: There are still zombie threads that couldn't be
terminated:
[junit4] 2> 1) Thread[id=7777,
name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0, state=RUNNABLE,
group=TGRP-CheckHdfsIndexTest]
[junit4] 2> at sun.nio.ch.IOUtil.drain(Native Method)
[junit4] 2> at
sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:109)
[junit4] 2> at
sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
[junit4] 2> at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
[junit4] 2> at
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
[junit4] 2> at java.lang.Thread.run(Thread.java:748)
[junit4] 2> NOTE: test params are: codec=Lucene62,
sim=RandomSimilarity(queryNorm=true,coord=yes): {field=LM
Jelinek-Mercer(0.700000), foo=ClassicSimilarity, titleTokenized=IB SPL-D2,
body=DFR I(F)L1}, locale=ga, timezone=Europe/Samara
[junit4] 2> NOTE: SunOS 5.11 amd64/Oracle Corporation 1.8.0_131
(64-bit)/cpus=3,threads=3,free=198249632,total=536870912
[junit4] 2> NOTE: All tests run in this JVM: [DirectoryFactoryTest,
TestSolrConfigHandlerConcurrent, CloudExitableDirectoryReaderTest,
OverriddenZkACLAndCredentialsProvidersTest, HdfsUnloadDistributedZkTest,
SolrSlf4jReporterTest, TestTrackingShardHandlerFactory, TestConfigReload,
LoggingHandlerTest, CdcrReplicationDistributedZkTest, TestManagedSchema,
OverseerModifyCollectionTest, GraphQueryTest, TestLazyCores,
HdfsRestartWhileUpdatingTest, AtomicUpdatesTest, TestDFRSimilarityFactory,
TestManagedSynonymFilterFactory, TestFieldTypeResource, SolrCloudExampleTest,
TestImplicitCoreProperties, TestTestInjection, PrimitiveFieldTypeTest,
DocValuesNotIndexedTest, TestRuleBasedAuthorizationPlugin, RulesTest,
TestRecoveryHdfs, MigrateRouteKeyTest, TestAtomicUpdateErrorCases,
TestRandomRequestDistribution, TestTrieFacet, SolrInfoMBeanTest,
TestMaxTokenLenTokenizer, FieldAnalysisRequestHandlerTest,
AnalyticsMergeStrategyTest, RollingRestartTest, TemplateUpdateProcessorTest,
CloneFieldUpdateProcessorFactoryTest, TestStressVersions,
TestReplicationHandlerBackup, SuggestComponentContextFilterQueryTest,
TestDefaultStatsCache, TestPseudoReturnFields, UpdateParamsTest,
SpellCheckCollatorWithCollapseTest, TestDistributedStatsComponentCardinality,
DirectUpdateHandlerTest, TestOrdValues, DeleteLastCustomShardedReplicaTest,
LeaderInitiatedRecoveryOnCommitTest, RequestLoggingTest,
TestLeaderElectionWithEmptyReplica, HdfsBasicDistributedZk2Test,
DistanceFunctionTest, DocumentBuilderTest, DistributedFacetPivotSmallTest,
DistributedSpellCheckComponentTest, TestCursorMarkWithoutUniqueKey,
TestQuerySenderNoQuery, AddSchemaFieldsUpdateProcessorFactoryTest,
SparseHLLTest, SortByFunctionTest, LeaderFailoverAfterPartitionTest,
SuggesterWFSTTest, TestDistributedSearch, TestQuerySenderListener,
TestFaceting, TestRawResponseWriter, TestFreeTextSuggestions,
CheckHdfsIndexTest]
[junit4] 2> NOTE: reproduce with: ant test -Dtestcase=CheckHdfsIndexTest
-Dtests.seed=7C4344571678932E -Dtests.slow=true -Dtests.locale=ga
-Dtests.timezone=Europe/Samara -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[junit4] ERROR 0.00s J0 | CheckHdfsIndexTest (suite) <<<
[junit4] > Throwable #1:
com.carrotsearch.randomizedtesting.ThreadLeakError: 5 threads leaked from SUITE
scope at org.apache.solr.index.hdfs.CheckHdfsIndexTest:
[junit4] > 1) Thread[id=7778, name=SessionTracker,
state=TIMED_WAITING, group=TGRP-CheckHdfsIndexTest]
[junit4] > at java.lang.Object.wait(Native Method)
[junit4] > at
org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:146)
[junit4] > 2) Thread[id=7780, name=ProcessThread(sid:0 cport:58165):,
state=WAITING, group=TGRP-CheckHdfsIndexTest]
[junit4] > at sun.misc.Unsafe.park(Native Method)
[junit4] > at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
[junit4] > at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
[junit4] > at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
[junit4] > at
org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:122)
[junit4] > 3) Thread[id=7777,
name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0, state=RUNNABLE,
group=TGRP-CheckHdfsIndexTest]
[junit4] > at sun.nio.ch.DevPollArrayWrapper.poll0(Native Method)
[junit4] > at
sun.nio.ch.DevPollArrayWrapper.poll(DevPollArrayWrapper.java:223)
[junit4] > at
sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:98)
[junit4] > at
sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
[junit4] > at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
[junit4] > at
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
[junit4] > at java.lang.Thread.run(Thread.java:748)
[junit4] > 4) Thread[id=7779, name=SyncThread:0, state=WAITING,
group=TGRP-CheckHdfsIndexTest]
[junit4] > at sun.misc.Unsafe.park(Native Method)
[junit4] > at
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
[junit4] > at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
[junit4] > at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
[junit4] > at
org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:127)
[junit4] > 5) Thread[id=7776, name=Thread-2981, state=WAITING,
group=TGRP-CheckHdfsIndexTest]
[junit4] > at java.lang.Object.wait(Native Method)
[junit4] > at java.lang.Thread.join(Thread.java:1252)
[junit4] > at java.lang.Thread.join(Thread.java:1326)
[junit4] > at
org.apache.zookeeper.server.NIOServerCnxnFactory.join(NIOServerCnxnFactory.java:297)
[junit4] > at
org.apache.solr.cloud.ZkTestServer$ZKServerMain.runFromConfig(ZkTestServer.java:309)
[junit4] > at
org.apache.solr.cloud.ZkTestServer$2.run(ZkTestServer.java:490)
[junit4] > at
__randomizedtesting.SeedInfo.seed([7C4344571678932E]:0)Throwable #2:
com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie
threads that couldn't be terminated:
[junit4] > 1) Thread[id=7777,
name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0, state=RUNNABLE,
group=TGRP-CheckHdfsIndexTest]
[junit4] > at sun.nio.ch.IOUtil.drain(Native Method)
[junit4] > at
sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:109)
[junit4] > at
sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
[junit4] > at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
[junit4] > at
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
[junit4] > at java.lang.Thread.run(Thread.java:748)
[junit4] > at
__randomizedtesting.SeedInfo.seed([7C4344571678932E]:0)
[junit4] Completed [117/713 (1!)] on J0 in 99.87s, 6 tests, 3 errors, 1
skipped <<< FAILURES!
[...truncated 45452 lines...]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]