Build: https://jenkins.thetaphi.de/job/Lucene-Solr-6.x-Solaris/946/
Java: 64bit/jdk1.8.0 -XX:+UseCompressedOops -XX:+UseG1GC

3 tests failed.
FAILED:  org.apache.solr.index.hdfs.CheckHdfsIndexTest.testChecksumsOnly

Error Message:
java.util.concurrent.TimeoutException: Could not connect to ZooKeeper 
127.0.0.1:58165 within 45000 ms

Stack Trace:
org.apache.solr.common.SolrException: java.util.concurrent.TimeoutException: 
Could not connect to ZooKeeper 127.0.0.1:58165 within 45000 ms
        at 
__randomizedtesting.SeedInfo.seed([7C4344571678932E:DEA68D74F4B76BEE]:0)
        at 
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:183)
        at 
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:117)
        at 
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:107)
        at 
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:86)
        at 
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:80)
        at 
org.apache.solr.cloud.AbstractDistribZkTestBase.distribSetUp(AbstractDistribZkTestBase.java:80)
        at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.distribSetUp(AbstractFullDistribZkTestBase.java:218)
        at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:962)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:916)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:802)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:852)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.util.concurrent.TimeoutException: Could not connect to 
ZooKeeper 127.0.0.1:58165 within 45000 ms
        at 
org.apache.solr.common.cloud.ConnectionManager.waitForConnected(ConnectionManager.java:233)
        at 
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:175)
        ... 38 more


FAILED:  junit.framework.TestSuite.org.apache.solr.index.hdfs.CheckHdfsIndexTest

Error Message:
5 threads leaked from SUITE scope at 
org.apache.solr.index.hdfs.CheckHdfsIndexTest:     1) Thread[id=7778, 
name=SessionTracker, state=TIMED_WAITING, group=TGRP-CheckHdfsIndexTest]        
 at java.lang.Object.wait(Native Method)         at 
org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:146) 
   2) Thread[id=7780, name=ProcessThread(sid:0 cport:58165):, state=WAITING, 
group=TGRP-CheckHdfsIndexTest]         at sun.misc.Unsafe.park(Native Method)   
      at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)      
   at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:122)
    3) Thread[id=7777, name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0, 
state=RUNNABLE, group=TGRP-CheckHdfsIndexTest]         at 
sun.nio.ch.DevPollArrayWrapper.poll0(Native Method)         at 
sun.nio.ch.DevPollArrayWrapper.poll(DevPollArrayWrapper.java:223)         at 
sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:98)         at 
sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)         at 
sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)         at 
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
         at java.lang.Thread.run(Thread.java:748)    4) Thread[id=7779, 
name=SyncThread:0, state=WAITING, group=TGRP-CheckHdfsIndexTest]         at 
sun.misc.Unsafe.park(Native Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:127)
    5) Thread[id=7776, name=Thread-2981, state=WAITING, 
group=TGRP-CheckHdfsIndexTest]         at java.lang.Object.wait(Native Method)  
       at java.lang.Thread.join(Thread.java:1252)         at 
java.lang.Thread.join(Thread.java:1326)         at 
org.apache.zookeeper.server.NIOServerCnxnFactory.join(NIOServerCnxnFactory.java:297)
         at 
org.apache.solr.cloud.ZkTestServer$ZKServerMain.runFromConfig(ZkTestServer.java:309)
         at org.apache.solr.cloud.ZkTestServer$2.run(ZkTestServer.java:490)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: 5 threads leaked from SUITE 
scope at org.apache.solr.index.hdfs.CheckHdfsIndexTest: 
   1) Thread[id=7778, name=SessionTracker, state=TIMED_WAITING, 
group=TGRP-CheckHdfsIndexTest]
        at java.lang.Object.wait(Native Method)
        at 
org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:146)
   2) Thread[id=7780, name=ProcessThread(sid:0 cport:58165):, state=WAITING, 
group=TGRP-CheckHdfsIndexTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:122)
   3) Thread[id=7777, name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0, 
state=RUNNABLE, group=TGRP-CheckHdfsIndexTest]
        at sun.nio.ch.DevPollArrayWrapper.poll0(Native Method)
        at sun.nio.ch.DevPollArrayWrapper.poll(DevPollArrayWrapper.java:223)
        at sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:98)
        at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
        at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
        at 
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
        at java.lang.Thread.run(Thread.java:748)
   4) Thread[id=7779, name=SyncThread:0, state=WAITING, 
group=TGRP-CheckHdfsIndexTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:127)
   5) Thread[id=7776, name=Thread-2981, state=WAITING, 
group=TGRP-CheckHdfsIndexTest]
        at java.lang.Object.wait(Native Method)
        at java.lang.Thread.join(Thread.java:1252)
        at java.lang.Thread.join(Thread.java:1326)
        at 
org.apache.zookeeper.server.NIOServerCnxnFactory.join(NIOServerCnxnFactory.java:297)
        at 
org.apache.solr.cloud.ZkTestServer$ZKServerMain.runFromConfig(ZkTestServer.java:309)
        at org.apache.solr.cloud.ZkTestServer$2.run(ZkTestServer.java:490)
        at __randomizedtesting.SeedInfo.seed([7C4344571678932E]:0)


FAILED:  junit.framework.TestSuite.org.apache.solr.index.hdfs.CheckHdfsIndexTest

Error Message:
There are still zombie threads that couldn't be terminated:    1) 
Thread[id=7777, name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0, state=RUNNABLE, 
group=TGRP-CheckHdfsIndexTest]         at sun.nio.ch.IOUtil.drain(Native 
Method)         at 
sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:109)         
at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)         at 
sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)         at 
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
         at java.lang.Thread.run(Thread.java:748)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie 
threads that couldn't be terminated:
   1) Thread[id=7777, name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0, 
state=RUNNABLE, group=TGRP-CheckHdfsIndexTest]
        at sun.nio.ch.IOUtil.drain(Native Method)
        at sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:109)
        at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
        at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
        at 
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
        at java.lang.Thread.run(Thread.java:748)
        at __randomizedtesting.SeedInfo.seed([7C4344571678932E]:0)




Build Log:
[...truncated 11338 lines...]
   [junit4] Suite: org.apache.solr.index.hdfs.CheckHdfsIndexTest
   [junit4]   2> Creating dataDir: 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/init-core-data-001
   [junit4]   2> 752884 WARN  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] 
o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=1 numCloses=1
   [junit4]   2> 752884 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] 
o.a.s.SolrTestCaseJ4 Using PointFields
   [junit4]   2> 752885 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] 
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via: 
@org.apache.solr.util.RandomizeSSL(reason=, value=NaN, ssl=NaN, clientAuth=NaN)
   [junit4]   2> 752885 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] 
o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 752907 WARN  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] 
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried 
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 752913 WARN  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 752915 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] o.m.log 
jetty-6.1.26
   [junit4]   2> 752925 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] o.m.log 
Extract 
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.2-tests.jar!/webapps/hdfs
 to ./temp/Jetty_serv1.vm3_33624_hdfs____tg9mmx/webapp
   [junit4]   2> 753434 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] o.m.log 
Started HttpServer2$SelectChannelConnectorWithSafeStartup@serv1-vm3:33624
   [junit4]   2> 753547 WARN  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 753548 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] o.m.log 
jetty-6.1.26
   [junit4]   2> 753612 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] o.m.log 
Extract 
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.2-tests.jar!/webapps/datanode
 to ./temp/Jetty_localhost_36641_datanode____.jmw2zt/webapp
   [junit4]   2> 754087 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] o.m.log 
Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:36641
   [junit4]   2> 754162 WARN  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 754177 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] o.m.log 
jetty-6.1.26
   [junit4]   2> 754195 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] o.m.log 
Extract 
jar:file:/export/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.2-tests.jar!/webapps/datanode
 to ./temp/Jetty_localhost_60939_datanode____g07nw/webapp
   [junit4]   2> 754240 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* processReport: from storage 
DS-705520ae-634a-486d-a7b9-524dede0ad8d node 
DatanodeRegistration(127.0.0.1:34809, 
datanodeUuid=2febcede-2f41-4aaa-9139-438e267ad018, infoPort=58165, 
infoSecurePort=0, ipcPort=36779, 
storageInfo=lv=-56;cid=testClusterID;nsid=1990931026;c=0), blocks: 0, 
hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 754240 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* processReport: from storage 
DS-8b47dc82-aa25-425b-a22e-33227a176a14 node 
DatanodeRegistration(127.0.0.1:34809, 
datanodeUuid=2febcede-2f41-4aaa-9139-438e267ad018, infoPort=58165, 
infoSecurePort=0, ipcPort=36779, 
storageInfo=lv=-56;cid=testClusterID;nsid=1990931026;c=0), blocks: 0, 
hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 754716 INFO  
(SUITE-CheckHdfsIndexTest-seed#[7C4344571678932E]-worker) [    ] o.m.log 
Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:60939
   [junit4]   2> 754819 INFO  (IPC Server handler 7 on 62193) [    ] 
BlockStateChange BLOCK* processReport: from storage 
DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5 node 
DatanodeRegistration(127.0.0.1:48010, 
datanodeUuid=142b4244-02b0-4ab4-9bac-7983fe2b175d, infoPort=58917, 
infoSecurePort=0, ipcPort=58493, 
storageInfo=lv=-56;cid=testClusterID;nsid=1990931026;c=0), blocks: 0, 
hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 754819 INFO  (IPC Server handler 7 on 62193) [    ] 
BlockStateChange BLOCK* processReport: from storage 
DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065 node 
DatanodeRegistration(127.0.0.1:48010, 
datanodeUuid=142b4244-02b0-4ab4-9bac-7983fe2b175d, infoPort=58917, 
infoSecurePort=0, ipcPort=58493, 
storageInfo=lv=-56;cid=testClusterID;nsid=1990931026;c=0), blocks: 0, 
hasStaleStorage: false, processing time: 1 msecs
   [junit4]   2> 754909 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnly-seed#[7C4344571678932E]) [    ] 
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 754910 INFO  (Thread-2981) [    ] o.a.s.c.ZkTestServer client 
port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 754910 INFO  (Thread-2981) [    ] o.a.s.c.ZkTestServer 
Starting server
   [junit4]   2> 754913 ERROR (Thread-2981) [    ] o.a.z.s.ZooKeeperServer 
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action 
on ERROR or SHUTDOWN server state changes
   [junit4]   2> 755010 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnly-seed#[7C4344571678932E]) [    ] 
o.a.s.c.ZkTestServer start zk server on port:58165
   [junit4]   2> 800020 WARN  
(TEST-CheckHdfsIndexTest.testChecksumsOnly-seed#[7C4344571678932E]-SendThread(127.0.0.1:58165))
 [    ] o.a.z.ClientCnxn Client session timed out, have not heard from server 
in 45009ms for sessionid 0x0
   [junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=CheckHdfsIndexTest 
-Dtests.method=testChecksumsOnly -Dtests.seed=7C4344571678932E 
-Dtests.slow=true -Dtests.locale=ga -Dtests.timezone=Europe/Samara 
-Dtests.asserts=true -Dtests.file.encoding=UTF-8
   [junit4] ERROR   45.5s J0 | CheckHdfsIndexTest.testChecksumsOnly <<<
   [junit4]    > Throwable #1: org.apache.solr.common.SolrException: 
java.util.concurrent.TimeoutException: Could not connect to ZooKeeper 
127.0.0.1:58165 within 45000 ms
   [junit4]    >        at 
__randomizedtesting.SeedInfo.seed([7C4344571678932E:DEA68D74F4B76BEE]:0)
   [junit4]    >        at 
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:183)
   [junit4]    >        at 
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:117)
   [junit4]    >        at 
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:107)
   [junit4]    >        at 
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:86)
   [junit4]    >        at 
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:80)
   [junit4]    >        at 
org.apache.solr.cloud.AbstractDistribZkTestBase.distribSetUp(AbstractDistribZkTestBase.java:80)
   [junit4]    >        at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.distribSetUp(AbstractFullDistribZkTestBase.java:218)
   [junit4]    >        at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:962)
   [junit4]    >        at java.lang.Thread.run(Thread.java:748)
   [junit4]    > Caused by: java.util.concurrent.TimeoutException: Could not 
connect to ZooKeeper 127.0.0.1:58165 within 45000 ms
   [junit4]    >        at 
org.apache.solr.common.cloud.ConnectionManager.waitForConnected(ConnectionManager.java:233)
   [junit4]    >        at 
org.apache.solr.common.cloud.SolrZkClient.<init>(SolrZkClient.java:175)
   [junit4]    >        ... 38 more
   [junit4]   2> 800425 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 800425 INFO  (Thread-2983) [    ] o.a.s.c.ZkTestServer client 
port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 800425 INFO  (Thread-2983) [    ] o.a.s.c.ZkTestServer 
Starting server
   [junit4]   2> 800427 ERROR (Thread-2983) [    ] o.a.z.s.ZooKeeperServer 
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action 
on ERROR or SHUTDOWN server state changes
   [junit4]   2> 800527 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.ZkTestServer start zk server on port:49657
   [junit4]   2> 800550 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
 to /configs/conf1/solrconfig.xml
   [junit4]   2> 800553 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/schema.xml
 to /configs/conf1/schema.xml
   [junit4]   2> 800556 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
 to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 800559 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
 to /configs/conf1/stopwords.txt
   [junit4]   2> 800564 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/protwords.txt
 to /configs/conf1/protwords.txt
   [junit4]   2> 800567 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/currency.xml
 to /configs/conf1/currency.xml
   [junit4]   2> 800572 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml
 to /configs/conf1/enumsConfig.xml
   [junit4]   2> 800575 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json
 to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 800578 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt
 to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 800581 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt
 to /configs/conf1/old_synonyms.txt
   [junit4]   2> 800583 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/synonyms.txt
 to /configs/conf1/synonyms.txt
   [junit4]   2> 800758 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.SolrTestCaseJ4 Writing core.properties file to 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/control-001/cores/collection1
   [junit4]   2> 800759 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.Server jetty-9.3.14.v20161028
   [junit4]   2> 800761 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@6eff6b56{/,null,AVAILABLE}
   [junit4]   2> 800761 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.AbstractConnector Started 
ServerConnector@4811b32f{HTTP/1.1,[http/1.1]}{127.0.0.1:58678}
   [junit4]   2> 800761 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.Server Started @803144ms
   [junit4]   2> 800761 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.s.e.JettySolrRunner Jetty properties: 
{solr.data.dir=hdfs://serv1-vm3:62193/hdfs__serv1-vm3_62193__export_home_jenkins_workspace_Lucene-Solr-6.x-Solaris_solr_build_solr-core_test_J0_temp_solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001_tempDir-003_control_data,
 hostContext=/, hostPort=58678, 
coreRootDirectory=/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/../../../../../../../../../../export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/control-001/cores}
   [junit4]   2> 800762 ERROR 
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 800762 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 
6.7.0
   [junit4]   2> 800762 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port 
null
   [junit4]   2> 800762 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 800762 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2017-07-03T09:54:15.678Z
   [junit4]   2> 800770 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in 
ZooKeeper)
   [junit4]   2> 800770 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.SolrXmlConfig Loading container configuration from 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/control-001/solr.xml
   [junit4]   2> 800785 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.u.UpdateShardHandler Creating UpdateShardHandler HTTP client with 
params: socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 800786 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:49657/solr
   [junit4]   2> 800840 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) 
[n:127.0.0.1:58678_    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 800841 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) 
[n:127.0.0.1:58678_    ] o.a.s.c.OverseerElectionContext I am going to be the 
leader 127.0.0.1:58678_
   [junit4]   2> 800843 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) 
[n:127.0.0.1:58678_    ] o.a.s.c.Overseer Overseer 
(id=98243422148493316-127.0.0.1:58678_-n_0000000000) starting
   [junit4]   2> 800856 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) 
[n:127.0.0.1:58678_    ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:58678_
   [junit4]   2> 800859 INFO  
(zkCallback-1148-thread-1-processing-n:127.0.0.1:58678_) [n:127.0.0.1:58678_    
] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 800991 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) 
[n:127.0.0.1:58678_    ] o.a.s.c.CorePropertiesLocator Found 1 core definitions 
underneath 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/../../../../../../../../../../export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/control-001/cores
   [junit4]   2> 800991 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) 
[n:127.0.0.1:58678_    ] o.a.s.c.CorePropertiesLocator Cores are: [collection1]
   [junit4]   2> 800995 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 
transient cores
   [junit4]   2> 800999 INFO  
(OverseerStateUpdate-98243422148493316-127.0.0.1:58678_-n_0000000000) 
[n:127.0.0.1:58678_    ] o.a.s.c.o.ReplicaMutator Assigning new node to shard 
shard=shard1
   [junit4]   2> 801019 WARN  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.c.Config 
Beginning with Solr 5.5, <mergePolicy> is deprecated, use <mergePolicyFactory> 
instead.
   [junit4]   2> 801020 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.c.SolrConfig 
Using Lucene MatchVersion: 6.7.0
   [junit4]   2> 801060 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.s.IndexSchema 
[collection1] Schema name=test
   [junit4]   2> 801163 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.s.IndexSchema 
Loaded schema test/1.0 with uniqueid field id
   [junit4]   2> 801191 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.c.CoreContainer 
Creating SolrCore 'collection1' using configuration from collection 
control_collection, trusted=true
   [junit4]   2> 801191 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory 
solr.hdfs.home=hdfs://serv1-vm3:62193/solr_hdfs_home
   [junit4]   2> 801191 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
   [junit4]   2> 801191 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.c.SolrCore 
solr.RecoveryStrategy.Builder
   [junit4]   2> 801191 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.c.SolrCore 
[[collection1] ] Opening new SolrCore at 
[/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/control-001/cores/collection1],
 
dataDir=[hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data/]
   [junit4]   2> 801191 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.JmxMonitoredMap JMX monitoring is enabled. Adding Solr mbeans to JMX 
Server: com.sun.jmx.mbeanserver.JmxMBeanServer@16fda84b
   [junit4]   2> 801192 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data/snapshot_metadata
   [junit4]   2> 801201 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct 
memory allocation set to [true]
   [junit4]   2> 801201 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of 
[8388608] will allocate [1] slabs and use ~[8388608] bytes
   [junit4]   2> 801208 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.s.b.BlockDirectory Block cache on write is disabled
   [junit4]   2> 801210 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data
   [junit4]   2> 801230 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data/index
   [junit4]   2> 801237 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct 
memory allocation set to [true]
   [junit4]   2> 801237 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of 
[8388608] will allocate [1] slabs and use ~[8388608] bytes
   [junit4]   2> 801242 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.s.b.BlockDirectory Block cache on write is disabled
   [junit4]   2> 801243 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: 
maxMergeAtOnce=18, maxMergeAtOnceExplicit=44, maxMergedSegmentMB=13.416015625, 
floorSegmentMB=1.5537109375, forceMergeDeletesPctAllowed=24.94718922912415, 
segmentsPerTier=15.0, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.2954474155087866
   [junit4]   2> 801266 INFO  (IPC Server handler 8 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741825_1001{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 801267 INFO  (IPC Server handler 4 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741825_1001{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED],
 
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
 size 0
   [junit4]   2> 801292 WARN  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = 
requestHandler,name = /dump,class = DumpRequestHandler,attributes = 
{initParams=a, name=/dump, class=DumpRequestHandler},args = 
{defaults={a=A,b=B}}}
   [junit4]   2> 801348 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.u.UpdateHandler 
Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
   [junit4]   2> 801348 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.u.UpdateLog 
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 801348 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.u.HdfsUpdateLog 
Initializing HdfsUpdateLog: tlogDfsReplication=2
   [junit4]   2> 801361 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.u.CommitTracker 
Hard AutoCommit: disabled
   [junit4]   2> 801361 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.u.CommitTracker 
Soft AutoCommit: disabled
   [junit4]   2> 801362 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy: 
minMergeSize=1000, mergeFactor=44, maxMergeSize=9223372036854775807, 
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false, 
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.8665622297044102]
   [junit4]   2> 801375 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.s.SolrIndexSearcher Opening [Searcher@753f972f[collection1] main]
   [junit4]   2> 801378 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 801378 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 801378 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.h.ReplicationHandler Commits will be reserved for  10000
   [junit4]   2> 801382 INFO  
(searcherExecutor-3028-thread-1-processing-n:127.0.0.1:58678_ x:collection1 
c:control_collection) [n:127.0.0.1:58678_ c:control_collection   x:collection1] 
o.a.s.c.SolrCore [collection1] Registered new searcher 
Searcher@753f972f[collection1] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 801385 INFO  
(coreLoadExecutor-3027-thread-1-processing-n:127.0.0.1:58678_) 
[n:127.0.0.1:58678_ c:control_collection   x:collection1] o.a.s.u.UpdateLog 
Could not find max version in index or recent updates, using new clock 
1571894755381477376
   [junit4]   2> 801397 INFO  
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1 
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1 
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext Enough replicas 
found to continue.
   [junit4]   2> 801397 INFO  
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1 
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1 
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext I may be the new 
leader - try and sync
   [junit4]   2> 801398 INFO  
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1 
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1 
r:core_node1 x:collection1] o.a.s.c.SyncStrategy Sync replicas to 
http://127.0.0.1:58678/collection1/
   [junit4]   2> 801398 INFO  
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1 
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1 
r:core_node1 x:collection1] o.a.s.c.SyncStrategy Sync Success - now sync 
replicas to me
   [junit4]   2> 801398 INFO  
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1 
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1 
r:core_node1 x:collection1] o.a.s.c.SyncStrategy 
http://127.0.0.1:58678/collection1/ has no replicas
   [junit4]   2> 801398 INFO  
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1 
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1 
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext Found all 
replicas participating in election, clear LIR
   [junit4]   2> 801406 INFO  
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1 
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1 
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext I am the new 
leader: http://127.0.0.1:58678/collection1/ shard1
   [junit4]   2> 801511 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 801514 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:49657/solr ready
   [junit4]   2> 801514 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection 
loss:false
   [junit4]   2> 801559 INFO  
(coreZkRegister-3020-thread-1-processing-n:127.0.0.1:58678_ x:collection1 
c:control_collection) [n:127.0.0.1:58678_ c:control_collection s:shard1 
r:core_node1 x:collection1] o.a.s.c.ZkController I am the leader, no recovery 
necessary
   [junit4]   2> 801662 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.SolrTestCaseJ4 Writing core.properties file to 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001/cores/collection1
   [junit4]   2> 801663 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001
   [junit4]   2> 801664 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.Server jetty-9.3.14.v20161028
   [junit4]   2> 801675 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@5e7d866{/,null,AVAILABLE}
   [junit4]   2> 801676 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.AbstractConnector Started 
ServerConnector@2127e18a{HTTP/1.1,[http/1.1]}{127.0.0.1:40887}
   [junit4]   2> 801684 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.Server Started @804066ms
   [junit4]   2> 801684 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.s.e.JettySolrRunner Jetty properties: 
{solr.data.dir=hdfs://serv1-vm3:62193/hdfs__serv1-vm3_62193__export_home_jenkins_workspace_Lucene-Solr-6.x-Solaris_solr_build_solr-core_test_J0_temp_solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001_tempDir-003_jetty1,
 solrconfig=solrconfig.xml, hostContext=/, hostPort=40887, 
coreRootDirectory=/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001/cores}
   [junit4]   2> 801684 ERROR 
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 801685 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 
6.7.0
   [junit4]   2> 801685 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port 
null
   [junit4]   2> 801685 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 801685 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2017-07-03T09:54:16.601Z
   [junit4]   2> 801695 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in 
ZooKeeper)
   [junit4]   2> 801695 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.SolrXmlConfig Loading container configuration from 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001/solr.xml
   [junit4]   2> 801702 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.u.UpdateShardHandler Creating UpdateShardHandler HTTP client with 
params: socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 801703 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:49657/solr
   [junit4]   2> 801733 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) 
[n:127.0.0.1:40887_    ] o.a.s.c.c.ZkStateReader Updated live nodes from 
ZooKeeper... (0) -> (1)
   [junit4]   2> 801738 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) 
[n:127.0.0.1:40887_    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 801742 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) 
[n:127.0.0.1:40887_    ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:40887_
   [junit4]   2> 801745 INFO  
(zkCallback-1148-thread-1-processing-n:127.0.0.1:58678_) [n:127.0.0.1:58678_    
] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 801745 INFO  (zkCallback-1152-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 801746 INFO  
(zkCallback-1157-thread-1-processing-n:127.0.0.1:40887_) [n:127.0.0.1:40887_    
] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 801816 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) 
[n:127.0.0.1:40887_    ] o.a.s.c.CorePropertiesLocator Found 1 core definitions 
underneath 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001/cores
   [junit4]   2> 801816 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) 
[n:127.0.0.1:40887_    ] o.a.s.c.CorePropertiesLocator Cores are: [collection1]
   [junit4]   2> 801817 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 
transient cores
   [junit4]   2> 801819 INFO  
(OverseerStateUpdate-98243422148493316-127.0.0.1:58678_-n_0000000000) 
[n:127.0.0.1:58678_    ] o.a.s.c.o.ReplicaMutator Assigning new node to shard 
shard=shard1
   [junit4]   2> 802844 WARN  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.Config Beginning 
with Solr 5.5, <mergePolicy> is deprecated, use <mergePolicyFactory> instead.
   [junit4]   2> 802845 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.SolrConfig Using 
Lucene MatchVersion: 6.7.0
   [junit4]   2> 802858 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.s.IndexSchema 
[collection1] Schema name=test
   [junit4]   2> 802962 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.s.IndexSchema Loaded 
schema test/1.0 with uniqueid field id
   [junit4]   2> 802988 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.CoreContainer 
Creating SolrCore 'collection1' using configuration from collection 
collection1, trusted=true
   [junit4]   2> 802988 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.HdfsDirectoryFactory 
solr.hdfs.home=hdfs://serv1-vm3:62193/solr_hdfs_home
   [junit4]   2> 802988 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.HdfsDirectoryFactory 
Solr Kerberos Authentication disabled
   [junit4]   2> 802988 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.SolrCore 
solr.RecoveryStrategy.Builder
   [junit4]   2> 802988 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.SolrCore 
[[collection1] ] Opening new SolrCore at 
[/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/build/solr-core/test/J0/temp/solr.index.hdfs.CheckHdfsIndexTest_7C4344571678932E-001/shard-1-001/cores/collection1],
 dataDir=[hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data/]
   [junit4]   2> 802989 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.JmxMonitoredMap JMX 
monitoring is enabled. Adding Solr mbeans to JMX Server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@16fda84b
   [junit4]   2> 802989 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.HdfsDirectoryFactory 
creating directory factory for path 
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data/snapshot_metadata
   [junit4]   2> 802999 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.HdfsDirectoryFactory 
Number of slabs of block cache [1] with direct memory allocation set to [true]
   [junit4]   2> 802999 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.HdfsDirectoryFactory 
Block cache target memory usage, slab size of [8388608] will allocate [1] slabs 
and use ~[8388608] bytes
   [junit4]   2> 803003 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.s.b.BlockDirectory 
Block cache on write is disabled
   [junit4]   2> 803004 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.HdfsDirectoryFactory 
creating directory factory for path 
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data
   [junit4]   2> 803024 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.HdfsDirectoryFactory 
creating directory factory for path 
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data/index
   [junit4]   2> 803033 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.HdfsDirectoryFactory 
Number of slabs of block cache [1] with direct memory allocation set to [true]
   [junit4]   2> 803033 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.HdfsDirectoryFactory 
Block cache target memory usage, slab size of [8388608] will allocate [1] slabs 
and use ~[8388608] bytes
   [junit4]   2> 803037 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.s.b.BlockDirectory 
Block cache on write is disabled
   [junit4]   2> 803038 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.u.RandomMergePolicy 
RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: 
[TieredMergePolicy: maxMergeAtOnce=18, maxMergeAtOnceExplicit=44, 
maxMergedSegmentMB=13.416015625, floorSegmentMB=1.5537109375, 
forceMergeDeletesPctAllowed=24.94718922912415, segmentsPerTier=15.0, 
maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.2954474155087866
   [junit4]   2> 803058 INFO  (IPC Server handler 5 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741826_1002{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 803061 INFO  (IPC Server handler 0 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741826_1002{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 803071 WARN  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.c.RequestHandlers 
INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class 
= DumpRequestHandler,attributes = {initParams=a, name=/dump, 
class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
   [junit4]   2> 803127 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.u.UpdateHandler Using 
UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
   [junit4]   2> 803127 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.u.UpdateLog 
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 803127 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.u.HdfsUpdateLog 
Initializing HdfsUpdateLog: tlogDfsReplication=2
   [junit4]   2> 803142 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.u.CommitTracker Hard 
AutoCommit: disabled
   [junit4]   2> 803142 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.u.CommitTracker Soft 
AutoCommit: disabled
   [junit4]   2> 803144 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.u.RandomMergePolicy 
RandomMergePolicy wrapping class org.apache.lucene.index.LogDocMergePolicy: 
[LogDocMergePolicy: minMergeSize=1000, mergeFactor=44, 
maxMergeSize=9223372036854775807, 
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false, 
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.8665622297044102]
   [junit4]   2> 803153 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.s.SolrIndexSearcher 
Opening [Searcher@471dd8eb[collection1] main]
   [junit4]   2> 803155 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 803156 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 803156 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.h.ReplicationHandler 
Commits will be reserved for  10000
   [junit4]   2> 803163 INFO  
(coreLoadExecutor-3038-thread-1-processing-n:127.0.0.1:40887_) 
[n:127.0.0.1:40887_ c:collection1   x:collection1] o.a.s.u.UpdateLog Could not 
find max version in index or recent updates, using new clock 1571894757245845504
   [junit4]   2> 803176 INFO  
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1 
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 
x:collection1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to 
continue.
   [junit4]   2> 803176 INFO  
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1 
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 
x:collection1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try 
and sync
   [junit4]   2> 803176 INFO  
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1 
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 
x:collection1] o.a.s.c.SyncStrategy Sync replicas to 
http://127.0.0.1:40887/collection1/
   [junit4]   2> 803176 INFO  
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1 
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 
x:collection1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 803176 INFO  
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1 
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 
x:collection1] o.a.s.c.SyncStrategy http://127.0.0.1:40887/collection1/ has no 
replicas
   [junit4]   2> 803176 INFO  
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1 
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 
x:collection1] o.a.s.c.ShardLeaderElectionContext Found all replicas 
participating in election, clear LIR
   [junit4]   2> 803180 INFO  
(searcherExecutor-3039-thread-1-processing-n:127.0.0.1:40887_ x:collection1 
c:collection1) [n:127.0.0.1:40887_ c:collection1   x:collection1] 
o.a.s.c.SolrCore [collection1] Registered new searcher 
Searcher@471dd8eb[collection1] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 803181 INFO  
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1 
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 
x:collection1] o.a.s.c.ShardLeaderElectionContext I am the new leader: 
http://127.0.0.1:40887/collection1/ shard1
   [junit4]   2> 803333 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.SolrTestCaseJ4 ###Starting testChecksumsOnlyVerbose
   [junit4]   2> 803336 INFO  
(coreZkRegister-3033-thread-1-processing-n:127.0.0.1:40887_ x:collection1 
c:collection1) [n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 
x:collection1] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 803425 INFO  (IPC Server handler 4 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741827_1003{UCState=COMMITTED, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW]]}
 size 261
   [junit4]   2> 803426 INFO  (IPC Server handler 1 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741827_1003 size 261
   [junit4]   2> 803834 INFO  (IPC Server handler 7 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741828_1004{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW]]}
 size 0
   [junit4]   2> 803834 INFO  (IPC Server handler 3 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741828_1004{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|FINALIZED]]}
 size 0
   [junit4]   2> 803846 INFO  (IPC Server handler 1 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741829_1005{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
 size 0
   [junit4]   2> 803846 INFO  (IPC Server handler 4 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741829_1005{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
 size 0
   [junit4]   2> 803857 INFO  (IPC Server handler 7 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741830_1006{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW]]}
 size 0
   [junit4]   2> 803858 INFO  (IPC Server handler 3 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741830_1006{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW]]}
 size 0
   [junit4]   2> 803868 INFO  (IPC Server handler 1 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741831_1007{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
 size 0
   [junit4]   2> 803869 INFO  (IPC Server handler 4 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741831_1007{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED],
 
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 803880 INFO  (IPC Server handler 7 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741832_1008{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW]]}
 size 0
   [junit4]   2> 803880 INFO  (IPC Server handler 3 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741832_1008{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW]]}
 size 0
   [junit4]   2> 803889 INFO  (IPC Server handler 0 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741833_1009{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 803889 INFO  (IPC Server handler 8 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741833_1009{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 803895 INFO  (IPC Server handler 2 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741834_1010{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 803899 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741834_1010{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED],
 
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|FINALIZED]]}
 size 0
   [junit4]   2> 803907 INFO  (IPC Server handler 9 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741835_1011{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
 size 0
   [junit4]   2> 803907 INFO  (IPC Server handler 9 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741835_1011{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
 size 0
   [junit4]   2> 803928 INFO  (IPC Server handler 1 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741836_1012{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 803928 INFO  (IPC Server handler 4 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741836_1012{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED],
 
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|FINALIZED]]}
 size 0
   [junit4]   2> 803996 INFO  (IPC Server handler 0 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741837_1013{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 803996 INFO  (IPC Server handler 8 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741837_1013{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED],
 
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|FINALIZED]]}
 size 0
   [junit4]   2> 804002 INFO  (IPC Server handler 2 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741838_1014{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|FINALIZED]]}
 size 0
   [junit4]   2> 804003 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741838_1014{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|FINALIZED],
 
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 804010 INFO  (IPC Server handler 9 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741839_1015{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW]]}
 size 0
   [junit4]   2> 804011 INFO  (IPC Server handler 5 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741839_1015 size 5711
   [junit4]   2> 804024 INFO  (IPC Server handler 1 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741840_1016{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 804025 INFO  (IPC Server handler 4 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741840_1016{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 804032 INFO  (IPC Server handler 3 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741841_1017{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW]]}
 size 0
   [junit4]   2> 804032 INFO  (IPC Server handler 9 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741841_1017{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW]]}
 size 0
   [junit4]   2> 804048 INFO  (IPC Server handler 3 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741842_1018{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW]]}
 size 0
   [junit4]   2> 804048 INFO  (IPC Server handler 9 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741842_1018{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-d0ba55ce-26f6-497c-8d24-cb7e5686e065:NORMAL:127.0.0.1:48010|RBW]]}
 size 0
   [junit4]   2> 804055 INFO  (IPC Server handler 1 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:48010 is 
added to blk_1073741843_1019{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-705520ae-634a-486d-a7b9-524dede0ad8d:NORMAL:127.0.0.1:34809|RBW],
 
ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW]]}
 size 0
   [junit4]   2> 804056 INFO  (IPC Server handler 2 on 62193) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:34809 is 
added to blk_1073741843_1019{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-4492cfe7-9e8d-445e-878b-fc9325bd7ee5:NORMAL:127.0.0.1:48010|RBW],
 
ReplicaUC[[DISK]DS-8b47dc82-aa25-425b-a22e-33227a176a14:NORMAL:127.0.0.1:34809|FINALIZED]]}
 size 0
   [junit4]   2> 804517 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.s.h.HdfsDirectory Closing hdfs directory hdfs://serv1-vm3:62193/solr
   [junit4]   2> 804518 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741831_1007 127.0.0.1:34809 
127.0.0.1:48010 
   [junit4]   2> 804518 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741832_1008 127.0.0.1:48010 
127.0.0.1:34809 
   [junit4]   2> 804518 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741833_1009 127.0.0.1:48010 
127.0.0.1:34809 
   [junit4]   2> 804518 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741834_1010 127.0.0.1:48010 
127.0.0.1:34809 
   [junit4]   2> 804518 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741841_1017 127.0.0.1:34809 
127.0.0.1:48010 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741827_1003 127.0.0.1:48010 
127.0.0.1:34809 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741828_1004 127.0.0.1:48010 
127.0.0.1:34809 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741842_1018 127.0.0.1:34809 
127.0.0.1:48010 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741835_1011 127.0.0.1:34809 
127.0.0.1:48010 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741836_1012 127.0.0.1:48010 
127.0.0.1:34809 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741839_1015 127.0.0.1:48010 
127.0.0.1:34809 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741840_1016 127.0.0.1:48010 
127.0.0.1:34809 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741837_1013 127.0.0.1:48010 
127.0.0.1:34809 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741838_1014 127.0.0.1:34809 
127.0.0.1:48010 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741829_1005 127.0.0.1:34809 
127.0.0.1:48010 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741830_1006 127.0.0.1:34809 
127.0.0.1:48010 
   [junit4]   2> 804519 INFO  (IPC Server handler 6 on 62193) [    ] 
BlockStateChange BLOCK* addToInvalidates: blk_1073741843_1019 127.0.0.1:34809 
127.0.0.1:48010 
   [junit4]   2> 804519 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.SolrTestCaseJ4 ###Ending testChecksumsOnlyVerbose
   [junit4]   2> 804519 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.ChaosMonkey monkey: stop shard! 58678
   [junit4]   2> 804519 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.CoreContainer Shutting down CoreContainer instance=1495132002
   [junit4]   2> 804519 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.node
   [junit4]   2> 804519 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.jvm
   [junit4]   2> 804519 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.jetty
   [junit4]   2> 804525 INFO  (coreCloseExecutor-3044-thread-1) 
[n:127.0.0.1:58678_ c:control_collection s:shard1 r:core_node1 x:collection1] 
o.a.s.c.SolrCore [collection1]  CLOSING SolrCore 
org.apache.solr.core.SolrCore@375d0a45
   [junit4]   2> 804548 INFO  
(org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@4cc02650)
 [    ] BlockStateChange BLOCK* BlockManager: ask 127.0.0.1:34809 to delete 
[blk_1073741827_1003, blk_1073741828_1004, blk_1073741829_1005, 
blk_1073741830_1006, blk_1073741831_1007, blk_1073741832_1008, 
blk_1073741833_1009, blk_1073741834_1010, blk_1073741835_1011, 
blk_1073741836_1012, blk_1073741837_1013, blk_1073741838_1014, 
blk_1073741839_1015, blk_1073741840_1016, blk_1073741841_1017, 
blk_1073741842_1018, blk_1073741843_1019]
   [junit4]   2> 804590 INFO  (coreCloseExecutor-3044-thread-1) 
[n:127.0.0.1:58678_ c:control_collection s:shard1 r:core_node1 x:collection1] 
o.a.s.m.SolrMetricManager Closing metric reporters for: solr.core.collection1
   [junit4]   2> 804594 INFO  (coreCloseExecutor-3044-thread-1) 
[n:127.0.0.1:58678_ c:control_collection s:shard1 r:core_node1 x:collection1] 
o.a.s.s.h.HdfsDirectory Closing hdfs directory 
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data/snapshot_metadata
   [junit4]   2> 804595 INFO  (coreCloseExecutor-3044-thread-1) 
[n:127.0.0.1:58678_ c:control_collection s:shard1 r:core_node1 x:collection1] 
o.a.s.s.h.HdfsDirectory Closing hdfs directory 
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data
   [junit4]   2> 804596 INFO  (coreCloseExecutor-3044-thread-1) 
[n:127.0.0.1:58678_ c:control_collection s:shard1 r:core_node1 x:collection1] 
o.a.s.s.h.HdfsDirectory Closing hdfs directory 
hdfs://serv1-vm3:62193/solr_hdfs_home/control_collection/core_node1/data/index
   [junit4]   2> 804596 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.Overseer Overseer 
(id=98243422148493316-127.0.0.1:58678_-n_0000000000) closing
   [junit4]   2> 804597 INFO  
(OverseerStateUpdate-98243422148493316-127.0.0.1:58678_-n_0000000000) 
[n:127.0.0.1:58678_    ] o.a.s.c.Overseer Overseer Loop exiting : 
127.0.0.1:58678_
   [junit4]   2> 804601 WARN  
(zkCallback-1148-thread-2-processing-n:127.0.0.1:58678_) [n:127.0.0.1:58678_    
] o.a.s.c.c.ZkStateReader ZooKeeper watch triggered, but Solr cannot talk to 
ZK: [KeeperErrorCode = Session expired for /live_nodes]
   [junit4]   2> 804601 INFO  (zkCallback-1152-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (1)
   [junit4]   2> 804601 INFO  
(zkCallback-1157-thread-2-processing-n:127.0.0.1:40887_) [n:127.0.0.1:40887_    
] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (1)
   [junit4]   2> 804601 INFO  
(zkCallback-1157-thread-1-processing-n:127.0.0.1:40887_) [n:127.0.0.1:40887_    
] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:40887_
   [junit4]   2> 804602 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.AbstractConnector Stopped 
ServerConnector@4811b32f{HTTP/1.1,[http/1.1]}{127.0.0.1:0}
   [junit4]   2> 804602 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.h.ContextHandler Stopped 
o.e.j.s.ServletContextHandler@6eff6b56{/,null,UNAVAILABLE}
   [junit4]   2> 804603 INFO  
(zkCallback-1157-thread-1-processing-n:127.0.0.1:40887_) [n:127.0.0.1:40887_    
] o.a.s.c.Overseer Overseer 
(id=98243422148493320-127.0.0.1:40887_-n_0000000001) starting
   [junit4]   2> 804606 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.ChaosMonkey monkey: stop shard! 40887
   [junit4]   2> 804606 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.CoreContainer Shutting down CoreContainer instance=1504621779
   [junit4]   2> 804607 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.node
   [junit4]   2> 804607 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.jvm
   [junit4]   2> 804607 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.m.SolrMetricManager Closing metric reporters for: solr.jetty
   [junit4]   2> 804611 INFO  (coreCloseExecutor-3046-thread-1) 
[n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 x:collection1] 
o.a.s.c.SolrCore [collection1]  CLOSING SolrCore 
org.apache.solr.core.SolrCore@22e16c08
   [junit4]   2> 804664 INFO  (coreCloseExecutor-3046-thread-1) 
[n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 x:collection1] 
o.a.s.m.SolrMetricManager Closing metric reporters for: solr.core.collection1
   [junit4]   2> 804664 INFO  (coreCloseExecutor-3046-thread-1) 
[n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 x:collection1] 
o.a.s.s.h.HdfsDirectory Closing hdfs directory 
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data
   [junit4]   2> 804665 INFO  (coreCloseExecutor-3046-thread-1) 
[n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 x:collection1] 
o.a.s.s.h.HdfsDirectory Closing hdfs directory 
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data/index
   [junit4]   2> 804666 INFO  (coreCloseExecutor-3046-thread-1) 
[n:127.0.0.1:40887_ c:collection1 s:shard1 r:core_node1 x:collection1] 
o.a.s.s.h.HdfsDirectory Closing hdfs directory 
hdfs://serv1-vm3:62193/solr_hdfs_home/collection1/core_node1/data/snapshot_metadata
   [junit4]   2> 804666 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.Overseer Overseer 
(id=98243422148493320-127.0.0.1:40887_-n_0000000001) closing
   [junit4]   2> 804666 INFO  
(OverseerStateUpdate-98243422148493320-127.0.0.1:40887_-n_0000000001) 
[n:127.0.0.1:40887_    ] o.a.s.c.Overseer Overseer Loop exiting : 
127.0.0.1:40887_
   [junit4]   2> 804669 INFO  (zkCallback-1152-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (0)
   [junit4]   2> 804671 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.AbstractConnector Stopped 
ServerConnector@2127e18a{HTTP/1.1,[http/1.1]}{127.0.0.1:0}
   [junit4]   2> 804671 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.e.j.s.h.ContextHandler Stopped 
o.e.j.s.ServletContextHandler@5e7d866{/,null,UNAVAILABLE}
   [junit4]   2> 804676 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.ChaosMonkey monkey: stop shard! 58678
   [junit4]   2> 804676 ERROR 
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper 
server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 804677 INFO  
(TEST-CheckHdfsIndexTest.testChecksumsOnlyVerbose-seed#[7C4344571678932E]) [    
] o.a.s.c.ZkTestServer connecting to 127.0.0.1:49657 49657
   [junit4]   2> 804744 INFO  (Thread-2983) [    ] o.a.s.c.ZkTestServer 
connecting to 127.0.0.1:49657 49657
   [junit4]   2> 804746 WARN  (Thread-2983) [    ] o.a.s.c.ZkTestServer Watch 
limit violations: 
   [junit4]   2> Maximum concurrent create/delete watches above limit:
   [junit4]   2> 
   [junit4]   2>        3       /solr/aliases.json
   [junit4]   2>        3       /solr/clusterprops.json
   [junit4]   2>        2       /solr/security.json
   [junit4]   2>        2       /solr/configs/conf1
   [junit4]   2> 
   [junit4]   2> Maximum concurrent data watches above limit:
   [junit4]   2> 
   [junit4]   2>        3       /solr/clusterstate.json
   [junit4]   2> 
   [junit4]   2> Maximum concurrent children watches above limit:
   [junit4]   2> 
   [junit4]   2>        3       /solr/live_nodes
   [junit4]   2>        3       /solr/collections
   [junit4]   2>        2       /solr/overseer/queue
   [junit4]   2>        2       /solr/overseer/collection-queue-work
   [junit4]   2> 
   [junit4]   2> 804781 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 804782 INFO  (Thread-3124) [    ] o.a.s.c.ZkTestServer client 
port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 804782 INFO  (Thread-3124) [    ] o.a.s.c.ZkTestServer 
Starting server
   [junit4]   2> 804783 ERROR (Thread-3124) [    ] o.a.z.s.ZooKeeperServer 
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action 
on ERROR or SHUTDOWN server state changes
   [junit4]   2> 804784 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.ZkTestServer start zk server on port:62052
   [junit4]   2> 804796 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
 to /configs/conf1/solrconfig.xml
   [junit4]   2> 804797 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/schema.xml
 to /configs/conf1/schema.xml
   [junit4]   2> 804799 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
 to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 804801 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
 to /configs/conf1/stopwords.txt
   [junit4]   2> 804803 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/protwords.txt
 to /configs/conf1/protwords.txt
   [junit4]   2> 804806 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/currency.xml
 to /configs/conf1/currency.xml
   [junit4]   2> 804807 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml
 to /configs/conf1/enumsConfig.xml
   [junit4]   2> 804809 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json
 to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 804810 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt
 to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 804812 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/export/home/jenkins/workspace/Lucene-Solr-6.x-Solaris/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt
 to /configs/conf1/old_synonyms.txt
   [junit4]   2> 804813 INFO  
(TEST-CheckHdfsIndexTest.testBogusTermVectors-seed#[7C4344571678932E]) 

[...truncated too long message...]

edSynchronizer.java:2048)
   [junit4]   2>        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
   [junit4]   2>        at 
org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:122)
   [junit4]   2> 849734 ERROR (ProcessThread(sid:0 cport:58165):) [    ] 
o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper 
server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 849734 ERROR (SessionTracker) [    ] 
o.a.z.s.ZooKeeperCriticalThread Severe unrecoverable error, from thread : 
SessionTracker
   [junit4]   2> java.lang.InterruptedException
   [junit4]   2>        at java.lang.Object.wait(Native Method)
   [junit4]   2>        at 
org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:146)
   [junit4]   2> 849736 ERROR (SessionTracker) [    ] o.a.z.s.ZooKeeperServer 
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action 
on ERROR or SHUTDOWN server state changes
   [junit4]   2> 849735 WARN  (Thread-2981) [    ] o.a.s.c.ZkTestServer Server 
interrupted
   [junit4]   2> java.lang.InterruptedException
   [junit4]   2>        at java.lang.Object.wait(Native Method)
   [junit4]   2>        at java.lang.Thread.join(Thread.java:1252)
   [junit4]   2>        at java.lang.Thread.join(Thread.java:1326)
   [junit4]   2>        at 
org.apache.zookeeper.server.NIOServerCnxnFactory.join(NIOServerCnxnFactory.java:297)
   [junit4]   2>        at 
org.apache.solr.cloud.ZkTestServer$ZKServerMain.runFromConfig(ZkTestServer.java:309)
   [junit4]   2>        at 
org.apache.solr.cloud.ZkTestServer$2.run(ZkTestServer.java:490)
   [junit4]   2> 849735 ERROR (SyncThread:0) [    ] 
o.a.z.s.ZooKeeperCriticalThread Severe unrecoverable error, from thread : 
SyncThread:0
   [junit4]   2> java.lang.InterruptedException
   [junit4]   2>        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.java:2014)
   [junit4]   2>        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2048)
   [junit4]   2>        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
   [junit4]   2>        at 
org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:127)
   [junit4]   2> 849738 ERROR (SyncThread:0) [    ] o.a.z.s.ZooKeeperServer 
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action 
on ERROR or SHUTDOWN server state changes
   [junit4]   2> Jul 03, 2017 9:55:07 AM 
com.carrotsearch.randomizedtesting.ThreadLeakControl tryToInterruptAll
   [junit4]   2> SEVERE: There are still zombie threads that couldn't be 
terminated:
   [junit4]   2>    1) Thread[id=7777, 
name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0, state=RUNNABLE, 
group=TGRP-CheckHdfsIndexTest]
   [junit4]   2>         at sun.nio.ch.IOUtil.drain(Native Method)
   [junit4]   2>         at 
sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:109)
   [junit4]   2>         at 
sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
   [junit4]   2>         at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
   [junit4]   2>         at 
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
   [junit4]   2>         at java.lang.Thread.run(Thread.java:748)
   [junit4]   2> NOTE: test params are: codec=Lucene62, 
sim=RandomSimilarity(queryNorm=true,coord=yes): {field=LM 
Jelinek-Mercer(0.700000), foo=ClassicSimilarity, titleTokenized=IB SPL-D2, 
body=DFR I(F)L1}, locale=ga, timezone=Europe/Samara
   [junit4]   2> NOTE: SunOS 5.11 amd64/Oracle Corporation 1.8.0_131 
(64-bit)/cpus=3,threads=3,free=198249632,total=536870912
   [junit4]   2> NOTE: All tests run in this JVM: [DirectoryFactoryTest, 
TestSolrConfigHandlerConcurrent, CloudExitableDirectoryReaderTest, 
OverriddenZkACLAndCredentialsProvidersTest, HdfsUnloadDistributedZkTest, 
SolrSlf4jReporterTest, TestTrackingShardHandlerFactory, TestConfigReload, 
LoggingHandlerTest, CdcrReplicationDistributedZkTest, TestManagedSchema, 
OverseerModifyCollectionTest, GraphQueryTest, TestLazyCores, 
HdfsRestartWhileUpdatingTest, AtomicUpdatesTest, TestDFRSimilarityFactory, 
TestManagedSynonymFilterFactory, TestFieldTypeResource, SolrCloudExampleTest, 
TestImplicitCoreProperties, TestTestInjection, PrimitiveFieldTypeTest, 
DocValuesNotIndexedTest, TestRuleBasedAuthorizationPlugin, RulesTest, 
TestRecoveryHdfs, MigrateRouteKeyTest, TestAtomicUpdateErrorCases, 
TestRandomRequestDistribution, TestTrieFacet, SolrInfoMBeanTest, 
TestMaxTokenLenTokenizer, FieldAnalysisRequestHandlerTest, 
AnalyticsMergeStrategyTest, RollingRestartTest, TemplateUpdateProcessorTest, 
CloneFieldUpdateProcessorFactoryTest, TestStressVersions, 
TestReplicationHandlerBackup, SuggestComponentContextFilterQueryTest, 
TestDefaultStatsCache, TestPseudoReturnFields, UpdateParamsTest, 
SpellCheckCollatorWithCollapseTest, TestDistributedStatsComponentCardinality, 
DirectUpdateHandlerTest, TestOrdValues, DeleteLastCustomShardedReplicaTest, 
LeaderInitiatedRecoveryOnCommitTest, RequestLoggingTest, 
TestLeaderElectionWithEmptyReplica, HdfsBasicDistributedZk2Test, 
DistanceFunctionTest, DocumentBuilderTest, DistributedFacetPivotSmallTest, 
DistributedSpellCheckComponentTest, TestCursorMarkWithoutUniqueKey, 
TestQuerySenderNoQuery, AddSchemaFieldsUpdateProcessorFactoryTest, 
SparseHLLTest, SortByFunctionTest, LeaderFailoverAfterPartitionTest, 
SuggesterWFSTTest, TestDistributedSearch, TestQuerySenderListener, 
TestFaceting, TestRawResponseWriter, TestFreeTextSuggestions, 
CheckHdfsIndexTest]
   [junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=CheckHdfsIndexTest 
-Dtests.seed=7C4344571678932E -Dtests.slow=true -Dtests.locale=ga 
-Dtests.timezone=Europe/Samara -Dtests.asserts=true -Dtests.file.encoding=UTF-8
   [junit4] ERROR   0.00s J0 | CheckHdfsIndexTest (suite) <<<
   [junit4]    > Throwable #1: 
com.carrotsearch.randomizedtesting.ThreadLeakError: 5 threads leaked from SUITE 
scope at org.apache.solr.index.hdfs.CheckHdfsIndexTest: 
   [junit4]    >    1) Thread[id=7778, name=SessionTracker, 
state=TIMED_WAITING, group=TGRP-CheckHdfsIndexTest]
   [junit4]    >         at java.lang.Object.wait(Native Method)
   [junit4]    >         at 
org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:146)
   [junit4]    >    2) Thread[id=7780, name=ProcessThread(sid:0 cport:58165):, 
state=WAITING, group=TGRP-CheckHdfsIndexTest]
   [junit4]    >         at sun.misc.Unsafe.park(Native Method)
   [junit4]    >         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
   [junit4]    >         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
   [junit4]    >         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
   [junit4]    >         at 
org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:122)
   [junit4]    >    3) Thread[id=7777, 
name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0, state=RUNNABLE, 
group=TGRP-CheckHdfsIndexTest]
   [junit4]    >         at sun.nio.ch.DevPollArrayWrapper.poll0(Native Method)
   [junit4]    >         at 
sun.nio.ch.DevPollArrayWrapper.poll(DevPollArrayWrapper.java:223)
   [junit4]    >         at 
sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:98)
   [junit4]    >         at 
sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
   [junit4]    >         at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
   [junit4]    >         at 
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
   [junit4]    >         at java.lang.Thread.run(Thread.java:748)
   [junit4]    >    4) Thread[id=7779, name=SyncThread:0, state=WAITING, 
group=TGRP-CheckHdfsIndexTest]
   [junit4]    >         at sun.misc.Unsafe.park(Native Method)
   [junit4]    >         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
   [junit4]    >         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
   [junit4]    >         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
   [junit4]    >         at 
org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:127)
   [junit4]    >    5) Thread[id=7776, name=Thread-2981, state=WAITING, 
group=TGRP-CheckHdfsIndexTest]
   [junit4]    >         at java.lang.Object.wait(Native Method)
   [junit4]    >         at java.lang.Thread.join(Thread.java:1252)
   [junit4]    >         at java.lang.Thread.join(Thread.java:1326)
   [junit4]    >         at 
org.apache.zookeeper.server.NIOServerCnxnFactory.join(NIOServerCnxnFactory.java:297)
   [junit4]    >         at 
org.apache.solr.cloud.ZkTestServer$ZKServerMain.runFromConfig(ZkTestServer.java:309)
   [junit4]    >         at 
org.apache.solr.cloud.ZkTestServer$2.run(ZkTestServer.java:490)
   [junit4]    >        at 
__randomizedtesting.SeedInfo.seed([7C4344571678932E]:0)Throwable #2: 
com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie 
threads that couldn't be terminated:
   [junit4]    >    1) Thread[id=7777, 
name=NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0, state=RUNNABLE, 
group=TGRP-CheckHdfsIndexTest]
   [junit4]    >         at sun.nio.ch.IOUtil.drain(Native Method)
   [junit4]    >         at 
sun.nio.ch.DevPollSelectorImpl.doSelect(DevPollSelectorImpl.java:109)
   [junit4]    >         at 
sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
   [junit4]    >         at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
   [junit4]    >         at 
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:173)
   [junit4]    >         at java.lang.Thread.run(Thread.java:748)
   [junit4]    >        at 
__randomizedtesting.SeedInfo.seed([7C4344571678932E]:0)
   [junit4] Completed [117/713 (1!)] on J0 in 99.87s, 6 tests, 3 errors, 1 
skipped <<< FAILURES!

[...truncated 45452 lines...]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to