Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-6.x/375/

1 tests failed.
FAILED:  org.apache.solr.cloud.hdfs.StressHdfsTest.test

Error Message:
Timeout occured while waiting response from server at: 
http://127.0.0.1:55311/nst/fw

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: Timeout occured while waiting 
response from server at: http://127.0.0.1:55311/nst/fw
        at 
__randomizedtesting.SeedInfo.seed([2C5F898E344E9D7D:A40BB6549AB2F085]:0)
        at 
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:605)
        at 
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:261)
        at 
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:250)
        at 
org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:477)
        at 
org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:407)
        at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1383)
        at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1134)
        at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:1073)
        at org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1219)
        at 
org.apache.solr.cloud.hdfs.StressHdfsTest.createAndDeleteCollection(StressHdfsTest.java:220)
        at 
org.apache.solr.cloud.hdfs.StressHdfsTest.test(StressHdfsTest.java:103)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1713)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:907)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:943)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:957)
        at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:992)
        at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:967)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:916)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:802)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:852)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.net.SocketTimeoutException: Read timed out
        at java.net.SocketInputStream.socketRead0(Native Method)
        at java.net.SocketInputStream.socketRead(SocketInputStream.java:116)
        at java.net.SocketInputStream.read(SocketInputStream.java:171)
        at java.net.SocketInputStream.read(SocketInputStream.java:141)
        at 
org.apache.http.impl.io.AbstractSessionInputBuffer.fillBuffer(AbstractSessionInputBuffer.java:160)
        at 
org.apache.http.impl.io.SocketInputBuffer.fillBuffer(SocketInputBuffer.java:84)
        at 
org.apache.http.impl.io.AbstractSessionInputBuffer.readLine(AbstractSessionInputBuffer.java:273)
        at 
org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:140)
        at 
org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:57)
        at 
org.apache.http.impl.io.AbstractMessageParser.parse(AbstractMessageParser.java:261)
        at 
org.apache.http.impl.AbstractHttpClientConnection.receiveResponseHeader(AbstractHttpClientConnection.java:283)
        at 
org.apache.http.impl.conn.DefaultClientConnection.receiveResponseHeader(DefaultClientConnection.java:251)
        at 
org.apache.http.impl.conn.ManagedClientConnectionImpl.receiveResponseHeader(ManagedClientConnectionImpl.java:197)
        at 
org.apache.http.protocol.HttpRequestExecutor.doReceiveResponse(HttpRequestExecutor.java:272)
        at 
org.apache.http.protocol.HttpRequestExecutor.execute(HttpRequestExecutor.java:124)
        at 
org.apache.http.impl.client.DefaultRequestDirector.tryExecute(DefaultRequestDirector.java:685)
        at 
org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:487)
        at 
org.apache.http.impl.client.AbstractHttpClient.doExecute(AbstractHttpClient.java:882)
        at 
org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82)
        at 
org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:55)
        at 
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:497)
        ... 51 more




Build Log:
[...truncated 13317 lines...]
   [junit4] Suite: org.apache.solr.cloud.hdfs.StressHdfsTest
   [junit4]   2> Creating dataDir: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/init-core-data-001
   [junit4]   2> 3197365 WARN  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] 
o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=41 numCloses=41
   [junit4]   2> 3197365 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] 
o.a.s.SolrTestCaseJ4 Using TrieFields
   [junit4]   2> 3197367 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] 
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via: 
@org.apache.solr.SolrTestCaseJ4$SuppressSSL(bugUrl=https://issues.apache.org/jira/browse/SOLR-5776)
   [junit4]   2> 3197367 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] 
o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /nst/fw
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 3197456 WARN  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] 
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried 
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 3197463 WARN  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 3197465 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] o.m.log 
jetty-6.1.26
   [junit4]   2> 3197475 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.2-tests.jar!/webapps/hdfs
 to ./temp/Jetty_localhost_45016_hdfs____eorwm6/webapp
   [junit4]   2> 3197825 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:45016
   [junit4]   2> 3197963 WARN  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 3197964 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] o.m.log 
jetty-6.1.26
   [junit4]   2> 3197977 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.2-tests.jar!/webapps/datanode
 to ./temp/Jetty_localhost_55427_datanode____.sopkg0/webapp
   [junit4]   2> 3198304 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:55427
   [junit4]   2> 3198362 WARN  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 3198371 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] o.m.log 
jetty-6.1.26
   [junit4]   2> 3198382 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.2-tests.jar!/webapps/datanode
 to ./temp/Jetty_localhost_43014_datanode____v2izx7/webapp
   [junit4]   2> 3198506 INFO  (IPC Server handler 4 on 36966) [    ] 
BlockStateChange BLOCK* processReport: from storage 
DS-96f0454d-71a1-441a-bb20-648eff417ff3 node 
DatanodeRegistration(127.0.0.1:33736, 
datanodeUuid=ac1db4d3-3667-4d48-bdee-b3b43daf786f, infoPort=34285, 
infoSecurePort=0, ipcPort=50116, 
storageInfo=lv=-56;cid=testClusterID;nsid=1185290609;c=0), blocks: 0, 
hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 3198506 INFO  (IPC Server handler 4 on 36966) [    ] 
BlockStateChange BLOCK* processReport: from storage 
DS-35bfeff3-61da-4bcf-b500-98d455d027e8 node 
DatanodeRegistration(127.0.0.1:33736, 
datanodeUuid=ac1db4d3-3667-4d48-bdee-b3b43daf786f, infoPort=34285, 
infoSecurePort=0, ipcPort=50116, 
storageInfo=lv=-56;cid=testClusterID;nsid=1185290609;c=0), blocks: 0, 
hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 3198765 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:43014
   [junit4]   2> 3198894 INFO  (IPC Server handler 3 on 36966) [    ] 
BlockStateChange BLOCK* processReport: from storage 
DS-43345fcd-5e09-4aa7-925b-a6e9805c0eaf node 
DatanodeRegistration(127.0.0.1:32931, 
datanodeUuid=e2ca1cf6-caa1-493a-ac43-2327149a533d, infoPort=37147, 
infoSecurePort=0, ipcPort=58714, 
storageInfo=lv=-56;cid=testClusterID;nsid=1185290609;c=0), blocks: 0, 
hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 3198894 INFO  (IPC Server handler 3 on 36966) [    ] 
BlockStateChange BLOCK* processReport: from storage 
DS-5284bd1e-8b3d-4e0a-9186-4be1c3da4cd2 node 
DatanodeRegistration(127.0.0.1:32931, 
datanodeUuid=e2ca1cf6-caa1-493a-ac43-2327149a533d, infoPort=37147, 
infoSecurePort=0, ipcPort=58714, 
storageInfo=lv=-56;cid=testClusterID;nsid=1185290609;c=0), blocks: 0, 
hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 3198936 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.c.ZkTestServer 
STARTING ZK TEST SERVER
   [junit4]   2> 3198936 INFO  (Thread-11045) [    ] o.a.s.c.ZkTestServer 
client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 3198936 INFO  (Thread-11045) [    ] o.a.s.c.ZkTestServer 
Starting server
   [junit4]   2> 3198938 ERROR (Thread-11045) [    ] o.a.z.s.ZooKeeperServer 
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action 
on ERROR or SHUTDOWN server state changes
   [junit4]   2> 3199036 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.c.ZkTestServer 
start zk server on port:58917
   [junit4]   2> 3199048 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
 to /configs/conf1/solrconfig.xml
   [junit4]   2> 3199049 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/core/src/test-files/solr/collection1/conf/schema.xml
 to /configs/conf1/schema.xml
   [junit4]   2> 3199051 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
 to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 3199052 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
 to /configs/conf1/stopwords.txt
   [junit4]   2> 3199054 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/core/src/test-files/solr/collection1/conf/protwords.txt
 to /configs/conf1/protwords.txt
   [junit4]   2> 3199055 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/core/src/test-files/solr/collection1/conf/currency.xml
 to /configs/conf1/currency.xml
   [junit4]   2> 3199056 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml
 to /configs/conf1/enumsConfig.xml
   [junit4]   2> 3199058 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json
 to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 3199059 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt
 to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 3199060 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt
 to /configs/conf1/old_synonyms.txt
   [junit4]   2> 3199062 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractZkTestCase put 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/core/src/test-files/solr/collection1/conf/synonyms.txt
 to /configs/conf1/synonyms.txt
   [junit4]   2> 3199156 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.SolrTestCaseJ4 
Writing core.properties file to 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/control-001/cores/collection1
   [junit4]   2> 3199157 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.e.j.s.Server 
jetty-9.3.14.v20161028
   [junit4]   2> 3199159 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@39d78eea{/nst/fw,null,AVAILABLE}
   [junit4]   2> 3199159 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@550e3866{HTTP/1.1,[http/1.1]}{127.0.0.1:55311}
   [junit4]   2> 3199159 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.e.j.s.Server 
Started @3201243ms
   [junit4]   2> 3199159 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: 
{solr.data.dir=hdfs://localhost:36966/hdfs__localhost_36966__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-6.x_checkout_solr_build_solr-core_test_J0_temp_solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001_tempDir-002_control_data,
 hostContext=/nst/fw, hostPort=55311, 
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/control-001/cores}
   [junit4]   2> 3199159 ERROR 
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 3199159 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 
6.7.0
   [junit4]   2> 3199159 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 3199159 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 3199160 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2017-06-19T11:52:59.227Z
   [junit4]   2> 3199164 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in 
ZooKeeper)
   [junit4]   2> 3199164 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.c.SolrXmlConfig 
Loading container configuration from 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/control-001/solr.xml
   [junit4]   2> 3199169 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.u.UpdateShardHandler Creating UpdateShardHandler HTTP client with params: 
socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 3199171 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.c.ZkContainer 
Zookeeper client=127.0.0.1:58917/solr
   [junit4]   2> 3199191 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:55311_nst%2Ffw  
  ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 3199192 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:55311_nst%2Ffw  
  ] o.a.s.c.OverseerElectionContext I am going to be the leader 
127.0.0.1:55311_nst%2Ffw
   [junit4]   2> 3199193 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:55311_nst%2Ffw  
  ] o.a.s.c.Overseer Overseer 
(id=98164616659206148-127.0.0.1:55311_nst%2Ffw-n_0000000000) starting
   [junit4]   2> 3199202 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:55311_nst%2Ffw  
  ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:55311_nst%2Ffw
   [junit4]   2> 3199204 INFO  
(zkCallback-31827-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw    ] o.a.s.c.c.ZkStateReader Updated live nodes 
from ZooKeeper... (0) -> (1)
   [junit4]   2> 3199374 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:55311_nst%2Ffw  
  ] o.a.s.c.CorePropertiesLocator Found 1 core definitions underneath 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/control-001/cores
   [junit4]   2> 3199374 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:55311_nst%2Ffw  
  ] o.a.s.c.CorePropertiesLocator Cores are: [collection1]
   [junit4]   2> 3199375 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 
transient cores
   [junit4]   2> 3199387 INFO  
(OverseerStateUpdate-98164616659206148-127.0.0.1:55311_nst%2Ffw-n_0000000000) 
[n:127.0.0.1:55311_nst%2Ffw    ] o.a.s.c.o.ReplicaMutator Assigning new node to 
shard shard=shard1
   [junit4]   2> 3200397 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.SolrConfig Using Lucene MatchVersion: 6.7.0
   [junit4]   2> 3200410 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.s.IndexSchema [collection1] Schema name=test
   [junit4]   2> 3200509 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.s.IndexSchema Loaded schema test/1.0 with uniqueid field id
   [junit4]   2> 3200517 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.CoreContainer Creating SolrCore 'collection1' using configuration from 
collection control_collection, trusted=true
   [junit4]   2> 3200517 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory 
solr.hdfs.home=hdfs://localhost:36966/solr_hdfs_home
   [junit4]   2> 3200517 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
   [junit4]   2> 3200517 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 3200517 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.SolrCore [[collection1] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/control-001/cores/collection1],
 
dataDir=[hdfs://localhost:36966/solr_hdfs_home/control_collection/core_node1/data/]
   [junit4]   2> 3200518 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.JmxMonitoredMap JMX monitoring is enabled. Adding Solr mbeans to JMX 
Server: com.sun.jmx.mbeanserver.JmxMBeanServer@78d9b800
   [junit4]   2> 3200518 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/control_collection/core_node1/data/snapshot_metadata
   [junit4]   2> 3200524 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct 
memory allocation set to [true]
   [junit4]   2> 3200524 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of 
[8388608] will allocate [1] slabs and use ~[8388608] bytes
   [junit4]   2> 3200524 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache
   [junit4]   2> 3200531 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.s.b.BlockDirectory Block cache on write is disabled
   [junit4]   2> 3200533 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/control_collection/core_node1/data
   [junit4]   2> 3200545 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/control_collection/core_node1/data/index
   [junit4]   2> 3200550 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct 
memory allocation set to [true]
   [junit4]   2> 3200550 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of 
[8388608] will allocate [1] slabs and use ~[8388608] bytes
   [junit4]   2> 3200550 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache
   [junit4]   2> 3200556 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.s.b.BlockDirectory Block cache on write is disabled
   [junit4]   2> 3200557 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy: 
minMergeSize=1000, mergeFactor=40, maxMergeSize=9223372036854775807, 
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false, 
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=1.0]
   [junit4]   2> 3200567 INFO  (IPC Server handler 4 on 36966) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33736 is 
added to blk_1073741825_1001{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-43345fcd-5e09-4aa7-925b-a6e9805c0eaf:NORMAL:127.0.0.1:32931|RBW],
 
ReplicaUC[[DISK]DS-96f0454d-71a1-441a-bb20-648eff417ff3:NORMAL:127.0.0.1:33736|RBW]]}
 size 0
   [junit4]   2> 3200568 INFO  (IPC Server handler 6 on 36966) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:32931 is 
added to blk_1073741825_1001{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-96f0454d-71a1-441a-bb20-648eff417ff3:NORMAL:127.0.0.1:33736|RBW],
 
ReplicaUC[[DISK]DS-5284bd1e-8b3d-4e0a-9186-4be1c3da4cd2:NORMAL:127.0.0.1:32931|FINALIZED]]}
 size 0
   [junit4]   2> 3200573 WARN  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = 
requestHandler,name = /dump,class = DumpRequestHandler,attributes = 
{initParams=a, name=/dump, class=DumpRequestHandler},args = 
{defaults={a=A,b=B}}}
   [junit4]   2> 3200600 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.u.UpdateHandler Using UpdateLog implementation: 
org.apache.solr.update.HdfsUpdateLog
   [junit4]   2> 3200600 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH 
numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 3200600 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=2
   [junit4]   2> 3200608 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 3200608 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 3200609 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.AlcoholicMergePolicy: [AlcoholicMergePolicy: 
minMergeSize=0, mergeFactor=10, maxMergeSize=1961424476, 
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, 
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.1]
   [junit4]   2> 3200612 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.s.SolrIndexSearcher Opening [Searcher@515d848f[collection1] main]
   [junit4]   2> 3200613 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 3200613 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 3200613 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.h.ReplicationHandler Commits will be reserved for  10000
   [junit4]   2> 3200615 INFO  
(searcherExecutor-13406-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw 
x:collection1 c:control_collection) [n:127.0.0.1:55311_nst%2Ffw 
c:control_collection   x:collection1] o.a.s.c.SolrCore [collection1] Registered 
new searcher Searcher@515d848f[collection1] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 3200616 INFO  
(coreLoadExecutor-13405-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw c:control_collection   x:collection1] 
o.a.s.u.UpdateLog Could not find max version in index or recent updates, using 
new clock 1570633868307857408
   [junit4]   2> 3200620 INFO  
(coreZkRegister-13398-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw 
x:collection1 c:control_collection) [n:127.0.0.1:55311_nst%2Ffw 
c:control_collection s:shard1 r:core_node1 x:collection1] 
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 3200620 INFO  
(coreZkRegister-13398-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw 
x:collection1 c:control_collection) [n:127.0.0.1:55311_nst%2Ffw 
c:control_collection s:shard1 r:core_node1 x:collection1] 
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 3200621 INFO  
(coreZkRegister-13398-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw 
x:collection1 c:control_collection) [n:127.0.0.1:55311_nst%2Ffw 
c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SyncStrategy 
Sync replicas to http://127.0.0.1:55311/nst/fw/collection1/
   [junit4]   2> 3200621 INFO  
(coreZkRegister-13398-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw 
x:collection1 c:control_collection) [n:127.0.0.1:55311_nst%2Ffw 
c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SyncStrategy 
Sync Success - now sync replicas to me
   [junit4]   2> 3200621 INFO  
(coreZkRegister-13398-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw 
x:collection1 c:control_collection) [n:127.0.0.1:55311_nst%2Ffw 
c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SyncStrategy 
http://127.0.0.1:55311/nst/fw/collection1/ has no replicas
   [junit4]   2> 3200621 INFO  
(coreZkRegister-13398-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw 
x:collection1 c:control_collection) [n:127.0.0.1:55311_nst%2Ffw 
c:control_collection s:shard1 r:core_node1 x:collection1] 
o.a.s.c.ShardLeaderElectionContext Found all replicas participating in 
election, clear LIR
   [junit4]   2> 3200623 INFO  
(coreZkRegister-13398-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw 
x:collection1 c:control_collection) [n:127.0.0.1:55311_nst%2Ffw 
c:control_collection s:shard1 r:core_node1 x:collection1] 
o.a.s.c.ShardLeaderElectionContext I am the new leader: 
http://127.0.0.1:55311/nst/fw/collection1/ shard1
   [junit4]   2> 3200774 INFO  
(coreZkRegister-13398-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw 
x:collection1 c:control_collection) [n:127.0.0.1:55311_nst%2Ffw 
c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ZkController 
I am the leader, no recovery necessary
   [junit4]   2> 3200898 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 3200899 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:58917/solr ready
   [junit4]   2> 3200899 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.c.ChaosMonkey 
monkey: init - expire sessions:false cause connection loss:false
   [junit4]   2> 3200988 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.SolrTestCaseJ4 
Writing core.properties file to 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-1-001/cores/collection1
   [junit4]   2> 3200988 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-1-001
   [junit4]   2> 3200989 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.e.j.s.Server 
jetty-9.3.14.v20161028
   [junit4]   2> 3200992 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@39ea6b5{/nst/fw,null,AVAILABLE}
   [junit4]   2> 3200992 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@2aa50014{HTTP/1.1,[http/1.1]}{127.0.0.1:38778}
   [junit4]   2> 3200992 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.e.j.s.Server 
Started @3203077ms
   [junit4]   2> 3200992 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: 
{solr.data.dir=hdfs://localhost:36966/hdfs__localhost_36966__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-6.x_checkout_solr_build_solr-core_test_J0_temp_solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001_tempDir-002_jetty1,
 solrconfig=solrconfig.xml, hostContext=/nst/fw, hostPort=38778, 
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-1-001/cores}
   [junit4]   2> 3200994 ERROR 
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 3200994 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 
6.7.0
   [junit4]   2> 3200995 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 3200995 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 3200995 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2017-06-19T11:53:01.062Z
   [junit4]   2> 3200997 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in 
ZooKeeper)
   [junit4]   2> 3200997 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.c.SolrXmlConfig 
Loading container configuration from 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-1-001/solr.xml
   [junit4]   2> 3201005 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.u.UpdateShardHandler Creating UpdateShardHandler HTTP client with params: 
socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 3201006 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.c.ZkContainer 
Zookeeper client=127.0.0.1:58917/solr
   [junit4]   2> 3201021 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:38778_nst%2Ffw  
  ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 3201023 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:38778_nst%2Ffw  
  ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 3201025 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:38778_nst%2Ffw  
  ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:38778_nst%2Ffw
   [junit4]   2> 3201026 INFO  
(zkCallback-31827-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw    ] o.a.s.c.c.ZkStateReader Updated live nodes 
from ZooKeeper... (1) -> (2)
   [junit4]   2> 3201030 INFO  (zkCallback-31831-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 3201033 INFO  
(zkCallback-31836-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw    ] o.a.s.c.c.ZkStateReader Updated live nodes 
from ZooKeeper... (1) -> (2)
   [junit4]   2> 3201115 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:38778_nst%2Ffw  
  ] o.a.s.c.CorePropertiesLocator Found 1 core definitions underneath 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-1-001/cores
   [junit4]   2> 3201115 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:38778_nst%2Ffw  
  ] o.a.s.c.CorePropertiesLocator Cores are: [collection1]
   [junit4]   2> 3201116 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 
transient cores
   [junit4]   2> 3201116 INFO  
(OverseerStateUpdate-98164616659206148-127.0.0.1:55311_nst%2Ffw-n_0000000000) 
[n:127.0.0.1:55311_nst%2Ffw    ] o.a.s.c.o.ReplicaMutator Assigning new node to 
shard shard=shard1
   [junit4]   2> 3202127 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] o.a.s.c.SolrConfig 
Using Lucene MatchVersion: 6.7.0
   [junit4]   2> 3202147 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] o.a.s.s.IndexSchema 
[collection1] Schema name=test
   [junit4]   2> 3202265 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] o.a.s.s.IndexSchema 
Loaded schema test/1.0 with uniqueid field id
   [junit4]   2> 3202284 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.CoreContainer Creating SolrCore 'collection1' using configuration from 
collection collection1, trusted=true
   [junit4]   2> 3202285 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory 
solr.hdfs.home=hdfs://localhost:36966/solr_hdfs_home
   [junit4]   2> 3202285 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
   [junit4]   2> 3202285 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] o.a.s.c.SolrCore 
solr.RecoveryStrategy.Builder
   [junit4]   2> 3202285 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] o.a.s.c.SolrCore 
[[collection1] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-1-001/cores/collection1],
 dataDir=[hdfs://localhost:36966/solr_hdfs_home/collection1/core_node1/data/]
   [junit4]   2> 3202285 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.JmxMonitoredMap JMX monitoring is enabled. Adding Solr mbeans to JMX 
Server: com.sun.jmx.mbeanserver.JmxMBeanServer@78d9b800
   [junit4]   2> 3202285 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/collection1/core_node1/data/snapshot_metadata
   [junit4]   2> 3202292 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct 
memory allocation set to [true]
   [junit4]   2> 3202292 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of 
[8388608] will allocate [1] slabs and use ~[8388608] bytes
   [junit4]   2> 3202292 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache
   [junit4]   2> 3202297 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.s.b.BlockDirectory Block cache on write is disabled
   [junit4]   2> 3202298 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/collection1/core_node1/data
   [junit4]   2> 3202312 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/collection1/core_node1/data/index
   [junit4]   2> 3202317 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct 
memory allocation set to [true]
   [junit4]   2> 3202317 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of 
[8388608] will allocate [1] slabs and use ~[8388608] bytes
   [junit4]   2> 3202317 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache
   [junit4]   2> 3202322 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.s.b.BlockDirectory Block cache on write is disabled
   [junit4]   2> 3202322 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy: 
minMergeSize=1000, mergeFactor=40, maxMergeSize=9223372036854775807, 
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false, 
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=1.0]
   [junit4]   2> 3202341 INFO  (IPC Server handler 2 on 36966) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:32931 is 
added to blk_1073741826_1002{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-43345fcd-5e09-4aa7-925b-a6e9805c0eaf:NORMAL:127.0.0.1:32931|RBW],
 
ReplicaUC[[DISK]DS-96f0454d-71a1-441a-bb20-648eff417ff3:NORMAL:127.0.0.1:33736|RBW]]}
 size 0
   [junit4]   2> 3202341 INFO  (IPC Server handler 1 on 36966) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33736 is 
added to blk_1073741826_1002{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-43345fcd-5e09-4aa7-925b-a6e9805c0eaf:NORMAL:127.0.0.1:32931|RBW],
 
ReplicaUC[[DISK]DS-35bfeff3-61da-4bcf-b500-98d455d027e8:NORMAL:127.0.0.1:33736|FINALIZED]]}
 size 0
   [junit4]   2> 3202346 WARN  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = 
requestHandler,name = /dump,class = DumpRequestHandler,attributes = 
{initParams=a, name=/dump, class=DumpRequestHandler},args = 
{defaults={a=A,b=B}}}
   [junit4]   2> 3203295 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.UpdateHandler Using UpdateLog implementation: 
org.apache.solr.update.HdfsUpdateLog
   [junit4]   2> 3203295 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] o.a.s.u.UpdateLog 
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 3203295 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=2
   [junit4]   2> 3203303 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 3203303 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 3203305 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.AlcoholicMergePolicy: [AlcoholicMergePolicy: 
minMergeSize=0, mergeFactor=10, maxMergeSize=1961424476, 
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, 
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.1]
   [junit4]   2> 3203308 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.s.SolrIndexSearcher Opening [Searcher@6eec6405[collection1] main]
   [junit4]   2> 3203309 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 3203309 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 3203310 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] 
o.a.s.h.ReplicationHandler Commits will be reserved for  10000
   [junit4]   2> 3203311 INFO  
(searcherExecutor-13417-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw 
x:collection1 c:collection1) [n:127.0.0.1:38778_nst%2Ffw c:collection1   
x:collection1] o.a.s.c.SolrCore [collection1] Registered new searcher 
Searcher@6eec6405[collection1] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 3203312 INFO  
(coreLoadExecutor-13416-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1   x:collection1] o.a.s.u.UpdateLog 
Could not find max version in index or recent updates, using new clock 
1570633871134818304
   [junit4]   2> 3203317 INFO  
(coreZkRegister-13411-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw 
x:collection1 c:collection1) [n:127.0.0.1:38778_nst%2Ffw c:collection1 s:shard1 
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext Enough replicas 
found to continue.
   [junit4]   2> 3203317 INFO  
(coreZkRegister-13411-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw 
x:collection1 c:collection1) [n:127.0.0.1:38778_nst%2Ffw c:collection1 s:shard1 
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext I may be the new 
leader - try and sync
   [junit4]   2> 3203317 INFO  
(coreZkRegister-13411-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw 
x:collection1 c:collection1) [n:127.0.0.1:38778_nst%2Ffw c:collection1 s:shard1 
r:core_node1 x:collection1] o.a.s.c.SyncStrategy Sync replicas to 
http://127.0.0.1:38778/nst/fw/collection1/
   [junit4]   2> 3203318 INFO  
(coreZkRegister-13411-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw 
x:collection1 c:collection1) [n:127.0.0.1:38778_nst%2Ffw c:collection1 s:shard1 
r:core_node1 x:collection1] o.a.s.c.SyncStrategy Sync Success - now sync 
replicas to me
   [junit4]   2> 3203319 INFO  
(coreZkRegister-13411-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw 
x:collection1 c:collection1) [n:127.0.0.1:38778_nst%2Ffw c:collection1 s:shard1 
r:core_node1 x:collection1] o.a.s.c.SyncStrategy 
http://127.0.0.1:38778/nst/fw/collection1/ has no replicas
   [junit4]   2> 3203319 INFO  
(coreZkRegister-13411-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw 
x:collection1 c:collection1) [n:127.0.0.1:38778_nst%2Ffw c:collection1 s:shard1 
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext Found all 
replicas participating in election, clear LIR
   [junit4]   2> 3203322 INFO  
(coreZkRegister-13411-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw 
x:collection1 c:collection1) [n:127.0.0.1:38778_nst%2Ffw c:collection1 s:shard1 
r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext I am the new 
leader: http://127.0.0.1:38778/nst/fw/collection1/ shard1
   [junit4]   2> 3203474 INFO  
(coreZkRegister-13411-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw 
x:collection1 c:collection1) [n:127.0.0.1:38778_nst%2Ffw c:collection1 s:shard1 
r:core_node1 x:collection1] o.a.s.c.ZkController I am the leader, no recovery 
necessary
   [junit4]   2> 3203837 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.SolrTestCaseJ4 
Writing core.properties file to 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-2-001/cores/collection1
   [junit4]   2> 3203838 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-2-001
   [junit4]   2> 3203838 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.e.j.s.Server 
jetty-9.3.14.v20161028
   [junit4]   2> 3203839 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@5ffa6be9{/nst/fw,null,AVAILABLE}
   [junit4]   2> 3203840 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@67961839{HTTP/1.1,[http/1.1]}{127.0.0.1:52478}
   [junit4]   2> 3203840 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.e.j.s.Server 
Started @3205924ms
   [junit4]   2> 3203840 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: 
{solr.data.dir=hdfs://localhost:36966/hdfs__localhost_36966__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-6.x_checkout_solr_build_solr-core_test_J0_temp_solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001_tempDir-002_jetty2,
 solrconfig=solrconfig.xml, hostContext=/nst/fw, hostPort=52478, 
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-2-001/cores}
   [junit4]   2> 3203840 ERROR 
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 3203841 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 
6.7.0
   [junit4]   2> 3203842 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 3203842 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 3203842 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2017-06-19T11:53:03.909Z
   [junit4]   2> 3203844 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in 
ZooKeeper)
   [junit4]   2> 3203844 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.c.SolrXmlConfig 
Loading container configuration from 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-2-001/solr.xml
   [junit4]   2> 3203850 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.u.UpdateShardHandler Creating UpdateShardHandler HTTP client with params: 
socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 3203850 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.c.ZkContainer 
Zookeeper client=127.0.0.1:58917/solr
   [junit4]   2> 3203857 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:52478_nst%2Ffw  
  ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 3203858 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:52478_nst%2Ffw  
  ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 3203860 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:52478_nst%2Ffw  
  ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:52478_nst%2Ffw
   [junit4]   2> 3203861 INFO  
(zkCallback-31836-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw    ] o.a.s.c.c.ZkStateReader Updated live nodes 
from ZooKeeper... (2) -> (3)
   [junit4]   2> 3203861 INFO  (zkCallback-31831-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 3203861 INFO  
(zkCallback-31827-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw    ] o.a.s.c.c.ZkStateReader Updated live nodes 
from ZooKeeper... (2) -> (3)
   [junit4]   2> 3203866 INFO  
(zkCallback-31842-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw    ] o.a.s.c.c.ZkStateReader Updated live nodes 
from ZooKeeper... (2) -> (3)
   [junit4]   2> 3203938 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:52478_nst%2Ffw  
  ] o.a.s.c.CorePropertiesLocator Found 1 core definitions underneath 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-2-001/cores
   [junit4]   2> 3203938 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:52478_nst%2Ffw  
  ] o.a.s.c.CorePropertiesLocator Cores are: [collection1]
   [junit4]   2> 3203938 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 
transient cores
   [junit4]   2> 3203940 INFO  
(OverseerStateUpdate-98164616659206148-127.0.0.1:55311_nst%2Ffw-n_0000000000) 
[n:127.0.0.1:55311_nst%2Ffw    ] o.a.s.c.o.ReplicaMutator Assigning new node to 
shard shard=shard1
   [junit4]   2> 3204950 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] o.a.s.c.SolrConfig 
Using Lucene MatchVersion: 6.7.0
   [junit4]   2> 3204962 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] o.a.s.s.IndexSchema 
[collection1] Schema name=test
   [junit4]   2> 3205063 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] o.a.s.s.IndexSchema 
Loaded schema test/1.0 with uniqueid field id
   [junit4]   2> 3205073 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.CoreContainer Creating SolrCore 'collection1' using configuration from 
collection collection1, trusted=true
   [junit4]   2> 3205074 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory 
solr.hdfs.home=hdfs://localhost:36966/solr_hdfs_home
   [junit4]   2> 3205074 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
   [junit4]   2> 3205074 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] o.a.s.c.SolrCore 
solr.RecoveryStrategy.Builder
   [junit4]   2> 3205074 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] o.a.s.c.SolrCore 
[[collection1] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-2-001/cores/collection1],
 dataDir=[hdfs://localhost:36966/solr_hdfs_home/collection1/core_node2/data/]
   [junit4]   2> 3205074 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.JmxMonitoredMap JMX monitoring is enabled. Adding Solr mbeans to JMX 
Server: com.sun.jmx.mbeanserver.JmxMBeanServer@78d9b800
   [junit4]   2> 3205075 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/collection1/core_node2/data/snapshot_metadata
   [junit4]   2> 3205084 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct 
memory allocation set to [true]
   [junit4]   2> 3205084 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of 
[8388608] will allocate [1] slabs and use ~[8388608] bytes
   [junit4]   2> 3205084 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache
   [junit4]   2> 3205091 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.s.b.BlockDirectory Block cache on write is disabled
   [junit4]   2> 3205092 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/collection1/core_node2/data
   [junit4]   2> 3205109 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/collection1/core_node2/data/index
   [junit4]   2> 3205116 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct 
memory allocation set to [true]
   [junit4]   2> 3205116 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of 
[8388608] will allocate [1] slabs and use ~[8388608] bytes
   [junit4]   2> 3205116 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache
   [junit4]   2> 3205123 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.s.b.BlockDirectory Block cache on write is disabled
   [junit4]   2> 3205123 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy: 
minMergeSize=1000, mergeFactor=40, maxMergeSize=9223372036854775807, 
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false, 
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=1.0]
   [junit4]   2> 3205130 INFO  (IPC Server handler 1 on 36966) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:32931 is 
added to blk_1073741827_1003{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-96f0454d-71a1-441a-bb20-648eff417ff3:NORMAL:127.0.0.1:33736|RBW],
 
ReplicaUC[[DISK]DS-5284bd1e-8b3d-4e0a-9186-4be1c3da4cd2:NORMAL:127.0.0.1:32931|FINALIZED]]}
 size 0
   [junit4]   2> 3205130 INFO  (IPC Server handler 4 on 36966) [    ] 
BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:33736 is 
added to blk_1073741827_1003{UCState=UNDER_CONSTRUCTION, truncateBlock=null, 
primaryNodeIndex=-1, 
replicas=[ReplicaUC[[DISK]DS-96f0454d-71a1-441a-bb20-648eff417ff3:NORMAL:127.0.0.1:33736|RBW],
 
ReplicaUC[[DISK]DS-5284bd1e-8b3d-4e0a-9186-4be1c3da4cd2:NORMAL:127.0.0.1:32931|FINALIZED]]}
 size 0
   [junit4]   2> 3205136 WARN  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = 
requestHandler,name = /dump,class = DumpRequestHandler,attributes = 
{initParams=a, name=/dump, class=DumpRequestHandler},args = 
{defaults={a=A,b=B}}}
   [junit4]   2> 3205171 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.UpdateHandler Using UpdateLog implementation: 
org.apache.solr.update.HdfsUpdateLog
   [junit4]   2> 3205171 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] o.a.s.u.UpdateLog 
Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 3205171 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=2
   [junit4]   2> 3205183 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 3205183 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 3205185 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class 
org.apache.lucene.index.AlcoholicMergePolicy: [AlcoholicMergePolicy: 
minMergeSize=0, mergeFactor=10, maxMergeSize=1961424476, 
maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, 
maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, 
noCFSRatio=0.1]
   [junit4]   2> 3205188 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.s.SolrIndexSearcher Opening [Searcher@668d9eeb[collection1] main]
   [junit4]   2> 3205189 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 3205189 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 3205189 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] 
o.a.s.h.ReplicationHandler Commits will be reserved for  10000
   [junit4]   2> 3205190 INFO  
(searcherExecutor-13428-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 c:collection1) [n:127.0.0.1:52478_nst%2Ffw c:collection1   
x:collection1] o.a.s.c.SolrCore [collection1] Registered new searcher 
Searcher@668d9eeb[collection1] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 3205192 INFO  
(coreLoadExecutor-13427-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw c:collection1   x:collection1] o.a.s.u.UpdateLog 
Could not find max version in index or recent updates, using new clock 
1570633873106141184
   [junit4]   2> 3205196 INFO  
(coreZkRegister-13422-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 c:collection1) [n:127.0.0.1:52478_nst%2Ffw c:collection1 s:shard1 
r:core_node2 x:collection1] o.a.s.c.ZkController Core needs to 
recover:collection1
   [junit4]   2> 3205196 INFO  
(updateExecutor-31839-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.u.DefaultSolrCoreState 
Running recovery
   [junit4]   2> 3205196 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.c.RecoveryStrategy 
Starting recovery process. recoveringAfterStartup=true
   [junit4]   2> 3205197 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.c.RecoveryStrategy 
###### startupVersions=[[]]
   [junit4]   2> 3205197 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.c.RecoveryStrategy 
Begin buffering updates. core=[collection1]
   [junit4]   2> 3205197 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.u.UpdateLog Starting 
to buffer updates. HDFSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 3205197 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.c.RecoveryStrategy 
Publishing state of core [collection1] as recovering, leader is 
[http://127.0.0.1:38778/nst/fw/collection1/] and I am 
[http://127.0.0.1:52478/nst/fw/collection1/]
   [junit4]   2> 3205199 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.c.RecoveryStrategy 
Sending prep recovery command to [http://127.0.0.1:38778/nst/fw]; 
[WaitForState: 
action=PREPRECOVERY&core=collection1&nodeName=127.0.0.1:52478_nst%252Ffw&coreNodeName=core_node2&state=recovering&checkLive=true&onlyIfLeader=true&onlyIfLeaderActive=true]
   [junit4]   2> 3205200 INFO  (qtp195928662-129720) 
[n:127.0.0.1:38778_nst%2Ffw    ] o.a.s.h.a.PrepRecoveryOp Going to wait for 
coreNodeName: core_node2, state: recovering, checkLive: true, onlyIfLeader: 
true, onlyIfLeaderActive: true
   [junit4]   2> 3205200 INFO  (qtp195928662-129720) 
[n:127.0.0.1:38778_nst%2Ffw    ] o.a.s.h.a.PrepRecoveryOp Will wait a max of 
183 seconds to see collection1 (shard1 of collection1) have state: recovering
   [junit4]   2> 3205200 INFO  (qtp195928662-129720) 
[n:127.0.0.1:38778_nst%2Ffw    ] o.a.s.h.a.PrepRecoveryOp In 
WaitForState(recovering): collection=collection1, shard=shard1, 
thisCore=collection1, leaderDoesNotNeedRecovery=false, isLeader? true, 
live=true, checkLive=true, currentState=down, localState=active, 
nodeName=127.0.0.1:52478_nst%2Ffw, coreNodeName=core_node2, 
onlyIfActiveCheckResult=false, nodeProps: 
core_node2:{"core":"collection1","base_url":"http://127.0.0.1:52478/nst/fw","node_name":"127.0.0.1:52478_nst%2Ffw","state":"down"}
   [junit4]   2> 3205526 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.SolrTestCaseJ4 
Writing core.properties file to 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-3-001/cores/collection1
   [junit4]   2> 3205526 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.AbstractFullDistribZkTestBase create jetty 3 in directory 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-3-001
   [junit4]   2> 3205527 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.e.j.s.Server 
jetty-9.3.14.v20161028
   [junit4]   2> 3205528 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@47490617{/nst/fw,null,AVAILABLE}
   [junit4]   2> 3205528 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@1cd8785c{HTTP/1.1,[http/1.1]}{127.0.0.1:54118}
   [junit4]   2> 3205529 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.e.j.s.Server 
Started @3207613ms
   [junit4]   2> 3205529 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: 
{solr.data.dir=hdfs://localhost:36966/hdfs__localhost_36966__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-6.x_checkout_solr_build_solr-core_test_J0_temp_solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001_tempDir-002_jetty3,
 solrconfig=solrconfig.xml, hostContext=/nst/fw, hostPort=54118, 
coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-3-001/cores}
   [junit4]   2> 3205530 ERROR 
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 3205530 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 
6.7.0
   [junit4]   2> 3205530 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 3205530 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 3205530 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2017-06-19T11:53:05.597Z
   [junit4]   2> 3205532 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in 
ZooKeeper)
   [junit4]   2> 3205533 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.c.SolrXmlConfig 
Loading container configuration from 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-3-001/solr.xml
   [junit4]   2> 3205538 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] 
o.a.s.u.UpdateShardHandler Creating UpdateShardHandler HTTP client with params: 
socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 3205538 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [    ] o.a.s.c.ZkContainer 
Zookeeper client=127.0.0.1:58917/solr
   [junit4]   2> 3205546 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:54118_nst%2Ffw  
  ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
   [junit4]   2> 3205547 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:54118_nst%2Ffw  
  ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 3205549 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:54118_nst%2Ffw  
  ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:54118_nst%2Ffw
   [junit4]   2> 3205550 INFO  
(zkCallback-31827-thread-1-processing-n:127.0.0.1:55311_nst%2Ffw) 
[n:127.0.0.1:55311_nst%2Ffw    ] o.a.s.c.c.ZkStateReader Updated live nodes 
from ZooKeeper... (3) -> (4)
   [junit4]   2> 3205550 INFO  
(zkCallback-31836-thread-1-processing-n:127.0.0.1:38778_nst%2Ffw) 
[n:127.0.0.1:38778_nst%2Ffw    ] o.a.s.c.c.ZkStateReader Updated live nodes 
from ZooKeeper... (3) -> (4)
   [junit4]   2> 3205550 INFO  (zkCallback-31831-thread-1) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 3205550 INFO  
(zkCallback-31842-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw) 
[n:127.0.0.1:52478_nst%2Ffw    ] o.a.s.c.c.ZkStateReader Updated live nodes 
from ZooKeeper... (3) -> (4)
   [junit4]   2> 3205551 INFO  
(zkCallback-31849-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw    ] o.a.s.c.c.ZkStateReader Updated live nodes 
from ZooKeeper... (3) -> (4)
   [junit4]   2> 3205655 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:54118_nst%2Ffw  
  ] o.a.s.c.CorePropertiesLocator Found 1 core definitions underneath 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-3-001/cores
   [junit4]   2> 3205655 INFO  
(TEST-StressHdfsTest.test-seed#[2C5F898E344E9D7D]) [n:127.0.0.1:54118_nst%2Ffw  
  ] o.a.s.c.CorePropertiesLocator Cores are: [collection1]
   [junit4]   2> 3205656 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 
transient cores
   [junit4]   2> 3205656 INFO  
(OverseerStateUpdate-98164616659206148-127.0.0.1:55311_nst%2Ffw-n_0000000000) 
[n:127.0.0.1:55311_nst%2Ffw    ] o.a.s.c.o.ReplicaMutator Assigning new node to 
shard shard=shard1
   [junit4]   2> 3206201 INFO  (qtp195928662-129720) 
[n:127.0.0.1:38778_nst%2Ffw    ] o.a.s.h.a.PrepRecoveryOp In 
WaitForState(recovering): collection=collection1, shard=shard1, 
thisCore=collection1, leaderDoesNotNeedRecovery=false, isLeader? true, 
live=true, checkLive=true, currentState=recovering, localState=active, 
nodeName=127.0.0.1:52478_nst%2Ffw, coreNodeName=core_node2, 
onlyIfActiveCheckResult=false, nodeProps: 
core_node2:{"core":"collection1","dataDir":"hdfs://localhost:36966/solr_hdfs_home/collection1/core_node2/data/","base_url":"http://127.0.0.1:52478/nst/fw","node_name":"127.0.0.1:52478_nst%2Ffw","state":"recovering","ulogDir":"hdfs://localhost:36966/solr_hdfs_home/collection1/core_node2/data/tlog"}
   [junit4]   2> 3206201 INFO  (qtp195928662-129720) 
[n:127.0.0.1:38778_nst%2Ffw    ] o.a.s.h.a.PrepRecoveryOp Waited coreNodeName: 
core_node2, state: recovering, checkLive: true, onlyIfLeader: true for: 1 
seconds.
   [junit4]   2> 3206201 INFO  (qtp195928662-129720) 
[n:127.0.0.1:38778_nst%2Ffw    ] o.a.s.s.HttpSolrCall [admin] webapp=null 
path=/admin/cores 
params={nodeName=127.0.0.1:52478_nst%252Ffw&onlyIfLeaderActive=true&core=collection1&coreNodeName=core_node2&action=PREPRECOVERY&checkLive=true&state=recovering&onlyIfLeader=true&wt=javabin&version=2}
 status=0 QTime=1000
   [junit4]   2> 3206676 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] o.a.s.c.SolrConfig 
Using Lucene MatchVersion: 6.7.0
   [junit4]   2> 3206687 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] o.a.s.s.IndexSchema 
[collection1] Schema name=test
   [junit4]   2> 3206701 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.c.RecoveryStrategy 
Attempting to PeerSync from [http://127.0.0.1:38778/nst/fw/collection1/] - 
recoveringAfterStartup=[true]
   [junit4]   2> 3206701 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.u.PeerSync PeerSync: 
core=collection1 url=http://127.0.0.1:52478/nst/fw START 
replicas=[http://127.0.0.1:38778/nst/fw/collection1/] nUpdates=100
   [junit4]   2> 3206704 INFO  (qtp195928662-129722) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1 s:shard1 r:core_node1 x:collection1] 
o.a.s.u.IndexFingerprint IndexFingerprint millis:0.0 
result:{maxVersionSpecified=9223372036854775807, maxVersionEncountered=0, 
maxInHash=0, versionsHash=0, numVersions=0, numDocs=0, maxDoc=0}
   [junit4]   2> 3206704 INFO  (qtp195928662-129722) 
[n:127.0.0.1:38778_nst%2Ffw c:collection1 s:shard1 r:core_node1 x:collection1] 
o.a.s.c.S.Request [collection1]  webapp=/nst/fw path=/get 
params={distrib=false&qt=/get&getFingerprint=9223372036854775807&wt=javabin&version=2}
 status=0 QTime=0
   [junit4]   2> 3206705 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.u.IndexFingerprint 
IndexFingerprint millis:0.0 result:{maxVersionSpecified=9223372036854775807, 
maxVersionEncountered=0, maxInHash=0, versionsHash=0, numVersions=0, numDocs=0, 
maxDoc=0}
   [junit4]   2> 3206705 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.u.PeerSync We are 
already in sync. No need to do a PeerSync 
   [junit4]   2> 3206705 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.u.DirectUpdateHandler2 
start 
commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3206705 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.u.DirectUpdateHandler2 
No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 3206705 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.u.DirectUpdateHandler2 
end_commit_flush
   [junit4]   2> 3206705 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.c.RecoveryStrategy 
PeerSync stage of recovery was successful.
   [junit4]   2> 3206705 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.c.RecoveryStrategy 
Replaying updates buffered during PeerSync.
   [junit4]   2> 3206706 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.c.RecoveryStrategy No 
replay needed.
   [junit4]   2> 3206706 INFO  
(recoveryExecutor-31840-thread-1-processing-n:127.0.0.1:52478_nst%2Ffw 
x:collection1 s:shard1 c:collection1 r:core_node2) [n:127.0.0.1:52478_nst%2Ffw 
c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.c.RecoveryStrategy 
Registering as Active after recovery.
   [junit4]   2> 3206787 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] o.a.s.s.IndexSchema 
Loaded schema test/1.0 with uniqueid field id
   [junit4]   2> 3206795 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.CoreContainer Creating SolrCore 'collection1' using configuration from 
collection collection1, trusted=true
   [junit4]   2> 3206795 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory 
solr.hdfs.home=hdfs://localhost:36966/solr_hdfs_home
   [junit4]   2> 3206795 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
   [junit4]   2> 3206796 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] o.a.s.c.SolrCore 
solr.RecoveryStrategy.Builder
   [junit4]   2> 3206796 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] o.a.s.c.SolrCore 
[[collection1] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001/shard-3-001/cores/collection1],
 dataDir=[hdfs://localhost:36966/solr_hdfs_home/collection1/core_node3/data/]
   [junit4]   2> 3206796 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.JmxMonitoredMap JMX monitoring is enabled. Adding Solr mbeans to JMX 
Server: com.sun.jmx.mbeanserver.JmxMBeanServer@78d9b800
   [junit4]   2> 3206796 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/collection1/core_node3/data/snapshot_metadata
   [junit4]   2> 3206802 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct 
memory allocation set to [true]
   [junit4]   2> 3206802 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of 
[8388608] will allocate [1] slabs and use ~[8388608] bytes
   [junit4]   2> 3206802 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache
   [junit4]   2> 3206809 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.s.b.BlockDirectory Block cache on write is disabled
   [junit4]   2> 3206809 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/collection1/core_node3/data
   [junit4]   2> 3206821 INFO  
(coreLoadExecutor-13438-thread-1-processing-n:127.0.0.1:54118_nst%2Ffw) 
[n:127.0.0.1:54118_nst%2Ffw c:collection1   x:collection1] 
o.a.s.c.HdfsDirectoryFactory creating directory factory for path 
hdfs://localhost:36966/solr_hdfs_home/collection1/core_node3/data/index
   [junit4]   2> 3206826 INFO  (coreLoadExecutor-13438-thread-1-processing

[...truncated too long message...]

ricsSystem.java:72)
   [junit4]   2>        at 
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.shutdown(DefaultMetricsSystem.java:68)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics.shutdown(NameNodeMetrics.java:171)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.stop(NameNode.java:872)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1726)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.MiniDFSCluster.shutdown(MiniDFSCluster.java:1705)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.HdfsTestUtil.teardownClass(HdfsTestUtil.java:203)
   [junit4]   2>        at 
org.apache.solr.cloud.hdfs.StressHdfsTest.teardownClass(StressHdfsTest.java:77)
   [junit4]   2>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
Method)
   [junit4]   2>        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   [junit4]   2>        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   [junit4]   2>        at java.lang.reflect.Method.invoke(Method.java:498)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1713)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:870)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
   [junit4]   2>        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
   [junit4]   2>        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   [junit4]   2>        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
   [junit4]   2>        at java.lang.Thread.run(Thread.java:748)
   [junit4]   2> Caused by: java.lang.NullPointerException
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.blockmanagement.BlocksMap.size(BlocksMap.java:203)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.getTotalBlocks(BlockManager.java:3370)
   [junit4]   2>        at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlocksTotal(FSNamesystem.java:5729)
   [junit4]   2>        ... 53 more
   [junit4]   2> 3384324 INFO  
(SUITE-StressHdfsTest-seed#[2C5F898E344E9D7D]-worker) [    ] 
o.a.s.SolrTestCaseJ4 ###deleteCore
   [junit4]   2> NOTE: leaving temporary files on disk at: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-6.x/checkout/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_2C5F898E344E9D7D-001
   [junit4]   2> Jun 19, 2017 11:56:04 AM 
com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
   [junit4]   2> WARNING: Will linger awaiting termination of 34 leaked 
thread(s).
   [junit4]   2> NOTE: test params are: codec=Asserting(Lucene62): 
{range_facet_l_dv=TestBloomFilteredLucenePostings(BloomFilteringPostingsFormat(Lucene50(blocksize=128))),
 multiDefault=PostingsFormat(name=LuceneVarGapFixedInterval), 
intDefault=PostingsFormat(name=Asserting), 
id_i1=PostingsFormat(name=Asserting), 
id=TestBloomFilteredLucenePostings(BloomFilteringPostingsFormat(Lucene50(blocksize=128))),
 range_facet_i_dv=PostingsFormat(name=LuceneVarGapFixedInterval), 
text=Lucene50(blocksize=128), 
range_facet_l=PostingsFormat(name=LuceneVarGapFixedInterval), 
timestamp=PostingsFormat(name=LuceneVarGapFixedInterval), 
txt_t=Lucene50(blocksize=128)}, 
docValues:{range_facet_l_dv=DocValuesFormat(name=Asserting), 
_version_=DocValuesFormat(name=Lucene54), 
multiDefault=DocValuesFormat(name=Memory), 
intDefault=DocValuesFormat(name=Lucene54), 
id_i1=DocValuesFormat(name=Lucene54), 
range_facet_i_dv=DocValuesFormat(name=Memory), 
id=DocValuesFormat(name=Asserting), text=DocValuesFormat(name=Lucene54), 
intDvoDefault=DocValuesFormat(name=Asserting), 
timestamp=DocValuesFormat(name=Memory), 
range_facet_l=DocValuesFormat(name=Memory), 
txt_t=DocValuesFormat(name=Lucene54)}, maxPointsInLeafNode=186, 
maxMBSortInHeap=6.205771100611032, 
sim=RandomSimilarity(queryNorm=true,coord=crazy): {}, locale=ar, 
timezone=Asia/Aqtau
   [junit4]   2> NOTE: Linux 3.13.0-88-generic amd64/Oracle Corporation 
1.8.0_131 (64-bit)/cpus=4,threads=2,free=268578224,total=535298048
   [junit4]   2> NOTE: All tests run in this JVM: [TestSortByMinMaxFunction, 
TestFieldCacheSanityChecker, CurrencyFieldOpenExchangeTest, PeerSyncTest, 
SpatialRPTFieldTypeTest, ScriptEngineTest, TestReloadAndDeleteDocs, 
TestDownShardTolerantSearch, TestHighFrequencyDictionaryFactory, 
TestFunctionQuery, TestValueSourceCache, TestReplicaProperties, LargeFieldTest, 
TestReRankQParserPlugin, UniqFieldsUpdateProcessorFactoryTest, 
ClusterStateTest, SignatureUpdateProcessorFactoryTest, AlternateDirectoryTest, 
TestScoreJoinQPScore, TestFieldSortValues, 
DistributedFacetPivotSmallAdvancedTest, PathHierarchyTokenizerFactoryTest, 
ShufflingReplicaListTransformerTest, TestOmitPositions, TestBadConfig, 
CustomCollectionTest, TestPushWriter, TestSolrCoreProperties, 
TestGeoJSONResponseWriter, DistributedQueryComponentCustomSortTest, 
OverseerStatusTest, SortSpecParsingTest, JsonValidatorTest, 
LeaderElectionIntegrationTest, TestJoin, FullHLLTest, TestCodecSupport, 
ChangedSchemaMergeTest, TestSchemaResource, TestReplicationHandler, 
TestAuthenticationFramework, TestCrossCoreJoin, BlockJoinFacetRandomTest, 
PolyFieldTest, AliasIntegrationTest, TestSolrDeletionPolicy2, 
TestSizeLimitedDistributedMap, SecurityConfHandlerTest, 
TestSha256AuthenticationProvider, TestQuerySenderNoQuery, TestSSLRandomization, 
TestJsonFacets, CoreAdminCreateDiscoverTest, TestBlendedInfixSuggestions, 
TestJmxIntegration, BasicZkTest, TestSolr4Spatial2, 
TestDocBasedVersionConstraints, DistributedFacetExistsSmallTest, 
TestHashQParserPlugin, SuggestComponentTest, TestInPlaceUpdatesDistrib, 
SyncSliceTest, RegexBoostProcessorTest, StandardRequestHandlerTest, 
TestIBSimilarityFactory, TestHttpShardHandlerFactory, TestJettySolrRunner, 
TestFoldingMultitermQuery, ReplicationFactorTest, TestRandomFaceting, 
TestSolrCloudWithHadoopAuthPlugin, HighlighterConfigTest, ReturnFieldsTest, 
TestXmlQParser, ExitableDirectoryReaderTest, TestStressLiveNodes, 
TestQueryWrapperFilter, QueryEqualityTest, TestMacros, UUIDFieldTest, 
TestBinaryResponseWriter, TestLazyCores, ConfigSetsAPITest, 
RecoveryAfterSoftCommitTest, TestClusterStateMutator, TestSolrCloudSnapshots, 
TestSerializedLuceneMatchVersion, SolrCLIZkUtilsTest, BasicAuthIntegrationTest, 
TestCloudRecovery, TestSolrConfigHandler, TestCloudNestedDocsSort, 
TestMissingGroups, SharedFSAutoReplicaFailoverTest, DocValuesNotIndexedTest, 
TestNumericTerms32, TestCoreAdminApis, TestInfoStreamLogging, 
OverseerCollectionConfigSetProcessorTest, 
DistributedQueryElevationComponentTest, 
IgnoreCommitOptimizeUpdateProcessorFactoryTest, 
TestSolrConfigHandlerConcurrent, DistribJoinFromCollectionTest, 
AnalyticsMergeStrategyTest, ConfigureRecoveryStrategyTest, 
TestClassNameShortening, TestFaceting, BaseCdcrDistributedZkTest, 
NumericFieldsTest, TestSolr4Spatial, HdfsBasicDistributedZk2Test, 
ClassificationUpdateProcessorTest, TestSchemalessBufferedUpdates, 
BasicDistributedZkTest, RecoveryZkTest, ClusterStateUpdateTest, 
ZkSolrClientTest, ShardRoutingCustomTest, TestHashPartitioner, 
TermVectorComponentDistributedTest, TestReload, TestRangeQuery, 
TestMergePolicyConfig, MultiTermTest, TestDocSet, TestSearchPerf, 
MinimalSchemaTest, TestConfig, OutputWriterTest, TestFuzzyAnalyzedSuggestions, 
ExternalFileFieldSortTest, DirectSolrConnectionTest, NotRequiredUniqueKeyTest, 
TestLuceneMatchVersion, SpellPossibilityIteratorTest, SynonymTokenizerTest, 
TestXIncludeConfig, EchoParamsTest, TestPerFieldSimilarity, 
TestLMJelinekMercerSimilarityFactory, TestFastWriter, TestJmxMonitoredMap, 
ResourceLoaderTest, OpenExchangeRatesOrgProviderTest, URLClassifyProcessorTest, 
PluginInfoTest, TestFastLRUCache, DateMathParserTest, 
ChaosMonkeyNothingIsSafeTest, CdcrReplicationHandlerTest, 
ConnectionManagerTest, DeleteNodeTest, DeleteStatusTest, 
DistribDocExpirationUpdateProcessorTest, MultiThreadedOCPTest, 
TestCloudJSONFacetJoinDomain, TestConfigSetsAPI, TestLeaderElectionZkExpiry, 
TestLocalFSCloudBackupRestore, TestOnReconnectListenerSupport, 
TestRandomFlRTGCloud, TestSolrCloudWithKerberosAlt, TestStressInPlaceUpdates, 
TestTolerantUpdateProcessorCloud, TestTolerantUpdateProcessorRandomCloud, 
HdfsChaosMonkeySafeLeaderTest, HdfsUnloadDistributedZkTest, StressHdfsTest]
   [junit4] Completed [596/715 (1!)] on J0 in 196.67s, 1 test, 1 error <<< 
FAILURES!

[...truncated 43820 lines...]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to