Build: https://ci-builds.apache.org/job/Lucene/job/Lucene-Solr-Tests-8.11/529/

2 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.api.collections.HdfsCloudIncrementalBackupTest

Error Message:
1 thread leaked from SUITE scope at 
org.apache.solr.cloud.api.collections.HdfsCloudIncrementalBackupTest:     1) 
Thread[id=9929, name=Command processor, state=WAITING, 
group=TGRP-HdfsCloudIncrementalBackupTest]         at 
sun.misc.Unsafe.park(Native Method)         at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)         at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
         at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)     
    at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1291)
         at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1275)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: 1 thread leaked from SUITE 
scope at org.apache.solr.cloud.api.collections.HdfsCloudIncrementalBackupTest: 
   1) Thread[id=9929, name=Command processor, state=WAITING, 
group=TGRP-HdfsCloudIncrementalBackupTest]
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
        at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
        at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1291)
        at 
org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1275)
        at __randomizedtesting.SeedInfo.seed([C0258A7B5C6D0A61]:0)


FAILED:  
org.apache.solr.cloud.api.collections.ShardSplitTest.testSplitMixedReplicaTypes

Error Message:
Timeout occurred while waiting response from server at: 
http://127.0.0.1:42131/onek/c

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: Timeout occurred while 
waiting response from server at: http://127.0.0.1:42131/onek/c
        at 
__randomizedtesting.SeedInfo.seed([C0258A7B5C6D0A61:78E6DEDBA0B6DF14]:0)
        at 
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:692)
        at 
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:266)
        at 
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:248)
        at 
org.apache.solr.client.solrj.impl.LBSolrClient.doRequest(LBSolrClient.java:369)
        at 
org.apache.solr.client.solrj.impl.LBSolrClient.request(LBSolrClient.java:297)
        at 
org.apache.solr.client.solrj.impl.BaseCloudSolrClient.sendRequest(BaseCloudSolrClient.java:1188)
        at 
org.apache.solr.client.solrj.impl.BaseCloudSolrClient.requestWithRetryOnStaleState(BaseCloudSolrClient.java:934)
        at 
org.apache.solr.client.solrj.impl.BaseCloudSolrClient.request(BaseCloudSolrClient.java:866)
        at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:225)
        at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:242)
        at 
org.apache.solr.cloud.api.collections.ShardSplitTest.doSplitMixedReplicaTypes(ShardSplitTest.java:365)
        at 
org.apache.solr.cloud.api.collections.ShardSplitTest.testSplitMixedReplicaTypes(ShardSplitTest.java:350)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
        at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:1084)
        at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:1055)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:80)
        at org.junit.rules.RunRules.evaluate(RunRules.java:20)
        at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at org.junit.rules.RunRules.evaluate(RunRules.java:20)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:80)
        at org.junit.rules.RunRules.evaluate(RunRules.java:20)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
        at org.junit.rules.RunRules.evaluate(RunRules.java:20)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
        at java.lang.Thread.run(Thread.java:750)
Caused by: java.net.SocketTimeoutException: Read timed out
        at java.net.SocketInputStream.socketRead0(Native Method)
        at java.net.SocketInputStream.socketRead(SocketInputStream.java:116)
        at java.net.SocketInputStream.read(SocketInputStream.java:171)
        at java.net.SocketInputStream.read(SocketInputStream.java:141)
        at 
org.apache.http.impl.io.SessionInputBufferImpl.streamRead(SessionInputBufferImpl.java:137)
        at 
org.apache.http.impl.io.SessionInputBufferImpl.fillBuffer(SessionInputBufferImpl.java:153)
        at 
org.apache.http.impl.io.SessionInputBufferImpl.readLine(SessionInputBufferImpl.java:280)
        at 
org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:138)
        at 
org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:56)
        at 
org.apache.http.impl.io.AbstractMessageParser.parse(AbstractMessageParser.java:259)
        at 
org.apache.http.impl.DefaultBHttpClientConnection.receiveResponseHeader(DefaultBHttpClientConnection.java:163)
        at 
org.apache.http.impl.conn.CPoolProxy.receiveResponseHeader(CPoolProxy.java:157)
        at 
org.apache.http.protocol.HttpRequestExecutor.doReceiveResponse(HttpRequestExecutor.java:273)
        at 
org.apache.http.protocol.HttpRequestExecutor.execute(HttpRequestExecutor.java:125)
        at 
org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:272)
        at 
org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:186)
        at org.apache.http.impl.execchain.RetryExec.execute(RetryExec.java:89)
        at 
org.apache.http.impl.execchain.RedirectExec.execute(RedirectExec.java:110)
        at 
org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:185)
        at 
org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:83)
        at 
org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:56)
        at 
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:571)
        ... 57 more




Build Log:
[...truncated 14292 lines...]
   [junit4] Suite: 
org.apache.solr.cloud.api.collections.HdfsCloudIncrementalBackupTest
   [junit4]   2> 565873 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.a.s.SolrTestCase Setting 'solr.default.confdir' system property to 
test-framework derived value of 
'/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/server/solr/configsets/_default/conf'
   [junit4]   2> 565873 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: 
test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> 565874 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.a.s.u.ErrorLogMuter Closing ErrorLogMuter-regex-172 after mutting 0 log 
messages
   [junit4]   2> 565874 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.a.s.u.ErrorLogMuter Creating ErrorLogMuter-regex-173 for ERROR logs matching 
regex: ignore_exception
   [junit4]   2> 565874 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.a.s.SolrTestCaseJ4 Created dataDir: 
/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/data-dir-86-001
   [junit4]   2> 565875 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) 
w/NUMERIC_DOCVALUES_SYSPROP=false
   [junit4]   2> 565875 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via: 
@org.apache.solr.util.RandomizeSSL(reason=, ssl=NaN, value=NaN, clientAuth=NaN)
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 567600 WARN  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.a.h.s.a.s.AuthenticationFilter Unable to initialize FileSignerSecretProvider, 
falling back to use random secrets. Reason: access denied 
("java.io.FilePermission" "/home/jenkins/hadoop-http-auth-signature-secret" 
"read")
   [junit4]   2> 567611 WARN  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 567644 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.e.j.s.Server jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 
27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 1.8.0_362-b09
   [junit4]   2> 567647 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 567647 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 567647 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 567650 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@55a64563{static,/static,jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.4-tests.jar!/webapps/static,AVAILABLE}
   [junit4]   2> 567852 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.w.WebAppContext@57fff38f{hdfs,/,file:///home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/jetty-localhost_localdomain-37109-hadoop-hdfs-3_2_4-tests_jar-_-any-5138460785843269654/webapp/,AVAILABLE}{jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.4-tests.jar!/webapps/hdfs}
   [junit4]   2> 567853 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.e.j.s.AbstractConnector Started ServerConnector@30ee72e0{HTTP/1.1, 
(http/1.1)}{localhost.localdomain:37109}
   [junit4]   2> 567853 INFO  
(SUITE-HdfsCloudIncrementalBackupTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.e.j.s.Server Started @567881ms
   [junit4]   2> 568707 WARN  (StorageLocationChecker thread 0) [     ] 
o.a.h.u.NativeCodeLoader Unable to load native-hadoop library for your 
platform... using builtin-java classes where applicable
   [junit4]   2> 568774 WARN  (Listener at localhost.localdomain/41515) [     ] 
o.a.h.s.a.s.AuthenticationFilter Unable to initialize FileSignerSecretProvider, 
falling back to use random secrets. Reason: access denied 
("java.io.FilePermission" "/home/jenkins/hadoop-http-auth-signature-secret" 
"read")
   [junit4]   2> 568774 WARN  (Listener at localhost.localdomain/41515) [     ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 568780 INFO  (Listener at localhost.localdomain/41515) [     ] 
o.e.j.s.Server jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 
27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 1.8.0_362-b09
   [junit4]   2> 568781 INFO  (Listener at localhost.localdomain/41515) [     ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 568781 INFO  (Listener at localhost.localdomain/41515) [     ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 568781 INFO  (Listener at localhost.localdomain/41515) [     ] 
o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 568781 INFO  (Listener at localhost.localdomain/41515) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@24e3b405{static,/static,jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.4-tests.jar!/webapps/static,AVAILABLE}
   [junit4]   2> 568902 INFO  (Listener at localhost.localdomain/41515) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.w.WebAppContext@faddfc2{datanode,/,file:///home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/jetty-localhost-43785-hadoop-hdfs-3_2_4-tests_jar-_-any-643071367416039975/webapp/,AVAILABLE}{jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.4-tests.jar!/webapps/datanode}
   [junit4]   2> 568902 INFO  (Listener at localhost.localdomain/41515) [     ] 
o.e.j.s.AbstractConnector Started ServerConnector@49ead34b{HTTP/1.1, 
(http/1.1)}{localhost:43785}
   [junit4]   2> 568902 INFO  (Listener at localhost.localdomain/41515) [     ] 
o.e.j.s.Server Started @568930ms
   [junit4]   2> 569491 WARN  (Listener at localhost.localdomain/45149) [     ] 
o.a.h.s.a.s.AuthenticationFilter Unable to initialize FileSignerSecretProvider, 
falling back to use random secrets. Reason: access denied 
("java.io.FilePermission" "/home/jenkins/hadoop-http-auth-signature-secret" 
"read")
   [junit4]   2> 569492 WARN  (Listener at localhost.localdomain/45149) [     ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 569493 INFO  (Listener at localhost.localdomain/45149) [     ] 
o.e.j.s.Server jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 
27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 1.8.0_362-b09
   [junit4]   2> 569500 INFO  (Listener at localhost.localdomain/45149) [     ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 569500 INFO  (Listener at localhost.localdomain/45149) [     ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 569500 INFO  (Listener at localhost.localdomain/45149) [     ] 
o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 569500 INFO  (Listener at localhost.localdomain/45149) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@1fdb6b45{static,/static,jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.4-tests.jar!/webapps/static,AVAILABLE}
   [junit4]   2> 569622 INFO  (Listener at localhost.localdomain/45149) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.w.WebAppContext@56894d49{datanode,/,file:///home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/jetty-localhost-45799-hadoop-hdfs-3_2_4-tests_jar-_-any-5425933224831959611/webapp/,AVAILABLE}{jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.4-tests.jar!/webapps/datanode}
   [junit4]   2> 569623 INFO  (Listener at localhost.localdomain/45149) [     ] 
o.e.j.s.AbstractConnector Started ServerConnector@74110532{HTTP/1.1, 
(http/1.1)}{localhost:45799}
   [junit4]   2> 569623 INFO  (Listener at localhost.localdomain/45149) [     ] 
o.e.j.s.Server Started @569651ms
   [junit4]   2> 570976 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xe2756d34f26365d4: Processing first storage report for 
DS-cc3a79b6-73df-42d0-8bdf-2f7676fe24f9 from datanode 
DatanodeRegistration(127.0.0.1:35283, 
datanodeUuid=d047c66a-5542-4881-bdc4-0cf1246c709b, infoPort=45735, 
infoSecurePort=0, ipcPort=37431, 
storageInfo=lv=-57;cid=testClusterID;nsid=1676083662;c=1707692208827)
   [junit4]   2> 570978 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xe2756d34f26365d4: from storage 
DS-cc3a79b6-73df-42d0-8bdf-2f7676fe24f9 node 
DatanodeRegistration(127.0.0.1:35283, 
datanodeUuid=d047c66a-5542-4881-bdc4-0cf1246c709b, infoPort=45735, 
infoSecurePort=0, ipcPort=37431, 
storageInfo=lv=-57;cid=testClusterID;nsid=1676083662;c=1707692208827), blocks: 
0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0
   [junit4]   2> 570978 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xe2756d34f26365d4: Processing first storage report for 
DS-27bfdf39-3fb3-43e2-b473-13319e9a8702 from datanode 
DatanodeRegistration(127.0.0.1:35283, 
datanodeUuid=d047c66a-5542-4881-bdc4-0cf1246c709b, infoPort=45735, 
infoSecurePort=0, ipcPort=37431, 
storageInfo=lv=-57;cid=testClusterID;nsid=1676083662;c=1707692208827)
   [junit4]   2> 570978 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xe2756d34f26365d4: from storage 
DS-27bfdf39-3fb3-43e2-b473-13319e9a8702 node 
DatanodeRegistration(127.0.0.1:35283, 
datanodeUuid=d047c66a-5542-4881-bdc4-0cf1246c709b, infoPort=45735, 
infoSecurePort=0, ipcPort=37431, 
storageInfo=lv=-57;cid=testClusterID;nsid=1676083662;c=1707692208827), blocks: 
0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0
   [junit4]   2> 570987 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xf1d1909aaaee23d6: Processing first storage report for 
DS-92c3b1f0-e656-4908-915f-e40f928006c9 from datanode 
DatanodeRegistration(127.0.0.1:43755, 
datanodeUuid=9afe006a-1d94-4d73-afff-47025b1ec31f, infoPort=46517, 
infoSecurePort=0, ipcPort=45149, 
storageInfo=lv=-57;cid=testClusterID;nsid=1676083662;c=1707692208827)
   [junit4]   2> 570987 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xf1d1909aaaee23d6: from storage 
DS-92c3b1f0-e656-4908-915f-e40f928006c9 node 
DatanodeRegistration(127.0.0.1:43755, 
datanodeUuid=9afe006a-1d94-4d73-afff-47025b1ec31f, infoPort=46517, 
infoSecurePort=0, ipcPort=45149, 
storageInfo=lv=-57;cid=testClusterID;nsid=1676083662;c=1707692208827), blocks: 
0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0
   [junit4]   2> 570987 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xf1d1909aaaee23d6: Processing first storage report for 
DS-6a26b5cb-03d7-4502-a504-ad92fbf548f0 from datanode 
DatanodeRegistration(127.0.0.1:43755, 
datanodeUuid=9afe006a-1d94-4d73-afff-47025b1ec31f, infoPort=46517, 
infoSecurePort=0, ipcPort=45149, 
storageInfo=lv=-57;cid=testClusterID;nsid=1676083662;c=1707692208827)
   [junit4]   2> 570987 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xf1d1909aaaee23d6: from storage 
DS-6a26b5cb-03d7-4502-a504-ad92fbf548f0 node 
DatanodeRegistration(127.0.0.1:43755, 
datanodeUuid=9afe006a-1d94-4d73-afff-47025b1ec31f, infoPort=46517, 
infoSecurePort=0, ipcPort=45149, 
storageInfo=lv=-57;cid=testClusterID;nsid=1676083662;c=1707692208827), blocks: 
0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0
   [junit4]   2> 571618 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.MiniSolrCloudCluster Starting cluster of 2 servers in 
/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/tempDir-002
   [junit4]   2> 571618 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 571621 INFO  (ZkTestServer Run Thread) [     ] 
o.a.s.c.ZkTestServer client port: 0.0.0.0/0.0.0.0:0
   [junit4]   2> 571621 INFO  (ZkTestServer Run Thread) [     ] 
o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 571622 WARN  (ZkTestServer Run Thread) [     ] 
o.a.z.s.ServerCnxnFactory maxCnxns is not configured, using default value 0.
   [junit4]   2> 571721 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.ZkTestServer start zk server on port: 38011
   [junit4]   2> 571721 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.ZkTestServer waitForServerUp: 127.0.0.1:38011
   [junit4]   2> 571721 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.ZkTestServer parse host and port list: 127.0.0.1:38011
   [junit4]   2> 571721 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.ZkTestServer connecting to 127.0.0.1 38011
   [junit4]   2> 571735 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 571738 INFO  (zkConnectionManagerCallback-6890-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 571738 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 571744 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 571747 INFO  (zkConnectionManagerCallback-6892-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 571747 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 571767 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 571775 INFO  (zkConnectionManagerCallback-6894-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 571775 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 571884 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.c.s.e.JettySolrRunner Start Jetty (configured port=0, binding port=0)
   [junit4]   2> 571884 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.c.s.e.JettySolrRunner Trying to start Jetty on port 0 try number 2 ...
   [junit4]   2> 571884 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.e.j.s.Server jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 
27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 1.8.0_362-b09
   [junit4]   2> 571884 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.c.s.e.JettySolrRunner Start Jetty (configured port=0, binding port=0)
   [junit4]   2> 571884 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.c.s.e.JettySolrRunner Trying to start Jetty on port 0 try number 2 ...
   [junit4]   2> 571885 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.e.j.s.Server jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 
27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 1.8.0_362-b09
   [junit4]   2> 571892 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 571892 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 571892 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 571900 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@557fabb5{/solr,null,AVAILABLE}
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.e.j.s.AbstractConnector Started ServerConnector@32a9ebc9{HTTP/1.1, (http/1.1, 
h2c)}{127.0.0.1:46335}
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.e.j.s.Server Started @571951ms
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, 
zkHost=127.0.0.1:38011/solr, hostPort=46335}
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 571923 ERROR (jetty-launcher-6895-thread-1) [     ] 
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter Using logger factory 
org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 
8.11.4
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: 
/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2024-02-11T22:56:53.727Z
   [junit4]   2> 571923 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@4de0756d{/solr,null,AVAILABLE}
   [junit4]   2> 571924 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.e.j.s.AbstractConnector Started ServerConnector@47ccb864{HTTP/1.1, (http/1.1, 
h2c)}{127.0.0.1:41643}
   [junit4]   2> 571924 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.e.j.s.Server Started @571952ms
   [junit4]   2> 571924 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, 
zkHost=127.0.0.1:38011/solr, hostPort=41643}
   [junit4]   2> 571924 ERROR (jetty-launcher-6895-thread-2) [     ] 
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 571924 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter Using logger factory 
org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 571924 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 
8.11.4
   [junit4]   2> 571924 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 571924 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: 
/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr
   [junit4]   2> 571924 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2024-02-11T22:56:53.728Z
   [junit4]   2> 571929 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 571929 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 571932 INFO  (zkConnectionManagerCallback-6897-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 571932 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 571932 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 571953 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 571954 INFO  (zkConnectionManagerCallback-6899-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 571954 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 571955 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 571955 WARN  
(jetty-launcher-6895-thread-1-SendThread(127.0.0.1:38011)) [     ] 
o.a.z.ClientCnxn An exception was thrown while closing send thread for session 
0x1006d6130130003.
   [junit4]   2>           => EndOfStreamException: Unable to read additional 
data from server sessionid 0x1006d6130130003, likely server has closed socket
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77)
   [junit4]   2> org.apache.zookeeper.ClientCnxn$EndOfStreamException: Unable 
to read additional data from server sessionid 0x1006d6130130003, likely server 
has closed socket
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77) 
~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350)
 ~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1275) 
[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 571958 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 572064 WARN  
(jetty-launcher-6895-thread-2-SendThread(127.0.0.1:38011)) [     ] 
o.a.z.ClientCnxn An exception was thrown while closing send thread for session 
0x1006d6130130004.
   [junit4]   2>           => EndOfStreamException: Unable to read additional 
data from server sessionid 0x1006d6130130004, likely server has closed socket
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77)
   [junit4]   2> org.apache.zookeeper.ClientCnxn$EndOfStreamException: Unable 
to read additional data from server sessionid 0x1006d6130130004, likely server 
has closed socket
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77) 
~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350)
 ~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1275) 
[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 574925 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.h.c.HttpShardHandlerFactory Host whitelist initialized: 
WhitelistHostChecker [whitelistHosts=null, whitelistHostCheckingEnabled=true]
   [junit4]   2> 574926 WARN  (jetty-launcher-6895-thread-1) [     ] 
o.e.j.u.s.S.config Trusting all certificates configured for 
Client@29ca6b47[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 574926 WARN  (jetty-launcher-6895-thread-1) [     ] 
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for 
Client@29ca6b47[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 574944 WARN  (jetty-launcher-6895-thread-1) [     ] 
o.e.j.u.s.S.config Trusting all certificates configured for 
Client@205c5e36[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 574944 WARN  (jetty-launcher-6895-thread-1) [     ] 
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for 
Client@205c5e36[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 574945 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:38011/solr
   [junit4]   2> 574949 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.h.c.HttpShardHandlerFactory Host whitelist initialized: 
WhitelistHostChecker [whitelistHosts=null, whitelistHostCheckingEnabled=true]
   [junit4]   2> 574955 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 574958 WARN  (jetty-launcher-6895-thread-2) [     ] 
o.e.j.u.s.S.config Trusting all certificates configured for 
Client@7830cacf[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 574958 WARN  (jetty-launcher-6895-thread-2) [     ] 
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for 
Client@7830cacf[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 574961 WARN  (jetty-launcher-6895-thread-2) [     ] 
o.e.j.u.s.S.config Trusting all certificates configured for 
Client@39ad0623[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 574961 WARN  (jetty-launcher-6895-thread-2) [     ] 
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for 
Client@39ad0623[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 574963 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:38011/solr
   [junit4]   2> 574971 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 574972 INFO  (zkConnectionManagerCallback-6913-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 574972 INFO  (jetty-launcher-6895-thread-1) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 574972 INFO  (zkConnectionManagerCallback-6921-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 574972 INFO  (jetty-launcher-6895-thread-2) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 575074 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.c.ConnectionManager Waiting for client to 
connect to ZooKeeper
   [junit4]   2> 575075 INFO  (zkConnectionManagerCallback-6923-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 575075 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.c.ConnectionManager Client is connected 
to ZooKeeper
   [junit4]   2> 575174 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.c.c.ConnectionManager Waiting for client to 
connect to ZooKeeper
   [junit4]   2> 575176 INFO  (zkConnectionManagerCallback-6927-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 575176 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.c.c.ConnectionManager Client is connected 
to ZooKeeper
   [junit4]   2> 575182 WARN  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.ZkController Contents of zookeeper 
/security.json are world-readable; consider setting up ACLs as described in 
https://solr.apache.org/guide/zookeeper-access-control.html
   [junit4]   2> 575188 WARN  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.c.ZkController Contents of zookeeper 
/security.json are world-readable; consider setting up ACLs as described in 
https://solr.apache.org/guide/zookeeper-access-control.html
   [junit4]   2> 575190 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.OverseerElectionContext I am going to be 
the leader 127.0.0.1:46335_solr
   [junit4]   2> 575191 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.Overseer Overseer 
(id=72177858223734791-127.0.0.1:46335_solr-n_0000000000) starting
   [junit4]   2> 575195 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:41643_solr
   [junit4]   2> 575197 INFO  (zkCallback-6922-thread-1) [     ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 575198 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.c.ZkController non-data nodes now []
   [junit4]   2> 575198 INFO  
(OverseerStateUpdate-72177858223734791-127.0.0.1:46335_solr-n_0000000000) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.Overseer Starting to work on the main 
queue : 127.0.0.1:46335_solr
   [junit4]   2> 575198 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.ZkController Publish 
node=127.0.0.1:46335_solr as DOWN
   [junit4]   2> 575198 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.TransientSolrCoreCacheDefault Allocating 
transient core cache for max 2147483647 cores with initial capacity of 1024
   [junit4]   2> 575198 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:46335_solr
   [junit4]   2> 575201 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.p.PackageLoader /packages.json updated to 
version -1
   [junit4]   2> 575202 WARN  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.c.CoreContainer Not all security plugins 
configured!  authentication=disabled authorization=disabled.  Solr is only as 
secure as you make it. Consider configuring authentication/authorization before 
exposing Solr to users internal or external.  See 
https://s.apache.org/solrsecurity for more info
   [junit4]   2> 575202 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup 
repository with configuration params {type = repository,name = 
trackingBackupRepository,class = 
org.apache.solr.core.TrackingBackupRepository,attributes = {default=true, 
name=trackingBackupRepository, 
class=org.apache.solr.core.TrackingBackupRepository},args = 
{delegateRepoName=hdfs}}
   [junit4]   2> 575202 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup 
repository with configuration params {type = repository,name = hdfs,class = 
org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes = 
{name=hdfs, 
class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args = 
{location=/backup,solr.hdfs.home=hdfs://localhost.localdomain:41515/solr,solr.hdfs.confdir=}}
   [junit4]   2> 575202 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Default 
configuration for backup repository is with configuration params {type = 
repository,name = trackingBackupRepository,class = 
org.apache.solr.core.TrackingBackupRepository,attributes = {default=true, 
name=trackingBackupRepository, 
class=org.apache.solr.core.TrackingBackupRepository},args = 
{delegateRepoName=hdfs}}
   [junit4]   2> 575207 INFO  
(OverseerStateUpdate-72177858223734791-127.0.0.1:46335_solr-n_0000000000) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.c.ZkStateReader Updated live nodes from 
ZooKeeper... (1) -> (2)
   [junit4]   2> 575207 INFO  (zkCallback-6926-thread-1) [     ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 575211 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.ZkController non-data nodes now []
   [junit4]   2> 575216 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.p.PackageLoader /packages.json updated to 
version -1
   [junit4]   2> 575216 WARN  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.CoreContainer Not all security plugins 
configured!  authentication=disabled authorization=disabled.  Solr is only as 
secure as you make it. Consider configuring authentication/authorization before 
exposing Solr to users internal or external.  See 
https://s.apache.org/solrsecurity for more info
   [junit4]   2> 575216 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup 
repository with configuration params {type = repository,name = 
trackingBackupRepository,class = 
org.apache.solr.core.TrackingBackupRepository,attributes = {default=true, 
name=trackingBackupRepository, 
class=org.apache.solr.core.TrackingBackupRepository},args = 
{delegateRepoName=hdfs}}
   [junit4]   2> 575216 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup 
repository with configuration params {type = repository,name = hdfs,class = 
org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes = 
{name=hdfs, 
class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args = 
{location=/backup,solr.hdfs.home=hdfs://localhost.localdomain:41515/solr,solr.hdfs.confdir=}}
   [junit4]   2> 575216 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Default 
configuration for backup repository is with configuration params {type = 
repository,name = trackingBackupRepository,class = 
org.apache.solr.core.TrackingBackupRepository,attributes = {default=true, 
name=trackingBackupRepository, 
class=org.apache.solr.core.TrackingBackupRepository},args = 
{delegateRepoName=hdfs}}
   [junit4]   2> 575218 INFO  (zkCallback-6926-thread-2) [     ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 575234 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.h.a.MetricsHistoryHandler No .system 
collection, keeping metrics history in memory.
   [junit4]   2> 575248 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.h.a.MetricsHistoryHandler No .system 
collection, keeping metrics history in memory.
   [junit4]   2> 575259 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.node' (registry 'solr.node') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93
   [junit4]   2> 575271 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jvm' (registry 'solr.jvm') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93
   [junit4]   2> 575272 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jetty' (registry 'solr.jetty') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93
   [junit4]   2> 575273 INFO  (jetty-launcher-6895-thread-2) 
[n:127.0.0.1:41643_solr     ] o.a.s.c.CorePropertiesLocator Found 0 core 
definitions underneath 
/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/tempDir-002/node2
   [junit4]   2> 575276 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.node' (registry 'solr.node') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93
   [junit4]   2> 575285 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jvm' (registry 'solr.jvm') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93
   [junit4]   2> 575286 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jetty' (registry 'solr.jetty') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93
   [junit4]   2> 575287 INFO  (jetty-launcher-6895-thread-1) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.CorePropertiesLocator Found 0 core 
definitions underneath 
/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/tempDir-002/node1
   [junit4]   2> 575318 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.MiniSolrCloudCluster waitForAllNodes: numServers=2
   [junit4]   2> 575319 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 575320 INFO  (zkConnectionManagerCallback-6950-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 575320 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 575322 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 575323 INFO  (Listener at localhost.localdomain/37431) [     ] 
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38011/solr ready
   [junit4]   2> 575376 INFO  
(TEST-HdfsCloudIncrementalBackupTest.testRestoreToOriginalCollection-seed#[C0258A7B5C6D0A61])
 [     ] o.a.s.SolrTestCaseJ4 ###Starting testRestoreToOriginalCollection
   [junit4]   2> 575379 INFO  (qtp1225482037-10043) [n:127.0.0.1:41643_solr     
] o.a.s.s.HttpSolrCall 
HttpSolrCall.init(http://127.0.0.1:41643/solr/admin/collections?action=CREATE&name=hdfsbackuprestore_testbackuprestoretooriginal&collection.configName=conf1&numShards=2&maxShardsPerNode=-1&nrtReplicas=2&wt=javabin&version=2)
   [junit4]   2> 575383 INFO  
(OverseerThreadFactory-6934-thread-1-processing-n:127.0.0.1:46335_solr) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.a.c.CreateCollectionCmd Create collection 
hdfsbackuprestore_testbackuprestoretooriginal
   [junit4]   2> 575504 INFO  
(OverseerStateUpdate-72177858223734791-127.0.0.1:46335_solr-n_0000000000) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore_testbackuprestoretooriginal",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   
"core":"hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "node_name":"127.0.0.1:41643_solr",
   [junit4]   2>   "base_url":"http://127.0.0.1:41643/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 575514 INFO  
(OverseerStateUpdate-72177858223734791-127.0.0.1:46335_solr-n_0000000000) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore_testbackuprestoretooriginal",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   
"core":"hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "node_name":"127.0.0.1:46335_solr",
   [junit4]   2>   "base_url":"http://127.0.0.1:46335/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 575518 INFO  
(OverseerStateUpdate-72177858223734791-127.0.0.1:46335_solr-n_0000000000) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore_testbackuprestoretooriginal",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   
"core":"hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "node_name":"127.0.0.1:41643_solr",
   [junit4]   2>   "base_url":"http://127.0.0.1:41643/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 575522 INFO  
(OverseerStateUpdate-72177858223734791-127.0.0.1:46335_solr-n_0000000000) 
[n:127.0.0.1:46335_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore_testbackuprestoretooriginal",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   
"core":"hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "node_name":"127.0.0.1:46335_solr",
   [junit4]   2>   "base_url":"http://127.0.0.1:46335/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 576012 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr     
] o.a.s.s.HttpSolrCall 
HttpSolrCall.init(http://127.0.0.1:41643/solr/admin/cores?null)
   [junit4]   2> 576012 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr    
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1&action=CREATE&numShards=2&collection=hdfsbackuprestore_testbackuprestoretooriginal&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 576013 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr    
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient core cache for max 
2147483647 cores with initial capacity of 1024
   [junit4]   2> 576014 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr     
] o.a.s.s.HttpSolrCall 
HttpSolrCall.init(http://127.0.0.1:46335/solr/admin/cores?null)
   [junit4]   2> 576014 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr    
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2&action=CREATE&numShards=2&collection=hdfsbackuprestore_testbackuprestoretooriginal&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 576022 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr     
] o.a.s.s.HttpSolrCall 
HttpSolrCall.init(http://127.0.0.1:46335/solr/admin/cores?null)
   [junit4]   2> 576023 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr    
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6&action=CREATE&numShards=2&collection=hdfsbackuprestore_testbackuprestoretooriginal&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 576023 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr     
] o.a.s.s.HttpSolrCall 
HttpSolrCall.init(http://127.0.0.1:41643/solr/admin/cores?null)
   [junit4]   2> 576023 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr    
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4&action=CREATE&numShards=2&collection=hdfsbackuprestore_testbackuprestoretooriginal&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 576043 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.11.4
   [junit4]   2> 576045 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.11.4
   [junit4]   2> 576046 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.11.4
   [junit4]   2> 576046 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.s.IndexSchema Schema name=minimal
   [junit4]   2> 576052 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 576052 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.c.CoreContainer Creating SolrCore 
'hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2' using 
configuration from configset conf1, trusted=true
   [junit4]   2> 576052 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.s.IndexSchema Schema name=minimal
   [junit4]   2> 576052 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 576052 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.11.4
   [junit4]   2> 576052 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.CoreContainer Creating SolrCore 
'hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6' using 
configuration from configset conf1, trusted=true
   [junit4]   2> 576053 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.core.hdfsbackuprestore_testbackuprestoretooriginal.shard2.replica_n6' 
(registry 
'solr.core.hdfsbackuprestore_testbackuprestoretooriginal.shard2.replica_n6') 
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93
   [junit4]   2> 576053 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.core.hdfsbackuprestore_testbackuprestoretooriginal.shard1.replica_n2' 
(registry 
'solr.core.hdfsbackuprestore_testbackuprestoretooriginal.shard1.replica_n2') 
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93
   [junit4]   2> 576053 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.SolrCore 
[[hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6] ] Opening 
new SolrCore at 
[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/tempDir-002/node1/hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6],
 
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/tempDir-002/node1/hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6/data/]
   [junit4]   2> 576053 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.c.SolrCore 
[[hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2] ] Opening 
new SolrCore at 
[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/tempDir-002/node1/hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2],
 
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/tempDir-002/node1/hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2/data/]
   [junit4]   2> 576053 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.s.IndexSchema Schema name=minimal
   [junit4]   2> 576053 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 576053 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.CoreContainer Creating SolrCore 
'hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1' using 
configuration from configset conf1, trusted=true
   [junit4]   2> 576054 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.core.hdfsbackuprestore_testbackuprestoretooriginal.shard1.replica_n1' 
(registry 
'solr.core.hdfsbackuprestore_testbackuprestoretooriginal.shard1.replica_n1') 
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93
   [junit4]   2> 576054 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.SolrCore 
[[hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1] ] Opening 
new SolrCore at 
[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/tempDir-002/node2/hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1],
 
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/tempDir-002/node2/hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1/data/]
   [junit4]   2> 576056 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.s.IndexSchema Schema name=minimal
   [junit4]   2> 576056 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 576056 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.c.CoreContainer Creating SolrCore 
'hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4' using 
configuration from configset conf1, trusted=true
   [junit4]   2> 576057 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.core.hdfsbackuprestore_testbackuprestoretooriginal.shard2.replica_n4' 
(registry 
'solr.core.hdfsbackuprestore_testbackuprestoretooriginal.shard2.replica_n4') 
enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@4d499d93
   [junit4]   2> 576057 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.c.SolrCore 
[[hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4] ] Opening 
new SolrCore at 
[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/tempDir-002/node2/hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4],
 
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.HdfsCloudIncrementalBackupTest_C0258A7B5C6D0A61-001/tempDir-002/node2/hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4/data/]
   [junit4]   2> 576184 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.u.UpdateHandler Using UpdateLog implementation: 
org.apache.solr.update.UpdateLog
   [junit4]   2> 576185 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH 
numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 576186 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 576186 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 576208 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.u.UpdateHandler Using UpdateLog implementation: 
org.apache.solr.update.UpdateLog
   [junit4]   2> 576208 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH 
numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 576209 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 576209 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 576210 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 576210 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 576211 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.h.ReplicationHandler Commits will be reserved for 10000 ms
   [junit4]   2> 576211 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.u.UpdateLog Could not find max version in index or recent updates, using 
new clock 1790645075197296640
   [junit4]   2> 576214 INFO  
(searcherExecutor-6956-thread-1-processing-n:127.0.0.1:41643_solr 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3) 
[n:127.0.0.1:41643_solr c:hdfsbackuprestore_testbackuprestoretooriginal 
s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.SolrCore 
[hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1]  Registered 
new searcher autowarm time: 0 ms
   [junit4]   2> 576215 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.u.UpdateHandler Using UpdateLog implementation: 
org.apache.solr.update.UpdateLog
   [junit4]   2> 576215 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH 
numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 576216 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 576216 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 576218 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 576218 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 576219 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.h.ReplicationHandler Commits will be reserved for 10000 ms
   [junit4]   2> 576219 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.u.UpdateLog Could not find max version in index or recent updates, using 
new clock 1790645075205685248
   [junit4]   2> 576219 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.ZkShardTerms Successful update of terms at 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/terms/shard1 to 
Terms{values={core_node3=0}, version=0}
   [junit4]   2> 576219 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/leaders/shard1
   [junit4]   2> 576220 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 576220 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 576221 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.h.ReplicationHandler Commits will be reserved for 10000 ms
   [junit4]   2> 576221 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.u.UpdateLog Could not find max version in index or recent updates, using 
new clock 1790645075207782400
   [junit4]   2> 576225 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for 
shard shard1: total=2 found=1 timeoute in=9998ms
   [junit4]   2> 576228 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.u.UpdateHandler Using UpdateLog implementation: 
org.apache.solr.update.UpdateLog
   [junit4]   2> 576228 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH 
numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 576229 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 576229 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 576232 INFO  
(searcherExecutor-6958-thread-1-processing-n:127.0.0.1:41643_solr 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7) 
[n:127.0.0.1:41643_solr c:hdfsbackuprestore_testbackuprestoretooriginal 
s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.c.SolrCore 
[hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4]  Registered 
new searcher autowarm time: 0 ms
   [junit4]   2> 576233 INFO  
(searcherExecutor-6954-thread-1-processing-n:127.0.0.1:46335_solr 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8) 
[n:127.0.0.1:46335_solr c:hdfsbackuprestore_testbackuprestoretooriginal 
s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.SolrCore 
[hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6]  Registered 
new searcher autowarm time: 0 ms
   [junit4]   2> 576233 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: 
/configs/conf1
   [junit4]   2> 576234 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 576237 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.h.ReplicationHandler Commits will be reserved for 10000 ms
   [junit4]   2> 576237 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.u.UpdateLog Could not find max version in index or recent updates, using 
new clock 1790645075224559616
   [junit4]   2> 576239 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.ZkShardTerms Successful update of terms at 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/terms/shard2 to 
Terms{values={core_node8=0}, version=0}
   [junit4]   2> 576239 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/leaders/shard2
   [junit4]   2> 576240 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.c.ZkShardTerms Successful update of terms at 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/terms/shard2 to 
Terms{values={core_node7=0, core_node8=0}, version=1}
   [junit4]   2> 576240 INFO  (qtp1225482037-10042) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/leaders/shard2
   [junit4]   2> 576241 INFO  
(searcherExecutor-6952-thread-1-processing-n:127.0.0.1:46335_solr 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5) 
[n:127.0.0.1:46335_solr c:hdfsbackuprestore_testbackuprestoretooriginal 
s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.c.SolrCore 
[hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2]  Registered 
new searcher autowarm time: 0 ms
   [junit4]   2> 576246 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 576247 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.c.ZkShardTerms Successful update of terms at 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/terms/shard1 to 
Terms{values={core_node3=0, core_node5=0}, version=1}
   [junit4]   2> 576247 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 576247 INFO  (qtp1013846580-10034) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/leaders/shard1
   [junit4]   2> 576247 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.SyncStrategy Sync replicas to 
http://127.0.0.1:46335/solr/hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6/
   [junit4]   2> 576247 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.u.PeerSync PeerSync: 
core=hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 
url=http://127.0.0.1:46335/solr  START 
replicas=[http://127.0.0.1:41643/solr/hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4/]
 nUpdates=100
   [junit4]   2> 576248 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.u.PeerSync PeerSync: 
core=hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 
url=http://127.0.0.1:46335/solr  DONE. We have no versions. sync failed.
   [junit4]   2> 576249 INFO  (qtp1225482037-10046) [n:127.0.0.1:41643_solr     
] o.a.s.s.HttpSolrCall 
HttpSolrCall.init(http://127.0.0.1:41643/solr/hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4/get?null)
   [junit4]   2> 576250 INFO  (qtp1225482037-10046) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node7 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4 ] 
o.a.s.c.S.Request 
[hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n4]  webapp=/solr 
path=/get 
params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2}
 status=0 QTime=0
   [junit4]   2> 576250 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.SyncStrategy Leader's attempt to sync with shard failed, moving to the 
next candidate
   [junit4]   2> 576251 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.ShardLeaderElectionContext We failed sync, but we have no versions - we 
can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 576251 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.ShardLeaderElectionContextBase Creating leader registration node 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/leaders/shard2/leader
 after winning as 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/leader_elect/shard2/election/72177858223734791-core_node8-n_0000000000
   [junit4]   2> 576253 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.ShardLeaderElectionContext I am the new leader: 
http://127.0.0.1:46335/solr/hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6/
 shard2
   [junit4]   2> 576355 INFO  (zkCallback-6922-thread-1) [     ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore_testbackuprestoretooriginal/state.json] for 
collection [hdfsbackuprestore_testbackuprestoretooriginal] has occurred - 
updating... (live nodes size: [2])
   [junit4]   2> 576356 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard2 r:core_node8 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6 ] 
o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 576360 INFO  (qtp1013846580-10035) [n:127.0.0.1:46335_solr     
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores 
params={qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_testbackuprestoretooriginal_shard2_replica_n6&action=CREATE&numShards=2&collection=hdfsbackuprestore_testbackuprestoretooriginal&shard=shard2&wt=javabin&version=2&replicaType=NRT}
 status=0 QTime=337
   [junit4]   2> 576459 INFO  (zkCallback-6922-thread-1) [     ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore_testbackuprestoretooriginal/state.json] for 
collection [hdfsbackuprestore_testbackuprestoretooriginal] has occurred - 
updating... (live nodes size: [2])
   [junit4]   2> 576459 INFO  (zkCallback-6922-thread-2) [     ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore_testbackuprestoretooriginal/state.json] for 
collection [hdfsbackuprestore_testbackuprestoretooriginal] has occurred - 
updating... (live nodes size: [2])
   [junit4]   2> 576726 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 576727 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 576727 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.SyncStrategy Sync replicas to 
http://127.0.0.1:41643/solr/hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1/
   [junit4]   2> 576727 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.u.PeerSync PeerSync: 
core=hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 
url=http://127.0.0.1:41643/solr  START 
replicas=[http://127.0.0.1:46335/solr/hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2/]
 nUpdates=100
   [junit4]   2> 576728 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.u.PeerSync PeerSync: 
core=hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 
url=http://127.0.0.1:41643/solr  DONE. We have no versions. sync failed.
   [junit4]   2> 576729 INFO  (qtp1013846580-10033) [n:127.0.0.1:46335_solr     
] o.a.s.s.HttpSolrCall 
HttpSolrCall.init(http://127.0.0.1:46335/solr/hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2/get?null)
   [junit4]   2> 576730 INFO  (qtp1013846580-10033) [n:127.0.0.1:46335_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node5 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2 ] 
o.a.s.c.S.Request 
[hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n2]  webapp=/solr 
path=/get 
params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2}
 status=0 QTime=1
   [junit4]   2> 576730 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.SyncStrategy Leader's attempt to sync with shard failed, moving to the 
next candidate
   [junit4]   2> 576730 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.ShardLeaderElectionContext We failed sync, but we have no versions - we 
can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 576730 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.ShardLeaderElectionContextBase Creating leader registration node 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/leaders/shard1/leader
 after winning as 
/collections/hdfsbackuprestore_testbackuprestoretooriginal/leader_elect/shard1/election/72177858223734792-core_node3-n_0000000000
   [junit4]   2> 576732 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.ShardLeaderElectionContext I am the new leader: 
http://127.0.0.1:41643/solr/hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1/
 shard1
   [junit4]   2> 576834 INFO  (zkCallback-6926-thread-2) [     ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore_testbackuprestoretooriginal/state.json] for 
collection [hdfsbackuprestore_testbackuprestoretooriginal] has occurred - 
updating... (live nodes size: [2])
   [junit4]   2> 576834 INFO  (zkCallback-6922-thread-1) [     ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore_testbackuprestoretooriginal/state.json] for 
collection [hdfsbackuprestore_testbackuprestoretooriginal] has occurred - 
updating... (live nodes size: [2])
   [junit4]   2> 576834 INFO  (zkCallback-6922-thread-2) [     ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore_testbackuprestoretooriginal/state.json] for 
collection [hdfsbackuprestore_testbackuprestoretooriginal] has occurred - 
updating... (live nodes size: [2])
   [junit4]   2> 576835 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr 
c:hdfsbackuprestore_testbackuprestoretooriginal s:shard1 r:core_node3 
x:hdfsbackuprestore_testbackuprestoretooriginal_shard1_replica_n1 ] 
o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 576838 INFO  (qtp1225482037-10045) [n:127.0.0.1:41643_solr     
] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores 
params={qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=hdfsbackuprestore_t

[...truncated too long message...]

ed type:None path:null
   [junit4]   2> 1407926 DEBUG 
(TEST-ShardSplitTest.testSplitStaticIndexReplicationLink-seed#[C0258A7B5C6D0A61]-EventThread)
 [     ] o.a.s.c.c.SolrZkClient Submitting job to respond to event WatchedEvent 
state:Closed type:None path:null
   [junit4]   2> 1407930 INFO  (closeThreadPool-12289-thread-7) [     ] 
o.e.j.s.AbstractConnector Stopped ServerConnector@4525cfca{HTTP/1.1, (http/1.1, 
h2c)}{127.0.0.1:46147}
   [junit4]   2> 1407930 INFO  (closeThreadPool-12289-thread-7) [     ] 
o.e.j.s.h.ContextHandler Stopped 
o.e.j.s.ServletContextHandler@56b6eca7{/onek/c,null,STOPPED}
   [junit4]   2> 1407930 INFO  (closeThreadPool-12289-thread-7) [     ] 
o.e.j.s.session node0 Stopped scavenging
   [junit4]   2> 1407931 INFO  
(TEST-ShardSplitTest.testSplitStaticIndexReplicationLink-seed#[C0258A7B5C6D0A61])
 [     ] o.a.s.u.ErrorLogMuter Closing ErrorLogMuter-regex-339 after mutting 0 
log messages
   [junit4]   2> 1407931 INFO  
(TEST-ShardSplitTest.testSplitStaticIndexReplicationLink-seed#[C0258A7B5C6D0A61])
 [     ] o.a.s.u.ErrorLogMuter Creating ErrorLogMuter-regex-340 for ERROR logs 
matching regex: ignore_exception
   [junit4]   2> 1407932 INFO  
(TEST-ShardSplitTest.testSplitStaticIndexReplicationLink-seed#[C0258A7B5C6D0A61])
 [     ] o.a.s.c.ZkTestServer Shutting down ZkTestServer.
   [junit4]   2> 1407933 WARN  
(TEST-ShardSplitTest.testSplitStaticIndexReplicationLink-seed#[C0258A7B5C6D0A61]-SendThread(127.0.0.1:44979))
 [     ] o.a.z.ClientCnxn An exception was thrown while closing send thread for 
session 0x1006d6d7d9d0000.
   [junit4]   2>           => EndOfStreamException: Unable to read additional 
data from server sessionid 0x1006d6d7d9d0000, likely server has closed socket
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77)
   [junit4]   2> org.apache.zookeeper.ClientCnxn$EndOfStreamException: Unable 
to read additional data from server sessionid 0x1006d6d7d9d0000, likely server 
has closed socket
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:77) 
~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350)
 ~[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2>        at 
org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1275) 
[zookeeper-3.6.2.jar:3.6.2]
   [junit4]   2> 1408033 DEBUG 
(TEST-ShardSplitTest.testSplitStaticIndexReplicationLink-seed#[C0258A7B5C6D0A61]-EventThread)
 [     ] o.a.s.c.c.SolrZkClient Submitting job to respond to event WatchedEvent 
state:Closed type:None path:null
   [junit4]   2> 1408134 DEBUG 
(TEST-ShardSplitTest.testSplitStaticIndexReplicationLink-seed#[C0258A7B5C6D0A61]-EventThread)
 [     ] o.a.s.c.c.SolrZkClient Submitting job to respond to event WatchedEvent 
state:Closed type:None path:null
   [junit4]   2> 1408135 WARN  (ZkTestServer Run Thread) [     ] 
o.a.s.c.ZkTestServer Watch limit violations: 
   [junit4]   2> Maximum concurrent create/delete watches above limit:
   [junit4]   2> 
   [junit4]   2>        11      
/solr/collections/testSplitStaticIndexReplication_link/terms/shard1_0
   [junit4]   2>        9       /solr/aliases.json
   [junit4]   2>        6       /solr/packages.json
   [junit4]   2>        6       /solr/security.json
   [junit4]   2>        6       /solr/configs/conf1
   [junit4]   2>        6       /solr/collections/collection1/terms/shard1
   [junit4]   2>        5       
/solr/collections/collection1/collectionprops.json
   [junit4]   2>        5       /solr/collections/collection1/terms/shard2
   [junit4]   2>        4       
/solr/collections/testSplitStaticIndexReplication_link/terms/shard1
   [junit4]   2>        4       
/solr/collections/testSplitStaticIndexReplication_link/terms/shard1_1
   [junit4]   2>        3       
/solr/collections/testSplitStaticIndexReplication_link/collectionprops.json
   [junit4]   2>        3       
/solr/collections/control_collection/terms/shard1
   [junit4]   2> 
   [junit4]   2> Maximum concurrent data watches above limit:
   [junit4]   2> 
   [junit4]   2>        70      /solr/collections/collection1/state.json
   [junit4]   2>        13      
/solr/collections/testSplitStaticIndexReplication_link/state.json
   [junit4]   2>        9       /solr/clusterprops.json
   [junit4]   2>        9       /solr/clusterstate.json
   [junit4]   2>        7       /solr/collections/control_collection/state.json
   [junit4]   2>        3       
/solr/overseer_elect/election/72177911064231948-127.0.0.1:41987_onek%2Fc-n_0000000001
   [junit4]   2>        2       
/solr/overseer_elect/election/72177911064231940-127.0.0.1:43681_onek%2Fc-n_0000000000
   [junit4]   2> 
   [junit4]   2> Maximum concurrent children watches above limit:
   [junit4]   2> 
   [junit4]   2>        65      /solr/overseer/queue
   [junit4]   2>        51      /solr/live_nodes
   [junit4]   2>        29      
/solr/collections/testSplitStaticIndexReplication_link/state.json
   [junit4]   2>        23      /solr/overseer/collection-queue-work
   [junit4]   2>        15      /solr/collections
   [junit4]   2>        8       /solr/collections/collection1/state.json
   [junit4]   2>        2       /solr/collections/control_collection/state.json
   [junit4]   2> 
   [junit4]   2> 1408139 INFO  
(TEST-ShardSplitTest.testSplitStaticIndexReplicationLink-seed#[C0258A7B5C6D0A61])
 [     ] o.a.s.c.ZkTestServer waitForServerDown: 127.0.0.1:44979
   [junit4]   2> 1408139 INFO  
(TEST-ShardSplitTest.testSplitStaticIndexReplicationLink-seed#[C0258A7B5C6D0A61])
 [     ] o.a.s.c.ZkTestServer parse host and port list: 127.0.0.1:44979
   [junit4]   2> 1408139 INFO  
(TEST-ShardSplitTest.testSplitStaticIndexReplicationLink-seed#[C0258A7B5C6D0A61])
 [     ] o.a.s.c.ZkTestServer connecting to 127.0.0.1 44979
   [junit4]   2> 1408141 INFO  
(SUITE-ShardSplitTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.a.s.u.ErrorLogMuter Closing ErrorLogMuter-regex-340 after mutting 0 log 
messages
   [junit4]   2> 1408141 INFO  
(SUITE-ShardSplitTest-seed#[C0258A7B5C6D0A61]-worker) [     ] 
o.a.s.u.ErrorLogMuter Creating ErrorLogMuter-regex-341 for ERROR logs matching 
regex: ignore_exception
   [junit4]   2> NOTE: leaving temporary files on disk at: 
/home/jenkins/jenkins-slave/workspace/Lucene/Lucene-Solr-Tests-8.11/solr/build/solr-core/test/J3/temp/solr.cloud.api.collections.ShardSplitTest_C0258A7B5C6D0A61-001
   [junit4]   2> Feb 11, 2024 11:10:49 PM 
com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
   [junit4]   2> WARNING: Will linger awaiting termination of 1 leaked 
thread(s).
   [junit4]   2> NOTE: test params are: codec=Asserting(Lucene87): 
{rnd_b=BlockTreeOrds(blocksize=128), 
a_t=PostingsFormat(name=LuceneVarGapDocFreqInterval), _root_=Lucene84, 
foo_s=PostingsFormat(name=LuceneVarGapDocFreqInterval), 
id=BlockTreeOrds(blocksize=128)}, 
docValues:{_version_=DocValuesFormat(name=Direct), 
a_i=DocValuesFormat(name=Lucene80)}, maxPointsInLeafNode=226, 
maxMBSortInHeap=6.985192169296044, 
sim=Asserting(RandomSimilarity(queryNorm=true): {}), locale=pt, 
timezone=Brazil/East
   [junit4]   2> NOTE: Linux 4.15.0-213-generic amd64/Temurin 1.8.0_362 
(64-bit)/cpus=4,threads=1,free=268215136,total=503840768
   [junit4]   2> NOTE: All tests run in this JVM: [SOLR749Test, 
TestStandardQParsers, TestAddFieldRealTimeGet, 
TestManagedSynonymGraphFilterFactory, HdfsRecoveryZkTest, 
CollectionsAPIDistributedZkTest, QueryResultKeyTest, TestMergePolicyConfig, 
TestUtils, StatelessScriptUpdateProcessorFactoryTest, TestGroupingSearch, 
NestedAtomicUpdateTest, TestSimDistribStateManager, BufferStoreTest, 
TestSubQueryTransformerDistrib, TestUnifiedSolrHighlighterWithoutStoredId, 
TestIncrementalCoreBackup, AlternateDirectoryTest, TestSolrFieldCacheBean, 
SplitHandlerTest, ChangedSchemaMergeTest, MissingSegmentRecoveryTest, 
ClassificationUpdateProcessorFactoryTest, TestHighFrequencyDictionaryFactory, 
TestExactStatsCache, TestConfigOverlay, PolyFieldTest, 
TestInPlaceUpdatesStandalone, TestNoOpRegenerator, TestIBSimilarityFactory, 
TestFaceting, DirectoryFactoryTest, TestSimPolicyCloud, TestCustomSort, 
TestSchemaNameResource, TermVectorComponentTest, TestNumericTerms32, 
CdcrOpsAndBoundariesTest, JWTAuthPluginIntegrationTest, 
AutoscalingHistoryHandlerTest, TestSizeLimitedDistributedMap, 
StatsComponentTest, SolrTestCaseJ4DeleteCoreTest, TestFastLRUCache, 
MinimalSchemaTest, TestAuthenticationFramework, RootFieldTest, 
TestNonDefinedSimilarityFactory, CloudExitableDirectoryReaderTest, 
TestBadConfig, PingRequestHandlerTest, TestDistributedGrouping, 
SolrPluginUtilsTest, TestJavabinTupleStreamParser, TestUseDocValuesAsStored, 
SecurityConfHandlerTest, HdfsTlogReplayBufferedWhileIndexingTest, 
TestDynamicFieldNamesIndexCorrectly, SolrJmxReporterCloudTest, 
MultiSolrCloudTestCaseTest, DistributedQueryElevationComponentTest, 
ConjunctionSolrSpellCheckerTest, HighlighterTest, TestSnapshotCloudManager, 
TestRequestStatusCollectionAPI, SystemLogListenerTest, TestNestedDocsSort, 
TestExportWriter, URLClassifyProcessorTest, HdfsCloudIncrementalBackupTest, 
HdfsChaosMonkeyNothingIsSafeTest, TestCollectionsAPIViaSolrCloudCluster, 
TestSchemaManager, TestPrepRecovery, AnalysisErrorHandlingTest, 
TestStressUserVersions, TestBulkSchemaConcurrent, SaslZkACLProviderTest, 
TestQuerySenderNoQuery, TestManagedSchemaThreadSafety, 
CrossCollectionJoinQueryTest, LargeFieldTest, DefaultValueUpdateProcessorTest, 
FacetPivot2CollectionsTest, TestSerializedLuceneMatchVersion, TestSolrXml, 
TestRandomCollapseQParserPlugin, ResponseLogComponentTest, 
TestDocTermOrdsUninvertLimit, ShardBackupIdTest, CursorPagingTest, 
DebugComponentTest, TransactionLogTest, TestFieldCacheSortRandom, 
QueryParsingTest, TestInitQParser, TestMacros, TestLeaderElectionZkExpiry, 
IndexSizeTriggerMixedBoundsTest, TestDelegationWithHadoopAuth, 
TestComplexPhraseLeadingWildcard, ImplicitSnitchTest, RandomizedTaggerTest, 
EnumFieldTest, TestSolr4Spatial, ConnectionManagerTest, 
TestZkAclsWithHadoopAuth, TestEmbeddedSolrServerSchemaAPI, 
ChaosMonkeySafeLeaderTest, DocumentAnalysisRequestHandlerTest, 
XMLAtomicUpdateMultivalueTest, DeleteStatusTest, NoCacheHeaderTest, 
AuthWithShardHandlerFactoryOverrideTest, ShardSplitTest]
   [junit4] Completed [521/959 (2!)] on J3 in 351.82s, 11 tests, 1 error, 3 
skipped <<< FAILURES!

[...truncated 55842 lines...]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to