Build: https://builds.apache.org/job/Lucene-Solr-Tests-8.x/491/

2 tests failed.
FAILED:  
org.apache.solr.cloud.api.collections.TestHdfsCloudBackupRestore.testRestoreFailure

Error Message:
Failed collection is still in the clusterstate: 
DocCollection(hdfsbackuprestore_testfailure_restored//collections/hdfsbackuprestore_testfailure_restored/state.json/2)={
   "pullReplicas":0,   "replicationFactor":1,   "shards":{     "shard2":{       
"range":"0-7fffffff",       "state":"construction",       
"replicas":{"core_node2":{           
"core":"hdfsbackuprestore_testfailure_restored_shard2_replica_n1",           
"base_url":"https://127.0.0.1:36659/solr";,           
"node_name":"127.0.0.1:36659_solr",           "state":"down",           
"type":"NRT",           "force_set_state":"false"}},       
"stateTimestamp":"1567059232049688251"},     "shard1":{       
"range":"80000000-ffffffff",       "state":"construction",       "replicas":{}, 
      "stateTimestamp":"1567059232049701653"}},   
"router":{"name":"compositeId"},   "maxShardsPerNode":"1",   
"autoAddReplicas":"false",   "nrtReplicas":1,   "tlogReplicas":0} Expected: not 
a collection containing "hdfsbackuprestore_testfailure_restored"      but: was 
<[hdfsbackuprestore_testok, hdfsbackuprestore_testfailure_restored, 
hdfsbackuprestore_testfailure, hdfsbackuprestore_testok_restored]>

Stack Trace:
java.lang.AssertionError: Failed collection is still in the clusterstate: 
DocCollection(hdfsbackuprestore_testfailure_restored//collections/hdfsbackuprestore_testfailure_restored/state.json/2)={
  "pullReplicas":0,
  "replicationFactor":1,
  "shards":{
    "shard2":{
      "range":"0-7fffffff",
      "state":"construction",
      "replicas":{"core_node2":{
          "core":"hdfsbackuprestore_testfailure_restored_shard2_replica_n1",
          "base_url":"https://127.0.0.1:36659/solr";,
          "node_name":"127.0.0.1:36659_solr",
          "state":"down",
          "type":"NRT",
          "force_set_state":"false"}},
      "stateTimestamp":"1567059232049688251"},
    "shard1":{
      "range":"80000000-ffffffff",
      "state":"construction",
      "replicas":{},
      "stateTimestamp":"1567059232049701653"}},
  "router":{"name":"compositeId"},
  "maxShardsPerNode":"1",
  "autoAddReplicas":"false",
  "nrtReplicas":1,
  "tlogReplicas":0}
Expected: not a collection containing "hdfsbackuprestore_testfailure_restored"
     but: was <[hdfsbackuprestore_testok, 
hdfsbackuprestore_testfailure_restored, hdfsbackuprestore_testfailure, 
hdfsbackuprestore_testok_restored]>
        at 
__randomizedtesting.SeedInfo.seed([E037D74065656872:C94B49654D3C6B5F]:0)
        at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
        at org.junit.Assert.assertThat(Assert.java:956)
        at 
org.apache.solr.cloud.api.collections.AbstractCloudBackupRestoreTestCase.testRestoreFailure(AbstractCloudBackupRestoreTestCase.java:211)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at java.lang.Thread.run(Thread.java:748)


FAILED:  
org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore.testRestoreFailure

Error Message:
Failed collection is still in the clusterstate: 
DocCollection(backuprestore_testfailure_restored//collections/backuprestore_testfailure_restored/state.json/2)={
   "pullReplicas":0,   "replicationFactor":1,   "shards":{     "shard2":{       
"range":"0-7fffffff",       "state":"construction",       
"replicas":{"core_node2":{           
"core":"backuprestore_testfailure_restored_shard2_replica_n1",           
"base_url":"http://127.0.0.1:33205/solr";,           
"node_name":"127.0.0.1:33205_solr",           "state":"down",           
"type":"NRT",           "force_set_state":"false"}},       
"stateTimestamp":"1567060879213084847"},     "shard1":{       
"range":"80000000-ffffffff",       "state":"construction",       "replicas":{}, 
      "stateTimestamp":"1567060879213099152"}},   
"router":{"name":"compositeId"},   "maxShardsPerNode":"1",   
"autoAddReplicas":"false",   "nrtReplicas":1,   "tlogReplicas":0} Expected: not 
a collection containing "backuprestore_testfailure_restored"      but: was 
<[backuprestore_testok, backuprestore_testfailure, 
backuprestore_testfailure_restored, backuprestore_testok_restored]>

Stack Trace:
java.lang.AssertionError: Failed collection is still in the clusterstate: 
DocCollection(backuprestore_testfailure_restored//collections/backuprestore_testfailure_restored/state.json/2)={
  "pullReplicas":0,
  "replicationFactor":1,
  "shards":{
    "shard2":{
      "range":"0-7fffffff",
      "state":"construction",
      "replicas":{"core_node2":{
          "core":"backuprestore_testfailure_restored_shard2_replica_n1",
          "base_url":"http://127.0.0.1:33205/solr";,
          "node_name":"127.0.0.1:33205_solr",
          "state":"down",
          "type":"NRT",
          "force_set_state":"false"}},
      "stateTimestamp":"1567060879213084847"},
    "shard1":{
      "range":"80000000-ffffffff",
      "state":"construction",
      "replicas":{},
      "stateTimestamp":"1567060879213099152"}},
  "router":{"name":"compositeId"},
  "maxShardsPerNode":"1",
  "autoAddReplicas":"false",
  "nrtReplicas":1,
  "tlogReplicas":0}
Expected: not a collection containing "backuprestore_testfailure_restored"
     but: was <[backuprestore_testok, backuprestore_testfailure, 
backuprestore_testfailure_restored, backuprestore_testok_restored]>
        at 
__randomizedtesting.SeedInfo.seed([E037D74065656872:C94B49654D3C6B5F]:0)
        at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
        at org.junit.Assert.assertThat(Assert.java:956)
        at 
org.apache.solr.cloud.api.collections.AbstractCloudBackupRestoreTestCase.testRestoreFailure(AbstractCloudBackupRestoreTestCase.java:211)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at java.lang.Thread.run(Thread.java:748)




Build Log:
[...truncated 13726 lines...]
   [junit4] Suite: 
org.apache.solr.cloud.api.collections.TestHdfsCloudBackupRestore
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 439279 WARN  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried 
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 439296 WARN  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 439298 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.Server jetty-9.4.19.v20190610; built: 2019-06-10T16:30:51.723Z; git: 
afcf563148970e98786327af5e07c261fda175d3; jvm 1.8.0_191-b12
   [junit4]   2> 439300 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 439300 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 439300 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 439301 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@2774068b{static,/static,jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.0-tests.jar!/webapps/static,AVAILABLE}
   [junit4]   2> 439459 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.w.WebAppContext@f007949{hdfs,/,file:///home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/jetty-localhost.localdomain-36239-hdfs-_-any-924387434669286531.dir/webapp/,AVAILABLE}{/hdfs}
   [junit4]   2> 439460 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@7f6b887c{HTTP/1.1,[http/1.1]}{localhost.localdomain:36239}
   [junit4]   2> 439461 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.Server Started @439527ms
   [junit4]   2> 439553 WARN  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 439556 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.Server jetty-9.4.19.v20190610; built: 2019-06-10T16:30:51.723Z; git: 
afcf563148970e98786327af5e07c261fda175d3; jvm 1.8.0_191-b12
   [junit4]   2> 439556 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 439556 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 439557 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 439557 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@6adf3fad{static,/static,jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.0-tests.jar!/webapps/static,AVAILABLE}
   [junit4]   2> 439714 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.w.WebAppContext@1c703108{datanode,/,file:///home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/jetty-localhost-37543-datanode-_-any-7314119788980653551.dir/webapp/,AVAILABLE}{/datanode}
   [junit4]   2> 439715 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@7162d9c9{HTTP/1.1,[http/1.1]}{localhost:37543}
   [junit4]   2> 439715 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.Server Started @439781ms
   [junit4]   2> 439791 WARN  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 439792 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.Server jetty-9.4.19.v20190610; built: 2019-06-10T16:30:51.723Z; git: 
afcf563148970e98786327af5e07c261fda175d3; jvm 1.8.0_191-b12
   [junit4]   2> 439794 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 439794 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 439794 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 439795 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@11532006{static,/static,jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.0-tests.jar!/webapps/static,AVAILABLE}
   [junit4]   2> 439974 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xcac9d599fa230d3: Processing first storage report for 
DS-d29fa2ae-e164-4cca-aa51-f36bddc1bd73 from datanode 
f9a9e1ed-6c2b-46ce-b8bb-7bae1b0f893d
   [junit4]   2> 439974 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xcac9d599fa230d3: from storage 
DS-d29fa2ae-e164-4cca-aa51-f36bddc1bd73 node 
DatanodeRegistration(127.0.0.1:38477, 
datanodeUuid=f9a9e1ed-6c2b-46ce-b8bb-7bae1b0f893d, infoPort=34367, 
infoSecurePort=0, ipcPort=41639, 
storageInfo=lv=-57;cid=testClusterID;nsid=968518402;c=1567059213337), blocks: 
0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0
   [junit4]   2> 439974 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xcac9d599fa230d3: Processing first storage report for 
DS-1c4da6b9-2544-4f1f-b527-c4142a5267fd from datanode 
f9a9e1ed-6c2b-46ce-b8bb-7bae1b0f893d
   [junit4]   2> 439974 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xcac9d599fa230d3: from storage 
DS-1c4da6b9-2544-4f1f-b527-c4142a5267fd node 
DatanodeRegistration(127.0.0.1:38477, 
datanodeUuid=f9a9e1ed-6c2b-46ce-b8bb-7bae1b0f893d, infoPort=34367, 
infoSecurePort=0, ipcPort=41639, 
storageInfo=lv=-57;cid=testClusterID;nsid=968518402;c=1567059213337), blocks: 
0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0
   [junit4]   2> 440012 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.w.WebAppContext@397cc67d{datanode,/,file:///home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/jetty-localhost-42969-datanode-_-any-8925702212772852981.dir/webapp/,AVAILABLE}{/datanode}
   [junit4]   2> 440012 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@4892d943{HTTP/1.1,[http/1.1]}{localhost:42969}
   [junit4]   2> 440012 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.e.j.s.Server Started @440079ms
   [junit4]   2> 440173 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xb57ad50ac595db20: Processing first storage report for 
DS-ae78d8eb-dd57-4c19-ae6e-ea8f8519c130 from datanode 
d9c8819b-1365-4c42-ae05-ffe965768d2c
   [junit4]   2> 440173 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xb57ad50ac595db20: from storage 
DS-ae78d8eb-dd57-4c19-ae6e-ea8f8519c130 node 
DatanodeRegistration(127.0.0.1:42413, 
datanodeUuid=d9c8819b-1365-4c42-ae05-ffe965768d2c, infoPort=40741, 
infoSecurePort=0, ipcPort=34355, 
storageInfo=lv=-57;cid=testClusterID;nsid=968518402;c=1567059213337), blocks: 
0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0
   [junit4]   2> 440173 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xb57ad50ac595db20: Processing first storage report for 
DS-ef48ae90-a08b-4a82-9795-00787d190e45 from datanode 
d9c8819b-1365-4c42-ae05-ffe965768d2c
   [junit4]   2> 440173 INFO  (Block report processor) [     ] BlockStateChange 
BLOCK* processReport 0xb57ad50ac595db20: from storage 
DS-ef48ae90-a08b-4a82-9795-00787d190e45 node 
DatanodeRegistration(127.0.0.1:42413, 
datanodeUuid=d9c8819b-1365-4c42-ae05-ffe965768d2c, infoPort=40741, 
infoSecurePort=0, ipcPort=34355, 
storageInfo=lv=-57;cid=testClusterID;nsid=968518402;c=1567059213337), blocks: 
0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0
   [junit4]   2> 440259 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.MiniSolrCloudCluster Starting cluster of 2 servers in 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002
   [junit4]   2> 440260 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 440260 INFO  (ZkTestServer Run Thread) [     ] 
o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 440260 INFO  (ZkTestServer Run Thread) [     ] 
o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 440360 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.ZkTestServer start zk server on port:45147
   [junit4]   2> 440360 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.ZkTestServer waitForServerUp: 127.0.0.1:45147
   [junit4]   2> 440360 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.ZkTestServer parse host and port list: 127.0.0.1:45147
   [junit4]   2> 440360 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.ZkTestServer connecting to 127.0.0.1 45147
   [junit4]   2> 440363 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 440367 INFO  (zkConnectionManagerCallback-2523-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 440367 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 440371 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 440372 INFO  (zkConnectionManagerCallback-2525-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 440372 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 440376 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 440378 INFO  (zkConnectionManagerCallback-2527-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 440378 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 440486 WARN  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.s.AbstractConnector Ignoring deprecated socket close linger time
   [junit4]   2> 440486 WARN  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.s.AbstractConnector Ignoring deprecated socket close linger time
   [junit4]   2> 440487 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.c.s.e.JettySolrRunner Start Jetty (original configured port=0)
   [junit4]   2> 440487 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.c.s.e.JettySolrRunner Start Jetty (original configured port=0)
   [junit4]   2> 440487 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.c.s.e.JettySolrRunner Trying to start Jetty on port 0 try number 1 ...
   [junit4]   2> 440487 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.c.s.e.JettySolrRunner Trying to start Jetty on port 0 try number 1 ...
   [junit4]   2> 440487 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.s.Server jetty-9.4.19.v20190610; built: 2019-06-10T16:30:51.723Z; git: 
afcf563148970e98786327af5e07c261fda175d3; jvm 1.8.0_191-b12
   [junit4]   2> 440487 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.s.Server jetty-9.4.19.v20190610; built: 2019-06-10T16:30:51.723Z; git: 
afcf563148970e98786327af5e07c261fda175d3; jvm 1.8.0_191-b12
   [junit4]   2> 440495 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 440495 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 440496 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 440496 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 440496 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 440496 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 440496 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@45e0eb05{/solr,null,AVAILABLE}
   [junit4]   2> 440496 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@2f3604af{/solr,null,AVAILABLE}
   [junit4]   2> 440497 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.s.AbstractConnector Started ServerConnector@224c8694{SSL,[ssl, 
http/1.1]}{127.0.0.1:36659}
   [junit4]   2> 440497 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.s.Server Started @440564ms
   [junit4]   2> 440497 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, 
hostPort=36659}
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.s.AbstractConnector Started ServerConnector@341995c6{SSL,[ssl, 
http/1.1]}{127.0.0.1:46735}
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.s.Server Started @440564ms
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, 
hostPort=46735}
   [junit4]   2> 440498 ERROR (jetty-launcher-2528-thread-2) [     ] 
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter Using logger factory 
org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 440498 ERROR (jetty-launcher-2528-thread-1) [     ] 
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 
8.3.0
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter Using logger factory 
org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 
8.3.0
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2019-08-29T06:13:34.605Z
   [junit4]   2> 440498 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2019-08-29T06:13:34.605Z
   [junit4]   2> 440500 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 440503 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 440503 INFO  (zkConnectionManagerCallback-2530-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 440503 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 440504 INFO  (zkConnectionManagerCallback-2532-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 440504 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 440504 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 440505 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 440523 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 440530 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 440887 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.h.c.HttpShardHandlerFactory Host whitelist initialized: 
WhitelistHostChecker [whitelistHosts=null, whitelistHostCheckingEnabled=true]
   [junit4]   2> 440888 WARN  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport 
since Java 8 or lower versions does not support SSL + HTTP/2
   [junit4]   2> 440889 WARN  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.u.s.S.config Trusting all certificates configured for 
Client@4b3ef0b7[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 440889 WARN  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for 
Client@4b3ef0b7[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 440890 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.h.c.HttpShardHandlerFactory Host whitelist initialized: 
WhitelistHostChecker [whitelistHosts=null, whitelistHostCheckingEnabled=true]
   [junit4]   2> 440891 WARN  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport 
since Java 8 or lower versions does not support SSL + HTTP/2
   [junit4]   2> 440893 WARN  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.u.s.S.config Trusting all certificates configured for 
Client@7637eb82[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 440893 WARN  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for 
Client@7637eb82[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 440895 WARN  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport 
since Java 8 or lower versions does not support SSL + HTTP/2
   [junit4]   2> 440901 WARN  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport 
since Java 8 or lower versions does not support SSL + HTTP/2
   [junit4]   2> 440901 WARN  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.u.s.S.config Trusting all certificates configured for 
Client@74bd9149[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 440901 WARN  (jetty-launcher-2528-thread-2) [     ] 
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for 
Client@74bd9149[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 440902 WARN  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.u.s.S.config Trusting all certificates configured for 
Client@18e664b8[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 440902 WARN  (jetty-launcher-2528-thread-1) [     ] 
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for 
Client@18e664b8[provider=null,keyStore=null,trustStore=null]
   [junit4]   2> 440903 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45147/solr
   [junit4]   2> 440903 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45147/solr
   [junit4]   2> 440905 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 440906 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 440906 INFO  (zkConnectionManagerCallback-2546-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 440906 INFO  (jetty-launcher-2528-thread-1) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 440911 INFO  (zkConnectionManagerCallback-2544-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 440911 INFO  (jetty-launcher-2528-thread-2) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 441011 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.c.ConnectionManager Waiting for client to 
connect to ZooKeeper
   [junit4]   2> 441014 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.c.ConnectionManager Waiting for client to 
connect to ZooKeeper
   [junit4]   2> 441014 INFO  (zkConnectionManagerCallback-2548-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 441014 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.c.ConnectionManager Client is connected 
to ZooKeeper
   [junit4]   2> 441024 INFO  (zkConnectionManagerCallback-2550-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 441024 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.c.ConnectionManager Client is connected 
to ZooKeeper
   [junit4]   2> 441203 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.OverseerElectionContext I am going to be 
the leader 127.0.0.1:46735_solr
   [junit4]   2> 441205 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.Overseer Overseer 
(id=72285712308305927-127.0.0.1:46735_solr-n_0000000000) starting
   [junit4]   2> 441223 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.c.ConnectionManager Waiting for client to 
connect to ZooKeeper
   [junit4]   2> 441231 INFO  (zkConnectionManagerCallback-2559-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 441231 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.c.ConnectionManager Client is connected 
to ZooKeeper
   [junit4]   2> 441237 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster 
at 127.0.0.1:45147/solr ready
   [junit4]   2> 441243 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:36659_solr
   [junit4]   2> 441246 INFO  
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.Overseer Starting to work on the main 
queue : 127.0.0.1:46735_solr
   [junit4]   2> 441247 INFO  
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.c.ZkStateReader Updated live nodes from 
ZooKeeper... (0) -> (1)
   [junit4]   2> 441257 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.ZkController Publish 
node=127.0.0.1:46735_solr as DOWN
   [junit4]   2> 441259 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.TransientSolrCoreCacheDefault Allocating 
transient cache for 2147483647 transient cores
   [junit4]   2> 441259 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:46735_solr
   [junit4]   2> 441262 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.PackageManager clusterprops.json changed 
, version 0
   [junit4]   2> 441262 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup 
repository with configuration params {type = repository,name = hdfs,class = 
org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes = 
{name=hdfs, 
class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args = 
{location=/backup,solr.hdfs.home=hdfs://localhost.localdomain:46481/solr,solr.hdfs.confdir=}}
   [junit4]   2> 441262 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup 
repository with configuration params {type = repository,name = poisioned,class 
= 
org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes
 = {default=true, name=poisioned, 
class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
   [junit4]   2> 441262 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Default 
configuration for backup repository is with configuration params {type = 
repository,name = poisioned,class = 
org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes
 = {default=true, name=poisioned, 
class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
   [junit4]   2> 441267 INFO  (zkCallback-2547-thread-1) [     ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 441274 INFO  (zkCallback-2558-thread-1) [     ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 441277 INFO  (zkCallback-2549-thread-1) [     ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 441280 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.c.ConnectionManager Waiting for client to 
connect to ZooKeeper
   [junit4]   2> 441301 INFO  (zkConnectionManagerCallback-2564-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 441301 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.c.ConnectionManager Client is connected 
to ZooKeeper
   [junit4]   2> 441302 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.c.ZkStateReader Updated live nodes from 
ZooKeeper... (0) -> (2)
   [junit4]   2> 441305 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster 
at 127.0.0.1:45147/solr ready
   [junit4]   2> 441306 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.PackageManager clusterprops.json changed 
, version 0
   [junit4]   2> 441306 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup 
repository with configuration params {type = repository,name = hdfs,class = 
org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes = 
{name=hdfs, 
class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args = 
{location=/backup,solr.hdfs.home=hdfs://localhost.localdomain:46481/solr,solr.hdfs.confdir=}}
   [junit4]   2> 441306 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Added backup 
repository with configuration params {type = repository,name = poisioned,class 
= 
org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes
 = {default=true, name=poisioned, 
class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
   [junit4]   2> 441306 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.b.r.BackupRepositoryFactory Default 
configuration for backup repository is with configuration params {type = 
repository,name = poisioned,class = 
org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes
 = {default=true, name=poisioned, 
class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
   [junit4]   2> 441332 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.h.a.MetricsHistoryHandler No .system 
collection, keeping metrics history in memory.
   [junit4]   2> 441378 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.h.a.MetricsHistoryHandler No .system 
collection, keeping metrics history in memory.
   [junit4]   2> 441409 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.node' (registry 'solr.node') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 441436 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.node' (registry 'solr.node') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 441444 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jvm' (registry 'solr.jvm') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 441444 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jetty' (registry 'solr.jetty') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 441445 INFO  (jetty-launcher-2528-thread-1) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.CorePropertiesLocator Found 0 core 
definitions underneath 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/.
   [junit4]   2> 441454 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jvm' (registry 'solr.jvm') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 441454 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jetty' (registry 'solr.jetty') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 441456 INFO  (jetty-launcher-2528-thread-2) 
[n:127.0.0.1:36659_solr     ] o.a.s.c.CorePropertiesLocator Found 0 core 
definitions underneath 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/.
   [junit4]   2> 441578 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.MiniSolrCloudCluster waitForAllNodes: numServers=2
   [junit4]   2> 441579 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 441584 INFO  (zkConnectionManagerCallback-2571-thread-1) [     
] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 441585 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 441588 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 441607 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [     ] 
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:45147/solr ready
   [junit4]   2> 441715 INFO  (qtp1840676713-6927) [n:127.0.0.1:36659_solr     
] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params 
collection.configName=conf1&router.name=implicit&version=2&pullReplicas=0&shards=shard1,shard2&property.customKey=customValue&maxShardsPerNode=3&router.field=shard_s&autoAddReplicas=true&name=hdfsbackuprestore_testok&nrtReplicas=2&action=CREATE&tlogReplicas=1&wt=javabin
 and sendToOCPQueue=true
   [junit4]   2> 441723 INFO  
(OverseerThreadFactory-1679-thread-1-processing-n:127.0.0.1:46735_solr) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.a.c.CreateCollectionCmd Create collection 
hdfsbackuprestore_testok
   [junit4]   2> 441830 WARN  
(OverseerThreadFactory-1679-thread-1-processing-n:127.0.0.1:46735_solr) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.a.c.CreateCollectionCmd Specified number 
of replicas of 3 on collection hdfsbackuprestore_testok is higher than the 
number of Solr instances currently live or live and part of your 
createNodeSet(2). It's unusual to run two replica of the same slice on the same 
Solr-instance.
   [junit4]   2> 441836 INFO  
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore_testok",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"hdfsbackuprestore_testok_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:46735/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 441841 INFO  
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore_testok",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"hdfsbackuprestore_testok_shard1_replica_n2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:36659/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 441847 INFO  
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore_testok",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"hdfsbackuprestore_testok_shard1_replica_t4",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:46735/solr";,
   [junit4]   2>   "type":"TLOG",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 441852 INFO  
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore_testok",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"hdfsbackuprestore_testok_shard2_replica_n7",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:36659/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 441855 INFO  
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore_testok",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"hdfsbackuprestore_testok_shard2_replica_n8",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:46735/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 441859 INFO  
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000) 
[n:127.0.0.1:46735_solr     ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore_testok",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"hdfsbackuprestore_testok_shard2_replica_t10",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:36659/solr";,
   [junit4]   2>   "type":"TLOG",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 442066 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr    
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.h.a.CoreAdminOperation 
core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node5&name=hdfsbackuprestore_testok_shard1_replica_n2&action=CREATE&numShards=2&shard=shard1&wt=javabin
   [junit4]   2> 442066 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr    
x:hdfsbackuprestore_testok_shard1_replica_n2 ] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 
transient cores
   [junit4]   2> 442078 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr    
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.h.a.CoreAdminOperation 
core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node9&name=hdfsbackuprestore_testok_shard2_replica_n7&action=CREATE&numShards=2&shard=shard2&wt=javabin
   [junit4]   2> 442084 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr    
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.h.a.CoreAdminOperation 
core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=TLOG&property.customKey=customValue&coreNodeName=core_node12&name=hdfsbackuprestore_testok_shard2_replica_t10&action=CREATE&numShards=2&shard=shard2&wt=javabin
   [junit4]   2> 442098 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr    
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.h.a.CoreAdminOperation 
core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=TLOG&property.customKey=customValue&coreNodeName=core_node6&name=hdfsbackuprestore_testok_shard1_replica_t4&action=CREATE&numShards=2&shard=shard1&wt=javabin
   [junit4]   2> 442101 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr    
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.h.a.CoreAdminOperation 
core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node3&name=hdfsbackuprestore_testok_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&wt=javabin
   [junit4]   2> 442112 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr    
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.h.a.CoreAdminOperation 
core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node11&name=hdfsbackuprestore_testok_shard2_replica_n8&action=CREATE&numShards=2&shard=shard2&wt=javabin
   [junit4]   2> 443135 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 8.3.0
   [junit4]   2> 443135 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 8.3.0
   [junit4]   2> 443138 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 8.3.0
   [junit4]   2> 443144 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 8.3.0
   [junit4]   2> 443144 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 8.3.0
   [junit4]   2> 443151 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 8.3.0
   [junit4]   2> 443188 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.s.IndexSchema 
[hdfsbackuprestore_testok_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 443204 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.s.IndexSchema 
[hdfsbackuprestore_testok_shard1_replica_t4] Schema name=minimal
   [junit4]   2> 443207 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.s.IndexSchema Loaded 
schema minimal/1.1 with uniqueid field id
   [junit4]   2> 443208 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.c.CoreContainer Creating 
SolrCore 'hdfsbackuprestore_testok_shard1_replica_t4' using configuration from 
collection hdfsbackuprestore_testok, trusted=true
   [junit4]   2> 443208 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.m.r.SolrJmxReporter JMX 
monitoring for 'solr.core.hdfsbackuprestore_testok.shard1.replica_t4' (registry 
'solr.core.hdfsbackuprestore_testok.shard1.replica_t4') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 443215 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.s.IndexSchema 
[hdfsbackuprestore_testok_shard2_replica_n7] Schema name=minimal
   [junit4]   2> 443218 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.s.IndexSchema 
[hdfsbackuprestore_testok_shard1_replica_n2] Schema name=minimal
   [junit4]   2> 443218 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.s.IndexSchema Loaded 
schema minimal/1.1 with uniqueid field id
   [junit4]   2> 443218 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.c.CoreContainer Creating 
SolrCore 'hdfsbackuprestore_testok_shard2_replica_n7' using configuration from 
collection hdfsbackuprestore_testok, trusted=true
   [junit4]   2> 443219 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.m.r.SolrJmxReporter JMX 
monitoring for 'solr.core.hdfsbackuprestore_testok.shard2.replica_n7' (registry 
'solr.core.hdfsbackuprestore_testok.shard2.replica_n7') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 443222 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.s.IndexSchema 
[hdfsbackuprestore_testok_shard2_replica_t10] Schema name=minimal
   [junit4]   2> 443226 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.s.IndexSchema Loaded 
schema minimal/1.1 with uniqueid field id
   [junit4]   2> 443226 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.c.CoreContainer Creating 
SolrCore 'hdfsbackuprestore_testok_shard1_replica_n1' using configuration from 
collection hdfsbackuprestore_testok, trusted=true
   [junit4]   2> 443227 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.m.r.SolrJmxReporter JMX 
monitoring for 'solr.core.hdfsbackuprestore_testok.shard1.replica_n1' (registry 
'solr.core.hdfsbackuprestore_testok.shard1.replica_n1') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 443230 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.s.IndexSchema Loaded 
schema minimal/1.1 with uniqueid field id
   [junit4]   2> 443231 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.c.CoreContainer Creating 
SolrCore 'hdfsbackuprestore_testok_shard1_replica_n2' using configuration from 
collection hdfsbackuprestore_testok, trusted=true
   [junit4]   2> 443231 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.c.SolrCore 
[[hdfsbackuprestore_testok_shard2_replica_n7] ] Opening new SolrCore at 
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/hdfsbackuprestore_testok_shard2_replica_n7],
 
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/./hdfsbackuprestore_testok_shard2_replica_n7/data/]
   [junit4]   2> 443232 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.m.r.SolrJmxReporter JMX 
monitoring for 'solr.core.hdfsbackuprestore_testok.shard1.replica_n2' (registry 
'solr.core.hdfsbackuprestore_testok.shard1.replica_n2') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 443232 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.c.SolrCore 
[[hdfsbackuprestore_testok_shard1_replica_n2] ] Opening new SolrCore at 
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/hdfsbackuprestore_testok_shard1_replica_n2],
 
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/./hdfsbackuprestore_testok_shard1_replica_n2/data/]
   [junit4]   2> 443233 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.c.SolrCore 
[[hdfsbackuprestore_testok_shard1_replica_t4] ] Opening new SolrCore at 
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/hdfsbackuprestore_testok_shard1_replica_t4],
 
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/./hdfsbackuprestore_testok_shard1_replica_t4/data/]
   [junit4]   2> 443234 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.s.IndexSchema Loaded 
schema minimal/1.1 with uniqueid field id
   [junit4]   2> 443234 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.c.CoreContainer Creating 
SolrCore 'hdfsbackuprestore_testok_shard2_replica_t10' using configuration from 
collection hdfsbackuprestore_testok, trusted=true
   [junit4]   2> 443235 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.m.r.SolrJmxReporter JMX 
monitoring for 'solr.core.hdfsbackuprestore_testok.shard2.replica_t10' 
(registry 'solr.core.hdfsbackuprestore_testok.shard2.replica_t10') enabled at 
server: com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 443235 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.c.SolrCore 
[[hdfsbackuprestore_testok_shard2_replica_t10] ] Opening new SolrCore at 
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/hdfsbackuprestore_testok_shard2_replica_t10],
 
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/./hdfsbackuprestore_testok_shard2_replica_t10/data/]
   [junit4]   2> 443236 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.c.SolrCore 
[[hdfsbackuprestore_testok_shard1_replica_n1] ] Opening new SolrCore at 
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/hdfsbackuprestore_testok_shard1_replica_n1],
 
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/./hdfsbackuprestore_testok_shard1_replica_n1/data/]
   [junit4]   2> 443240 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.s.IndexSchema 
[hdfsbackuprestore_testok_shard2_replica_n8] Schema name=minimal
   [junit4]   2> 443243 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.s.IndexSchema Loaded 
schema minimal/1.1 with uniqueid field id
   [junit4]   2> 443243 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.c.CoreContainer Creating 
SolrCore 'hdfsbackuprestore_testok_shard2_replica_n8' using configuration from 
collection hdfsbackuprestore_testok, trusted=true
   [junit4]   2> 443244 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.m.r.SolrJmxReporter JMX 
monitoring for 'solr.core.hdfsbackuprestore_testok.shard2.replica_n8' (registry 
'solr.core.hdfsbackuprestore_testok.shard2.replica_n8') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
   [junit4]   2> 443244 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.c.SolrCore 
[[hdfsbackuprestore_testok_shard2_replica_n8] ] Opening new SolrCore at 
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/hdfsbackuprestore_testok_shard2_replica_n8],
 
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/./hdfsbackuprestore_testok_shard2_replica_n8/data/]
   [junit4]   2> 443410 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.UpdateHandler Using 
UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 443410 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.UpdateLog Initializing 
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 443412 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.CommitTracker Hard 
AutoCommit: disabled
   [junit4]   2> 443412 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.CommitTracker Soft 
AutoCommit: disabled
   [junit4]   2> 443418 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.s.SolrIndexSearcher 
Opening [Searcher@5d6f00f3[hdfsbackuprestore_testok_shard1_replica_n1] main]
   [junit4]   2> 443427 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 443428 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.r.ManagedResourceStorage 
Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 443432 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.h.ReplicationHandler 
Commits will be reserved for 10000ms.
   [junit4]   2> 443432 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.UpdateLog Could not find 
max version in index or recent updates, using new clock 1643180686090174464
   [junit4]   2> 443440 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.c.ZkShardTerms Successful 
update of terms at /collections/hdfsbackuprestore_testok/terms/shard1 to 
Terms{values={core_node3=0}, version=0}
   [junit4]   2> 443441 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] 
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created 
/collections/hdfsbackuprestore_testok/leaders/shard1
   [junit4]   2> 443444 INFO  
(searcherExecutor-1690-thread-1-processing-n:127.0.0.1:46735_solr 
x:hdfsbackuprestore_testok_shard1_replica_n1 c:hdfsbackuprestore_testok 
s:shard1 r:core_node3) [n:127.0.0.1:46735_solr c:hdfsbackuprestore_testok 
s:shard1 r:core_node3 x:hdfsbackuprestore_testok_shard1_replica_n1 ] 
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard1_replica_n1] Registered new 
searcher Searcher@5d6f00f3[hdfsbackuprestore_testok_shard1_replica_n1] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 443449 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] 
o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for 
shard shard1: total=3 found=1 timeoutin=9999ms
   [junit4]   2> 443462 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.u.UpdateHandler Using 
UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 443462 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.u.UpdateLog Initializing 
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 443462 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.u.UpdateHandler Using 
UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 443462 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.u.UpdateLog Initializing 
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 443463 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.u.CommitTracker Hard 
AutoCommit: disabled
   [junit4]   2> 443463 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.u.CommitTracker Soft 
AutoCommit: disabled
   [junit4]   2> 443463 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.u.CommitTracker Hard 
AutoCommit: disabled
   [junit4]   2> 443464 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.u.CommitTracker Soft 
AutoCommit: disabled
   [junit4]   2> 443466 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.u.UpdateHandler Using 
UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 443466 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.u.UpdateLog Initializing 
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 443467 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.s.SolrIndexSearcher 
Opening [Searcher@7add75ac[hdfsbackuprestore_testok_shard1_replica_n2] main]
   [junit4]   2> 443467 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.u.CommitTracker Hard 
AutoCommit: disabled
   [junit4]   2> 443467 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.u.CommitTracker Soft 
AutoCommit: disabled
   [junit4]   2> 443470 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.s.SolrIndexSearcher 
Opening [Searcher@abe7412[hdfsbackuprestore_testok_shard1_replica_t4] main]
   [junit4]   2> 443473 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 443475 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.r.ManagedResourceStorage 
Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 443475 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.h.ReplicationHandler 
Commits will be reserved for 10000ms.
   [junit4]   2> 443476 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.u.UpdateLog Could not find 
max version in index or recent updates, using new clock 1643180686136311808
   [junit4]   2> 443484 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 443484 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.s.SolrIndexSearcher 
Opening [Searcher@40bbb2b8[hdfsbackuprestore_testok_shard2_replica_n8] main]
   [junit4]   2> 443485 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.r.ManagedResourceStorage 
Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 443485 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.h.ReplicationHandler 
Commits will be reserved for 10000ms.
   [junit4]   2> 443485 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.u.UpdateLog Could not find 
max version in index or recent updates, using new clock 1643180686145748992
   [junit4]   2> 443487 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 443488 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.r.ManagedResourceStorage 
Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 443488 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.h.ReplicationHandler 
Commits will be reserved for 10000ms.
   [junit4]   2> 443489 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.u.UpdateLog Could not find 
max version in index or recent updates, using new clock 1643180686148894720
   [junit4]   2> 443494 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.u.UpdateHandler Using 
UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 443494 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.u.UpdateLog Initializing 
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 443494 INFO  
(searcherExecutor-1691-thread-1-processing-n:127.0.0.1:36659_solr 
x:hdfsbackuprestore_testok_shard1_replica_n2 c:hdfsbackuprestore_testok 
s:shard1 r:core_node5) [n:127.0.0.1:36659_solr c:hdfsbackuprestore_testok 
s:shard1 r:core_node5 x:hdfsbackuprestore_testok_shard1_replica_n2 ] 
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard1_replica_n2] Registered new 
searcher Searcher@7add75ac[hdfsbackuprestore_testok_shard1_replica_n2] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 443495 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.u.CommitTracker Hard 
AutoCommit: disabled
   [junit4]   2> 443495 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.u.CommitTracker Soft 
AutoCommit: disabled
   [junit4]   2> 443496 INFO  
(searcherExecutor-1693-thread-1-processing-n:127.0.0.1:46735_solr 
x:hdfsbackuprestore_testok_shard2_replica_n8 c:hdfsbackuprestore_testok 
s:shard2 r:core_node11) [n:127.0.0.1:46735_solr c:hdfsbackuprestore_testok 
s:shard2 r:core_node11 x:hdfsbackuprestore_testok_shard2_replica_n8 ] 
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard2_replica_n8] Registered new 
searcher Searcher@40bbb2b8[hdfsbackuprestore_testok_shard2_replica_n8] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 443497 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.c.ZkShardTerms Successful 
update of terms at /collections/hdfsbackuprestore_testok/terms/shard1 to 
Terms{values={core_node6=0, core_node3=0}, version=1}
   [junit4]   2> 443498 INFO  (qtp2078506737-6920) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node6 
x:hdfsbackuprestore_testok_shard1_replica_t4 ] 
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created 
/collections/hdfsbackuprestore_testok/leaders/shard1
   [junit4]   2> 443500 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.s.SolrIndexSearcher 
Opening [Searcher@69173d6[hdfsbackuprestore_testok_shard2_replica_t10] main]
   [junit4]   2> 443501 INFO  
(searcherExecutor-1688-thread-1-processing-n:127.0.0.1:46735_solr 
x:hdfsbackuprestore_testok_shard1_replica_t4 c:hdfsbackuprestore_testok 
s:shard1 r:core_node6) [n:127.0.0.1:46735_solr c:hdfsbackuprestore_testok 
s:shard1 r:core_node6 x:hdfsbackuprestore_testok_shard1_replica_t4 ] 
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard1_replica_t4] Registered new 
searcher Searcher@abe7412[hdfsbackuprestore_testok_shard1_replica_t4] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 443503 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 443504 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.r.ManagedResourceStorage 
Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 443505 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.h.ReplicationHandler 
Commits will be reserved for 10000ms.
   [junit4]   2> 443505 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.u.UpdateLog Could not 
find max version in index or recent updates, using new clock 1643180686166720512
   [junit4]   2> 443505 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.c.ZkShardTerms Successful 
update of terms at /collections/hdfsbackuprestore_testok/terms/shard1 to 
Terms{values={core_node6=0, core_node3=0, core_node5=0}, version=2}
   [junit4]   2> 443505 INFO  (qtp1840676713-6923) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] 
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created 
/collections/hdfsbackuprestore_testok/leaders/shard1
   [junit4]   2> 443514 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.c.ZkShardTerms Successful 
update of terms at /collections/hdfsbackuprestore_testok/terms/shard2 to 
Terms{values={core_node11=0}, version=0}
   [junit4]   2> 443514 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] 
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created 
/collections/hdfsbackuprestore_testok/leaders/shard2
   [junit4]   2> 443520 INFO  
(searcherExecutor-1692-thread-1-processing-n:127.0.0.1:36659_solr 
x:hdfsbackuprestore_testok_shard2_replica_t10 c:hdfsbackuprestore_testok 
s:shard2 r:core_node12) [n:127.0.0.1:36659_solr c:hdfsbackuprestore_testok 
s:shard2 r:core_node12 x:hdfsbackuprestore_testok_shard2_replica_t10 ] 
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard2_replica_t10] Registered new 
searcher Searcher@69173d6[hdfsbackuprestore_testok_shard2_replica_t10] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 443521 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.c.ZkShardTerms Successful 
update of terms at /collections/hdfsbackuprestore_testok/terms/shard2 to 
Terms{values={core_node12=0, core_node11=0}, version=1}
   [junit4]   2> 443524 INFO  (qtp2078506737-6924) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node11 
x:hdfsbackuprestore_testok_shard2_replica_n8 ] 
o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for 
shard shard2: total=3 found=1 timeoutin=9998ms
   [junit4]   2> 443530 INFO  (qtp1840676713-6921) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node12 
x:hdfsbackuprestore_testok_shard2_replica_t10 ] 
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created 
/collections/hdfsbackuprestore_testok/leaders/shard2
   [junit4]   2> 443530 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.u.UpdateHandler Using 
UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 443530 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.u.UpdateLog Initializing 
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 443532 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.u.CommitTracker Hard 
AutoCommit: disabled
   [junit4]   2> 443532 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.u.CommitTracker Soft 
AutoCommit: disabled
   [junit4]   2> 443536 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.s.SolrIndexSearcher 
Opening [Searcher@50d2c5a[hdfsbackuprestore_testok_shard2_replica_n7] main]
   [junit4]   2> 443538 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 443538 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.r.ManagedResourceStorage 
Loaded null at path _rest_managed.json using 
ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 443539 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.h.ReplicationHandler 
Commits will be reserved for 10000ms.
   [junit4]   2> 443539 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.u.UpdateLog Could not find 
max version in index or recent updates, using new clock 1643180686202372096
   [junit4]   2> 443545 INFO  
(searcherExecutor-1689-thread-1-processing-n:127.0.0.1:36659_solr 
x:hdfsbackuprestore_testok_shard2_replica_n7 c:hdfsbackuprestore_testok 
s:shard2 r:core_node9) [n:127.0.0.1:36659_solr c:hdfsbackuprestore_testok 
s:shard2 r:core_node9 x:hdfsbackuprestore_testok_shard2_replica_n7 ] 
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard2_replica_n7] Registered new 
searcher Searcher@50d2c5a[hdfsbackuprestore_testok_shard2_replica_n7] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 443546 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.c.ZkShardTerms Successful 
update of terms at /collections/hdfsbackuprestore_testok/terms/shard2 to 
Terms{values={core_node12=0, core_node11=0, core_node9=0}, version=2}
   [junit4]   2> 443546 INFO  (qtp1840676713-6919) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard2 r:core_node9 
x:hdfsbackuprestore_testok_shard2_replica_n7 ] 
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created 
/collections/hdfsbackuprestore_testok/leaders/shard2
   [junit4]   2> 443952 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] 
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 443952 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] 
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 443952 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.c.SyncStrategy Sync 
replicas to 
https://127.0.0.1:46735/solr/hdfsbackuprestore_testok_shard1_replica_n1/
   [junit4]   2> 443953 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.PeerSync PeerSync: 
core=hdfsbackuprestore_testok_shard1_replica_n1 
url=https://127.0.0.1:46735/solr START 
replicas=[https://127.0.0.1:36659/solr/hdfsbackuprestore_testok_shard1_replica_n2/,
 https://127.0.0.1:46735/solr/hdfsbackuprestore_testok_shard1_replica_t4/] 
nUpdates=100
   [junit4]   2> 443954 INFO  (qtp2078506737-6918) [n:127.0.0.1:46735_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node3 
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.PeerSync PeerSync: 
core=hdfsbackuprestore_testok_shard1_replica_n1 
url=https://127.0.0.1:46735/solr DONE.  We have no versions.  sync failed.
   [junit4]   2> 443961 INFO  (qtp1840676713-6925) [n:127.0.0.1:36659_solr 
c:hdfsbackuprestore_testok s:shard1 r:core_node5 
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.c.S.Request [hdfsbac

[...truncated too long message...]

 loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml

resolve:

jar-checksums:
    [mkdir] Created dir: 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/null699052273
     [copy] Copying 249 files to 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/null699052273
   [delete] Deleting directory 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/null699052273

check-working-copy:
[ivy:cachepath] :: resolving dependencies :: #;working@lucene1-us-west
[ivy:cachepath]         confs: [default]
[ivy:cachepath]         found 
org.eclipse.jgit#org.eclipse.jgit;5.3.0.201903130848-r in public
[ivy:cachepath]         found com.jcraft#jsch;0.1.54 in public
[ivy:cachepath]         found com.jcraft#jzlib;1.1.1 in public
[ivy:cachepath]         found com.googlecode.javaewah#JavaEWAH;1.1.6 in public
[ivy:cachepath]         found org.slf4j#slf4j-api;1.7.2 in public
[ivy:cachepath]         found org.bouncycastle#bcpg-jdk15on;1.60 in public
[ivy:cachepath]         found org.bouncycastle#bcprov-jdk15on;1.60 in public
[ivy:cachepath]         found org.bouncycastle#bcpkix-jdk15on;1.60 in public
[ivy:cachepath]         found org.slf4j#slf4j-nop;1.7.2 in public
[ivy:cachepath] :: resolution report :: resolve 30ms :: artifacts dl 2ms
        ---------------------------------------------------------------------
        |                  |            modules            ||   artifacts   |
        |       conf       | number| search|dwnlded|evicted|| number|dwnlded|
        ---------------------------------------------------------------------
        |      default     |   9   |   0   |   0   |   0   ||   9   |   0   |
        ---------------------------------------------------------------------
[wc-checker] Initializing working copy...
[wc-checker] Checking working copy status...

-jenkins-base:

BUILD SUCCESSFUL
Total time: 118 minutes 58 seconds
Archiving artifacts
java.lang.InterruptedException: no matches found within 10000
        at hudson.FilePath$ValidateAntFileMask.hasMatch(FilePath.java:2847)
        at hudson.FilePath$ValidateAntFileMask.invoke(FilePath.java:2726)
        at hudson.FilePath$ValidateAntFileMask.invoke(FilePath.java:2707)
        at hudson.FilePath$FileCallableWrapper.call(FilePath.java:3086)
Also:   hudson.remoting.Channel$CallSiteStackTrace: Remote call to lucene
                at 
hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1741)
                at 
hudson.remoting.UserRequest$ExceptionResponse.retrieve(UserRequest.java:357)
                at hudson.remoting.Channel.call(Channel.java:955)
                at hudson.FilePath.act(FilePath.java:1072)
                at hudson.FilePath.act(FilePath.java:1061)
                at hudson.FilePath.validateAntFileMask(FilePath.java:2705)
                at 
hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
                at 
hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
                at 
hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
                at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
                at 
hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
                at hudson.model.Build$BuildExecution.post2(Build.java:186)
                at 
hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
                at hudson.model.Run.execute(Run.java:1835)
                at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
                at 
hudson.model.ResourceController.execute(ResourceController.java:97)
                at hudson.model.Executor.run(Executor.java:429)
Caused: hudson.FilePath$TunneledInterruptedException
        at hudson.FilePath$FileCallableWrapper.call(FilePath.java:3088)
        at hudson.remoting.UserRequest.perform(UserRequest.java:212)
        at hudson.remoting.UserRequest.perform(UserRequest.java:54)
        at hudson.remoting.Request$2.run(Request.java:369)
        at 
hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:744)
Caused: java.lang.InterruptedException: java.lang.InterruptedException: no 
matches found within 10000
        at hudson.FilePath.act(FilePath.java:1074)
        at hudson.FilePath.act(FilePath.java:1061)
        at hudson.FilePath.validateAntFileMask(FilePath.java:2705)
        at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
        at 
hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
        at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
        at hudson.model.Build$BuildExecution.post2(Build.java:186)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
        at hudson.model.Run.execute(Run.java:1835)
        at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
        at hudson.model.ResourceController.execute(ResourceController.java:97)
        at hudson.model.Executor.run(Executor.java:429)
No artifacts found that match the file pattern 
"**/*.events,heapdumps/**,**/hs_err_pid*". Configuration error?
Recording test results
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Email was triggered for: Unstable (Test Failures)
Sending email for trigger: Unstable (Test Failures)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to