Build: https://builds.apache.org/job/Lucene-Solr-Tests-8.x/491/
2 tests failed.
FAILED:
org.apache.solr.cloud.api.collections.TestHdfsCloudBackupRestore.testRestoreFailure
Error Message:
Failed collection is still in the clusterstate:
DocCollection(hdfsbackuprestore_testfailure_restored//collections/hdfsbackuprestore_testfailure_restored/state.json/2)={
"pullReplicas":0, "replicationFactor":1, "shards":{ "shard2":{
"range":"0-7fffffff", "state":"construction",
"replicas":{"core_node2":{
"core":"hdfsbackuprestore_testfailure_restored_shard2_replica_n1",
"base_url":"https://127.0.0.1:36659/solr",
"node_name":"127.0.0.1:36659_solr", "state":"down",
"type":"NRT", "force_set_state":"false"}},
"stateTimestamp":"1567059232049688251"}, "shard1":{
"range":"80000000-ffffffff", "state":"construction", "replicas":{},
"stateTimestamp":"1567059232049701653"}},
"router":{"name":"compositeId"}, "maxShardsPerNode":"1",
"autoAddReplicas":"false", "nrtReplicas":1, "tlogReplicas":0} Expected: not
a collection containing "hdfsbackuprestore_testfailure_restored" but: was
<[hdfsbackuprestore_testok, hdfsbackuprestore_testfailure_restored,
hdfsbackuprestore_testfailure, hdfsbackuprestore_testok_restored]>
Stack Trace:
java.lang.AssertionError: Failed collection is still in the clusterstate:
DocCollection(hdfsbackuprestore_testfailure_restored//collections/hdfsbackuprestore_testfailure_restored/state.json/2)={
"pullReplicas":0,
"replicationFactor":1,
"shards":{
"shard2":{
"range":"0-7fffffff",
"state":"construction",
"replicas":{"core_node2":{
"core":"hdfsbackuprestore_testfailure_restored_shard2_replica_n1",
"base_url":"https://127.0.0.1:36659/solr",
"node_name":"127.0.0.1:36659_solr",
"state":"down",
"type":"NRT",
"force_set_state":"false"}},
"stateTimestamp":"1567059232049688251"},
"shard1":{
"range":"80000000-ffffffff",
"state":"construction",
"replicas":{},
"stateTimestamp":"1567059232049701653"}},
"router":{"name":"compositeId"},
"maxShardsPerNode":"1",
"autoAddReplicas":"false",
"nrtReplicas":1,
"tlogReplicas":0}
Expected: not a collection containing "hdfsbackuprestore_testfailure_restored"
but: was <[hdfsbackuprestore_testok,
hdfsbackuprestore_testfailure_restored, hdfsbackuprestore_testfailure,
hdfsbackuprestore_testok_restored]>
at
__randomizedtesting.SeedInfo.seed([E037D74065656872:C94B49654D3C6B5F]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
at org.junit.Assert.assertThat(Assert.java:956)
at
org.apache.solr.cloud.api.collections.AbstractCloudBackupRestoreTestCase.testRestoreFailure(AbstractCloudBackupRestoreTestCase.java:211)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
FAILED:
org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore.testRestoreFailure
Error Message:
Failed collection is still in the clusterstate:
DocCollection(backuprestore_testfailure_restored//collections/backuprestore_testfailure_restored/state.json/2)={
"pullReplicas":0, "replicationFactor":1, "shards":{ "shard2":{
"range":"0-7fffffff", "state":"construction",
"replicas":{"core_node2":{
"core":"backuprestore_testfailure_restored_shard2_replica_n1",
"base_url":"http://127.0.0.1:33205/solr",
"node_name":"127.0.0.1:33205_solr", "state":"down",
"type":"NRT", "force_set_state":"false"}},
"stateTimestamp":"1567060879213084847"}, "shard1":{
"range":"80000000-ffffffff", "state":"construction", "replicas":{},
"stateTimestamp":"1567060879213099152"}},
"router":{"name":"compositeId"}, "maxShardsPerNode":"1",
"autoAddReplicas":"false", "nrtReplicas":1, "tlogReplicas":0} Expected: not
a collection containing "backuprestore_testfailure_restored" but: was
<[backuprestore_testok, backuprestore_testfailure,
backuprestore_testfailure_restored, backuprestore_testok_restored]>
Stack Trace:
java.lang.AssertionError: Failed collection is still in the clusterstate:
DocCollection(backuprestore_testfailure_restored//collections/backuprestore_testfailure_restored/state.json/2)={
"pullReplicas":0,
"replicationFactor":1,
"shards":{
"shard2":{
"range":"0-7fffffff",
"state":"construction",
"replicas":{"core_node2":{
"core":"backuprestore_testfailure_restored_shard2_replica_n1",
"base_url":"http://127.0.0.1:33205/solr",
"node_name":"127.0.0.1:33205_solr",
"state":"down",
"type":"NRT",
"force_set_state":"false"}},
"stateTimestamp":"1567060879213084847"},
"shard1":{
"range":"80000000-ffffffff",
"state":"construction",
"replicas":{},
"stateTimestamp":"1567060879213099152"}},
"router":{"name":"compositeId"},
"maxShardsPerNode":"1",
"autoAddReplicas":"false",
"nrtReplicas":1,
"tlogReplicas":0}
Expected: not a collection containing "backuprestore_testfailure_restored"
but: was <[backuprestore_testok, backuprestore_testfailure,
backuprestore_testfailure_restored, backuprestore_testok_restored]>
at
__randomizedtesting.SeedInfo.seed([E037D74065656872:C94B49654D3C6B5F]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
at org.junit.Assert.assertThat(Assert.java:956)
at
org.apache.solr.cloud.api.collections.AbstractCloudBackupRestoreTestCase.testRestoreFailure(AbstractCloudBackupRestoreTestCase.java:211)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)
Build Log:
[...truncated 13726 lines...]
[junit4] Suite:
org.apache.solr.cloud.api.collections.TestHdfsCloudBackupRestore
[junit4] 1> Formatting using clusterid: testClusterID
[junit4] 2> 439279 WARN
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[junit4] 2> 439296 WARN
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 439298 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.Server jetty-9.4.19.v20190610; built: 2019-06-10T16:30:51.723Z; git:
afcf563148970e98786327af5e07c261fda175d3; jvm 1.8.0_191-b12
[junit4] 2> 439300 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 439300 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 439300 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 439301 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@2774068b{static,/static,jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.0-tests.jar!/webapps/static,AVAILABLE}
[junit4] 2> 439459 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.w.WebAppContext@f007949{hdfs,/,file:///home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/jetty-localhost.localdomain-36239-hdfs-_-any-924387434669286531.dir/webapp/,AVAILABLE}{/hdfs}
[junit4] 2> 439460 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@7f6b887c{HTTP/1.1,[http/1.1]}{localhost.localdomain:36239}
[junit4] 2> 439461 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.Server Started @439527ms
[junit4] 2> 439553 WARN
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 439556 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.Server jetty-9.4.19.v20190610; built: 2019-06-10T16:30:51.723Z; git:
afcf563148970e98786327af5e07c261fda175d3; jvm 1.8.0_191-b12
[junit4] 2> 439556 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 439556 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 439557 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 439557 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@6adf3fad{static,/static,jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.0-tests.jar!/webapps/static,AVAILABLE}
[junit4] 2> 439714 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.w.WebAppContext@1c703108{datanode,/,file:///home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/jetty-localhost-37543-datanode-_-any-7314119788980653551.dir/webapp/,AVAILABLE}{/datanode}
[junit4] 2> 439715 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@7162d9c9{HTTP/1.1,[http/1.1]}{localhost:37543}
[junit4] 2> 439715 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.Server Started @439781ms
[junit4] 2> 439791 WARN
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 439792 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.Server jetty-9.4.19.v20190610; built: 2019-06-10T16:30:51.723Z; git:
afcf563148970e98786327af5e07c261fda175d3; jvm 1.8.0_191-b12
[junit4] 2> 439794 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 439794 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 439794 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 439795 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@11532006{static,/static,jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-3.2.0-tests.jar!/webapps/static,AVAILABLE}
[junit4] 2> 439974 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xcac9d599fa230d3: Processing first storage report for
DS-d29fa2ae-e164-4cca-aa51-f36bddc1bd73 from datanode
f9a9e1ed-6c2b-46ce-b8bb-7bae1b0f893d
[junit4] 2> 439974 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xcac9d599fa230d3: from storage
DS-d29fa2ae-e164-4cca-aa51-f36bddc1bd73 node
DatanodeRegistration(127.0.0.1:38477,
datanodeUuid=f9a9e1ed-6c2b-46ce-b8bb-7bae1b0f893d, infoPort=34367,
infoSecurePort=0, ipcPort=41639,
storageInfo=lv=-57;cid=testClusterID;nsid=968518402;c=1567059213337), blocks:
0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0
[junit4] 2> 439974 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xcac9d599fa230d3: Processing first storage report for
DS-1c4da6b9-2544-4f1f-b527-c4142a5267fd from datanode
f9a9e1ed-6c2b-46ce-b8bb-7bae1b0f893d
[junit4] 2> 439974 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xcac9d599fa230d3: from storage
DS-1c4da6b9-2544-4f1f-b527-c4142a5267fd node
DatanodeRegistration(127.0.0.1:38477,
datanodeUuid=f9a9e1ed-6c2b-46ce-b8bb-7bae1b0f893d, infoPort=34367,
infoSecurePort=0, ipcPort=41639,
storageInfo=lv=-57;cid=testClusterID;nsid=968518402;c=1567059213337), blocks:
0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0
[junit4] 2> 440012 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.w.WebAppContext@397cc67d{datanode,/,file:///home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/jetty-localhost-42969-datanode-_-any-8925702212772852981.dir/webapp/,AVAILABLE}{/datanode}
[junit4] 2> 440012 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.AbstractConnector Started
ServerConnector@4892d943{HTTP/1.1,[http/1.1]}{localhost:42969}
[junit4] 2> 440012 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.e.j.s.Server Started @440079ms
[junit4] 2> 440173 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xb57ad50ac595db20: Processing first storage report for
DS-ae78d8eb-dd57-4c19-ae6e-ea8f8519c130 from datanode
d9c8819b-1365-4c42-ae05-ffe965768d2c
[junit4] 2> 440173 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xb57ad50ac595db20: from storage
DS-ae78d8eb-dd57-4c19-ae6e-ea8f8519c130 node
DatanodeRegistration(127.0.0.1:42413,
datanodeUuid=d9c8819b-1365-4c42-ae05-ffe965768d2c, infoPort=40741,
infoSecurePort=0, ipcPort=34355,
storageInfo=lv=-57;cid=testClusterID;nsid=968518402;c=1567059213337), blocks:
0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0
[junit4] 2> 440173 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xb57ad50ac595db20: Processing first storage report for
DS-ef48ae90-a08b-4a82-9795-00787d190e45 from datanode
d9c8819b-1365-4c42-ae05-ffe965768d2c
[junit4] 2> 440173 INFO (Block report processor) [ ] BlockStateChange
BLOCK* processReport 0xb57ad50ac595db20: from storage
DS-ef48ae90-a08b-4a82-9795-00787d190e45 node
DatanodeRegistration(127.0.0.1:42413,
datanodeUuid=d9c8819b-1365-4c42-ae05-ffe965768d2c, infoPort=40741,
infoSecurePort=0, ipcPort=34355,
storageInfo=lv=-57;cid=testClusterID;nsid=968518402;c=1567059213337), blocks:
0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0
[junit4] 2> 440259 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.MiniSolrCloudCluster Starting cluster of 2 servers in
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002
[junit4] 2> 440260 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 440260 INFO (ZkTestServer Run Thread) [ ]
o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 440260 INFO (ZkTestServer Run Thread) [ ]
o.a.s.c.ZkTestServer Starting server
[junit4] 2> 440360 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.ZkTestServer start zk server on port:45147
[junit4] 2> 440360 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.ZkTestServer waitForServerUp: 127.0.0.1:45147
[junit4] 2> 440360 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.ZkTestServer parse host and port list: 127.0.0.1:45147
[junit4] 2> 440360 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.ZkTestServer connecting to 127.0.0.1 45147
[junit4] 2> 440363 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
[junit4] 2> 440367 INFO (zkConnectionManagerCallback-2523-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 440367 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
[junit4] 2> 440371 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
[junit4] 2> 440372 INFO (zkConnectionManagerCallback-2525-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 440372 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
[junit4] 2> 440376 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
[junit4] 2> 440378 INFO (zkConnectionManagerCallback-2527-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 440378 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
[junit4] 2> 440486 WARN (jetty-launcher-2528-thread-1) [ ]
o.e.j.s.AbstractConnector Ignoring deprecated socket close linger time
[junit4] 2> 440486 WARN (jetty-launcher-2528-thread-2) [ ]
o.e.j.s.AbstractConnector Ignoring deprecated socket close linger time
[junit4] 2> 440487 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.c.s.e.JettySolrRunner Start Jetty (original configured port=0)
[junit4] 2> 440487 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.c.s.e.JettySolrRunner Start Jetty (original configured port=0)
[junit4] 2> 440487 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.c.s.e.JettySolrRunner Trying to start Jetty on port 0 try number 1 ...
[junit4] 2> 440487 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.c.s.e.JettySolrRunner Trying to start Jetty on port 0 try number 1 ...
[junit4] 2> 440487 INFO (jetty-launcher-2528-thread-2) [ ]
o.e.j.s.Server jetty-9.4.19.v20190610; built: 2019-06-10T16:30:51.723Z; git:
afcf563148970e98786327af5e07c261fda175d3; jvm 1.8.0_191-b12
[junit4] 2> 440487 INFO (jetty-launcher-2528-thread-1) [ ]
o.e.j.s.Server jetty-9.4.19.v20190610; built: 2019-06-10T16:30:51.723Z; git:
afcf563148970e98786327af5e07c261fda175d3; jvm 1.8.0_191-b12
[junit4] 2> 440495 INFO (jetty-launcher-2528-thread-1) [ ]
o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 440495 INFO (jetty-launcher-2528-thread-1) [ ]
o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 440496 INFO (jetty-launcher-2528-thread-1) [ ]
o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 440496 INFO (jetty-launcher-2528-thread-2) [ ]
o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 440496 INFO (jetty-launcher-2528-thread-2) [ ]
o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 440496 INFO (jetty-launcher-2528-thread-2) [ ]
o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 440496 INFO (jetty-launcher-2528-thread-1) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@45e0eb05{/solr,null,AVAILABLE}
[junit4] 2> 440496 INFO (jetty-launcher-2528-thread-2) [ ]
o.e.j.s.h.ContextHandler Started
o.e.j.s.ServletContextHandler@2f3604af{/solr,null,AVAILABLE}
[junit4] 2> 440497 INFO (jetty-launcher-2528-thread-2) [ ]
o.e.j.s.AbstractConnector Started ServerConnector@224c8694{SSL,[ssl,
http/1.1]}{127.0.0.1:36659}
[junit4] 2> 440497 INFO (jetty-launcher-2528-thread-2) [ ]
o.e.j.s.Server Started @440564ms
[junit4] 2> 440497 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr,
hostPort=36659}
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-1) [ ]
o.e.j.s.AbstractConnector Started ServerConnector@341995c6{SSL,[ssl,
http/1.1]}{127.0.0.1:46735}
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-1) [ ]
o.e.j.s.Server Started @440564ms
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr,
hostPort=46735}
[junit4] 2> 440498 ERROR (jetty-launcher-2528-thread-2) [ ]
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.s.SolrDispatchFilter Using logger factory
org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 440498 ERROR (jetty-launcher-2528-thread-1) [ ]
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be
missing or incomplete.
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr? version
8.3.0
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.s.SolrDispatchFilter Using logger factory
org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr? version
8.3.0
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2019-08-29T06:13:34.605Z
[junit4] 2> 440498 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time:
2019-08-29T06:13:34.605Z
[junit4] 2> 440500 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
[junit4] 2> 440503 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
[junit4] 2> 440503 INFO (zkConnectionManagerCallback-2530-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 440503 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
[junit4] 2> 440504 INFO (zkConnectionManagerCallback-2532-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 440504 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
[junit4] 2> 440504 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 440505 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 440523 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.c.SolrXmlConfig MBean server found:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267, but no JMX reporters were
configured - adding default JMX reporter.
[junit4] 2> 440530 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.c.SolrXmlConfig MBean server found:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267, but no JMX reporters were
configured - adding default JMX reporter.
[junit4] 2> 440887 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.h.c.HttpShardHandlerFactory Host whitelist initialized:
WhitelistHostChecker [whitelistHosts=null, whitelistHostCheckingEnabled=true]
[junit4] 2> 440888 WARN (jetty-launcher-2528-thread-2) [ ]
o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport
since Java 8 or lower versions does not support SSL + HTTP/2
[junit4] 2> 440889 WARN (jetty-launcher-2528-thread-2) [ ]
o.e.j.u.s.S.config Trusting all certificates configured for
Client@4b3ef0b7[provider=null,keyStore=null,trustStore=null]
[junit4] 2> 440889 WARN (jetty-launcher-2528-thread-2) [ ]
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for
Client@4b3ef0b7[provider=null,keyStore=null,trustStore=null]
[junit4] 2> 440890 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.h.c.HttpShardHandlerFactory Host whitelist initialized:
WhitelistHostChecker [whitelistHosts=null, whitelistHostCheckingEnabled=true]
[junit4] 2> 440891 WARN (jetty-launcher-2528-thread-1) [ ]
o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport
since Java 8 or lower versions does not support SSL + HTTP/2
[junit4] 2> 440893 WARN (jetty-launcher-2528-thread-1) [ ]
o.e.j.u.s.S.config Trusting all certificates configured for
Client@7637eb82[provider=null,keyStore=null,trustStore=null]
[junit4] 2> 440893 WARN (jetty-launcher-2528-thread-1) [ ]
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for
Client@7637eb82[provider=null,keyStore=null,trustStore=null]
[junit4] 2> 440895 WARN (jetty-launcher-2528-thread-2) [ ]
o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport
since Java 8 or lower versions does not support SSL + HTTP/2
[junit4] 2> 440901 WARN (jetty-launcher-2528-thread-1) [ ]
o.a.s.c.s.i.Http2SolrClient Create Http2SolrClient with HTTP/1.1 transport
since Java 8 or lower versions does not support SSL + HTTP/2
[junit4] 2> 440901 WARN (jetty-launcher-2528-thread-2) [ ]
o.e.j.u.s.S.config Trusting all certificates configured for
Client@74bd9149[provider=null,keyStore=null,trustStore=null]
[junit4] 2> 440901 WARN (jetty-launcher-2528-thread-2) [ ]
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for
Client@74bd9149[provider=null,keyStore=null,trustStore=null]
[junit4] 2> 440902 WARN (jetty-launcher-2528-thread-1) [ ]
o.e.j.u.s.S.config Trusting all certificates configured for
Client@18e664b8[provider=null,keyStore=null,trustStore=null]
[junit4] 2> 440902 WARN (jetty-launcher-2528-thread-1) [ ]
o.e.j.u.s.S.config No Client EndPointIdentificationAlgorithm configured for
Client@18e664b8[provider=null,keyStore=null,trustStore=null]
[junit4] 2> 440903 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45147/solr
[junit4] 2> 440903 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:45147/solr
[junit4] 2> 440905 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
[junit4] 2> 440906 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
[junit4] 2> 440906 INFO (zkConnectionManagerCallback-2546-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 440906 INFO (jetty-launcher-2528-thread-1) [ ]
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
[junit4] 2> 440911 INFO (zkConnectionManagerCallback-2544-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 440911 INFO (jetty-launcher-2528-thread-2) [ ]
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
[junit4] 2> 441011 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.c.ConnectionManager Waiting for client to
connect to ZooKeeper
[junit4] 2> 441014 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.c.ConnectionManager Waiting for client to
connect to ZooKeeper
[junit4] 2> 441014 INFO (zkConnectionManagerCallback-2548-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 441014 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.c.ConnectionManager Client is connected
to ZooKeeper
[junit4] 2> 441024 INFO (zkConnectionManagerCallback-2550-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 441024 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.c.ConnectionManager Client is connected
to ZooKeeper
[junit4] 2> 441203 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.OverseerElectionContext I am going to be
the leader 127.0.0.1:46735_solr
[junit4] 2> 441205 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.Overseer Overseer
(id=72285712308305927-127.0.0.1:46735_solr-n_0000000000) starting
[junit4] 2> 441223 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.c.ConnectionManager Waiting for client to
connect to ZooKeeper
[junit4] 2> 441231 INFO (zkConnectionManagerCallback-2559-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 441231 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.c.ConnectionManager Client is connected
to ZooKeeper
[junit4] 2> 441237 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster
at 127.0.0.1:45147/solr ready
[junit4] 2> 441243 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:36659_solr
[junit4] 2> 441246 INFO
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000)
[n:127.0.0.1:46735_solr ] o.a.s.c.Overseer Starting to work on the main
queue : 127.0.0.1:46735_solr
[junit4] 2> 441247 INFO
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000)
[n:127.0.0.1:46735_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (1)
[junit4] 2> 441257 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.ZkController Publish
node=127.0.0.1:46735_solr as DOWN
[junit4] 2> 441259 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating
transient cache for 2147483647 transient cores
[junit4] 2> 441259 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.ZkController Register node as live in
ZooKeeper:/live_nodes/127.0.0.1:46735_solr
[junit4] 2> 441262 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.PackageManager clusterprops.json changed
, version 0
[junit4] 2> 441262 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.b.r.BackupRepositoryFactory Added backup
repository with configuration params {type = repository,name = hdfs,class =
org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes =
{name=hdfs,
class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args =
{location=/backup,solr.hdfs.home=hdfs://localhost.localdomain:46481/solr,solr.hdfs.confdir=}}
[junit4] 2> 441262 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.b.r.BackupRepositoryFactory Added backup
repository with configuration params {type = repository,name = poisioned,class
=
org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes
= {default=true, name=poisioned,
class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
[junit4] 2> 441262 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.b.r.BackupRepositoryFactory Default
configuration for backup repository is with configuration params {type =
repository,name = poisioned,class =
org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes
= {default=true, name=poisioned,
class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
[junit4] 2> 441267 INFO (zkCallback-2547-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 441274 INFO (zkCallback-2558-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 441277 INFO (zkCallback-2549-thread-1) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 441280 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.c.ConnectionManager Waiting for client to
connect to ZooKeeper
[junit4] 2> 441301 INFO (zkConnectionManagerCallback-2564-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 441301 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.c.ConnectionManager Client is connected
to ZooKeeper
[junit4] 2> 441302 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from
ZooKeeper... (0) -> (2)
[junit4] 2> 441305 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster
at 127.0.0.1:45147/solr ready
[junit4] 2> 441306 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.PackageManager clusterprops.json changed
, version 0
[junit4] 2> 441306 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.b.r.BackupRepositoryFactory Added backup
repository with configuration params {type = repository,name = hdfs,class =
org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes =
{name=hdfs,
class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args =
{location=/backup,solr.hdfs.home=hdfs://localhost.localdomain:46481/solr,solr.hdfs.confdir=}}
[junit4] 2> 441306 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.b.r.BackupRepositoryFactory Added backup
repository with configuration params {type = repository,name = poisioned,class
=
org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes
= {default=true, name=poisioned,
class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
[junit4] 2> 441306 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.b.r.BackupRepositoryFactory Default
configuration for backup repository is with configuration params {type =
repository,name = poisioned,class =
org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository,attributes
= {default=true, name=poisioned,
class=org.apache.solr.cloud.api.collections.TestLocalFSCloudBackupRestore$PoinsionedRepository},}
[junit4] 2> 441332 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.h.a.MetricsHistoryHandler No .system
collection, keeping metrics history in memory.
[junit4] 2> 441378 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.h.a.MetricsHistoryHandler No .system
collection, keeping metrics history in memory.
[junit4] 2> 441409 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.node' (registry 'solr.node') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 441436 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.node' (registry 'solr.node') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 441444 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jvm' (registry 'solr.jvm') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 441444 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jetty' (registry 'solr.jetty') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 441445 INFO (jetty-launcher-2528-thread-1)
[n:127.0.0.1:46735_solr ] o.a.s.c.CorePropertiesLocator Found 0 core
definitions underneath
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/.
[junit4] 2> 441454 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jvm' (registry 'solr.jvm') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 441454 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for
'solr.jetty' (registry 'solr.jetty') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 441456 INFO (jetty-launcher-2528-thread-2)
[n:127.0.0.1:36659_solr ] o.a.s.c.CorePropertiesLocator Found 0 core
definitions underneath
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/.
[junit4] 2> 441578 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.MiniSolrCloudCluster waitForAllNodes: numServers=2
[junit4] 2> 441579 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
[junit4] 2> 441584 INFO (zkConnectionManagerCallback-2571-thread-1) [
] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 441585 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
[junit4] 2> 441588 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 441607 INFO
(SUITE-TestHdfsCloudBackupRestore-seed#[E037D74065656872]-worker) [ ]
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:45147/solr ready
[junit4] 2> 441715 INFO (qtp1840676713-6927) [n:127.0.0.1:36659_solr
] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params
collection.configName=conf1&router.name=implicit&version=2&pullReplicas=0&shards=shard1,shard2&property.customKey=customValue&maxShardsPerNode=3&router.field=shard_s&autoAddReplicas=true&name=hdfsbackuprestore_testok&nrtReplicas=2&action=CREATE&tlogReplicas=1&wt=javabin
and sendToOCPQueue=true
[junit4] 2> 441723 INFO
(OverseerThreadFactory-1679-thread-1-processing-n:127.0.0.1:46735_solr)
[n:127.0.0.1:46735_solr ] o.a.s.c.a.c.CreateCollectionCmd Create collection
hdfsbackuprestore_testok
[junit4] 2> 441830 WARN
(OverseerThreadFactory-1679-thread-1-processing-n:127.0.0.1:46735_solr)
[n:127.0.0.1:46735_solr ] o.a.s.c.a.c.CreateCollectionCmd Specified number
of replicas of 3 on collection hdfsbackuprestore_testok is higher than the
number of Solr instances currently live or live and part of your
createNodeSet(2). It's unusual to run two replica of the same slice on the same
Solr-instance.
[junit4] 2> 441836 INFO
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000)
[n:127.0.0.1:46735_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"hdfsbackuprestore_testok",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"hdfsbackuprestore_testok_shard1_replica_n1",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:46735/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 441841 INFO
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000)
[n:127.0.0.1:46735_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"hdfsbackuprestore_testok",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"hdfsbackuprestore_testok_shard1_replica_n2",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:36659/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 441847 INFO
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000)
[n:127.0.0.1:46735_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"hdfsbackuprestore_testok",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"hdfsbackuprestore_testok_shard1_replica_t4",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:46735/solr",
[junit4] 2> "type":"TLOG",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 441852 INFO
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000)
[n:127.0.0.1:46735_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"hdfsbackuprestore_testok",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"hdfsbackuprestore_testok_shard2_replica_n7",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:36659/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 441855 INFO
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000)
[n:127.0.0.1:46735_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"hdfsbackuprestore_testok",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"hdfsbackuprestore_testok_shard2_replica_n8",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:46735/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 441859 INFO
(OverseerStateUpdate-72285712308305927-127.0.0.1:46735_solr-n_0000000000)
[n:127.0.0.1:46735_solr ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"hdfsbackuprestore_testok",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"hdfsbackuprestore_testok_shard2_replica_t10",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:36659/solr",
[junit4] 2> "type":"TLOG",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 442066 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.h.a.CoreAdminOperation
core create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node5&name=hdfsbackuprestore_testok_shard1_replica_n2&action=CREATE&numShards=2&shard=shard1&wt=javabin
[junit4] 2> 442066 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
x:hdfsbackuprestore_testok_shard1_replica_n2 ]
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647
transient cores
[junit4] 2> 442078 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.h.a.CoreAdminOperation
core create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node9&name=hdfsbackuprestore_testok_shard2_replica_n7&action=CREATE&numShards=2&shard=shard2&wt=javabin
[junit4] 2> 442084 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.h.a.CoreAdminOperation
core create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=TLOG&property.customKey=customValue&coreNodeName=core_node12&name=hdfsbackuprestore_testok_shard2_replica_t10&action=CREATE&numShards=2&shard=shard2&wt=javabin
[junit4] 2> 442098 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.h.a.CoreAdminOperation
core create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=TLOG&property.customKey=customValue&coreNodeName=core_node6&name=hdfsbackuprestore_testok_shard1_replica_t4&action=CREATE&numShards=2&shard=shard1&wt=javabin
[junit4] 2> 442101 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.h.a.CoreAdminOperation
core create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node3&name=hdfsbackuprestore_testok_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&wt=javabin
[junit4] 2> 442112 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.h.a.CoreAdminOperation
core create command
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore_testok&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node11&name=hdfsbackuprestore_testok_shard2_replica_n8&action=CREATE&numShards=2&shard=shard2&wt=javabin
[junit4] 2> 443135 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.c.SolrConfig Using Lucene
MatchVersion: 8.3.0
[junit4] 2> 443135 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.c.SolrConfig Using Lucene
MatchVersion: 8.3.0
[junit4] 2> 443138 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.c.SolrConfig Using Lucene
MatchVersion: 8.3.0
[junit4] 2> 443144 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.c.SolrConfig Using Lucene
MatchVersion: 8.3.0
[junit4] 2> 443144 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.c.SolrConfig Using Lucene
MatchVersion: 8.3.0
[junit4] 2> 443151 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.c.SolrConfig Using Lucene
MatchVersion: 8.3.0
[junit4] 2> 443188 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.s.IndexSchema
[hdfsbackuprestore_testok_shard1_replica_n1] Schema name=minimal
[junit4] 2> 443204 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.s.IndexSchema
[hdfsbackuprestore_testok_shard1_replica_t4] Schema name=minimal
[junit4] 2> 443207 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.s.IndexSchema Loaded
schema minimal/1.1 with uniqueid field id
[junit4] 2> 443208 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.c.CoreContainer Creating
SolrCore 'hdfsbackuprestore_testok_shard1_replica_t4' using configuration from
collection hdfsbackuprestore_testok, trusted=true
[junit4] 2> 443208 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.m.r.SolrJmxReporter JMX
monitoring for 'solr.core.hdfsbackuprestore_testok.shard1.replica_t4' (registry
'solr.core.hdfsbackuprestore_testok.shard1.replica_t4') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 443215 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.s.IndexSchema
[hdfsbackuprestore_testok_shard2_replica_n7] Schema name=minimal
[junit4] 2> 443218 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.s.IndexSchema
[hdfsbackuprestore_testok_shard1_replica_n2] Schema name=minimal
[junit4] 2> 443218 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.s.IndexSchema Loaded
schema minimal/1.1 with uniqueid field id
[junit4] 2> 443218 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.c.CoreContainer Creating
SolrCore 'hdfsbackuprestore_testok_shard2_replica_n7' using configuration from
collection hdfsbackuprestore_testok, trusted=true
[junit4] 2> 443219 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.m.r.SolrJmxReporter JMX
monitoring for 'solr.core.hdfsbackuprestore_testok.shard2.replica_n7' (registry
'solr.core.hdfsbackuprestore_testok.shard2.replica_n7') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 443222 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.s.IndexSchema
[hdfsbackuprestore_testok_shard2_replica_t10] Schema name=minimal
[junit4] 2> 443226 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.s.IndexSchema Loaded
schema minimal/1.1 with uniqueid field id
[junit4] 2> 443226 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.c.CoreContainer Creating
SolrCore 'hdfsbackuprestore_testok_shard1_replica_n1' using configuration from
collection hdfsbackuprestore_testok, trusted=true
[junit4] 2> 443227 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.m.r.SolrJmxReporter JMX
monitoring for 'solr.core.hdfsbackuprestore_testok.shard1.replica_n1' (registry
'solr.core.hdfsbackuprestore_testok.shard1.replica_n1') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 443230 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.s.IndexSchema Loaded
schema minimal/1.1 with uniqueid field id
[junit4] 2> 443231 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.c.CoreContainer Creating
SolrCore 'hdfsbackuprestore_testok_shard1_replica_n2' using configuration from
collection hdfsbackuprestore_testok, trusted=true
[junit4] 2> 443231 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.c.SolrCore
[[hdfsbackuprestore_testok_shard2_replica_n7] ] Opening new SolrCore at
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/hdfsbackuprestore_testok_shard2_replica_n7],
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/./hdfsbackuprestore_testok_shard2_replica_n7/data/]
[junit4] 2> 443232 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.m.r.SolrJmxReporter JMX
monitoring for 'solr.core.hdfsbackuprestore_testok.shard1.replica_n2' (registry
'solr.core.hdfsbackuprestore_testok.shard1.replica_n2') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 443232 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.c.SolrCore
[[hdfsbackuprestore_testok_shard1_replica_n2] ] Opening new SolrCore at
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/hdfsbackuprestore_testok_shard1_replica_n2],
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/./hdfsbackuprestore_testok_shard1_replica_n2/data/]
[junit4] 2> 443233 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.c.SolrCore
[[hdfsbackuprestore_testok_shard1_replica_t4] ] Opening new SolrCore at
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/hdfsbackuprestore_testok_shard1_replica_t4],
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/./hdfsbackuprestore_testok_shard1_replica_t4/data/]
[junit4] 2> 443234 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.s.IndexSchema Loaded
schema minimal/1.1 with uniqueid field id
[junit4] 2> 443234 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.c.CoreContainer Creating
SolrCore 'hdfsbackuprestore_testok_shard2_replica_t10' using configuration from
collection hdfsbackuprestore_testok, trusted=true
[junit4] 2> 443235 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.m.r.SolrJmxReporter JMX
monitoring for 'solr.core.hdfsbackuprestore_testok.shard2.replica_t10'
(registry 'solr.core.hdfsbackuprestore_testok.shard2.replica_t10') enabled at
server: com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 443235 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.c.SolrCore
[[hdfsbackuprestore_testok_shard2_replica_t10] ] Opening new SolrCore at
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/hdfsbackuprestore_testok_shard2_replica_t10],
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node2/./hdfsbackuprestore_testok_shard2_replica_t10/data/]
[junit4] 2> 443236 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.c.SolrCore
[[hdfsbackuprestore_testok_shard1_replica_n1] ] Opening new SolrCore at
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/hdfsbackuprestore_testok_shard1_replica_n1],
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/./hdfsbackuprestore_testok_shard1_replica_n1/data/]
[junit4] 2> 443240 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.s.IndexSchema
[hdfsbackuprestore_testok_shard2_replica_n8] Schema name=minimal
[junit4] 2> 443243 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.s.IndexSchema Loaded
schema minimal/1.1 with uniqueid field id
[junit4] 2> 443243 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.c.CoreContainer Creating
SolrCore 'hdfsbackuprestore_testok_shard2_replica_n8' using configuration from
collection hdfsbackuprestore_testok, trusted=true
[junit4] 2> 443244 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.m.r.SolrJmxReporter JMX
monitoring for 'solr.core.hdfsbackuprestore_testok.shard2.replica_n8' (registry
'solr.core.hdfsbackuprestore_testok.shard2.replica_n8') enabled at server:
com.sun.jmx.mbeanserver.JmxMBeanServer@481a4267
[junit4] 2> 443244 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.c.SolrCore
[[hdfsbackuprestore_testok_shard2_replica_n8] ] Opening new SolrCore at
[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/hdfsbackuprestore_testok_shard2_replica_n8],
dataDir=[/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/build/solr-core/test/J0/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_E037D74065656872-001/tempDir-002/node1/./hdfsbackuprestore_testok_shard2_replica_n8/data/]
[junit4] 2> 443410 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.UpdateHandler Using
UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 443410 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.UpdateLog Initializing
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 443412 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.CommitTracker Hard
AutoCommit: disabled
[junit4] 2> 443412 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.CommitTracker Soft
AutoCommit: disabled
[junit4] 2> 443418 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.s.SolrIndexSearcher
Opening [Searcher@5d6f00f3[hdfsbackuprestore_testok_shard1_replica_n1] main]
[junit4] 2> 443427 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.r.ManagedResourceStorage
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 443428 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.r.ManagedResourceStorage
Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 443432 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.h.ReplicationHandler
Commits will be reserved for 10000ms.
[junit4] 2> 443432 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.UpdateLog Could not find
max version in index or recent updates, using new clock 1643180686090174464
[junit4] 2> 443440 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.c.ZkShardTerms Successful
update of terms at /collections/hdfsbackuprestore_testok/terms/shard1 to
Terms{values={core_node3=0}, version=0}
[junit4] 2> 443441 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ]
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created
/collections/hdfsbackuprestore_testok/leaders/shard1
[junit4] 2> 443444 INFO
(searcherExecutor-1690-thread-1-processing-n:127.0.0.1:46735_solr
x:hdfsbackuprestore_testok_shard1_replica_n1 c:hdfsbackuprestore_testok
s:shard1 r:core_node3) [n:127.0.0.1:46735_solr c:hdfsbackuprestore_testok
s:shard1 r:core_node3 x:hdfsbackuprestore_testok_shard1_replica_n1 ]
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard1_replica_n1] Registered new
searcher Searcher@5d6f00f3[hdfsbackuprestore_testok_shard1_replica_n1]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 443449 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ]
o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for
shard shard1: total=3 found=1 timeoutin=9999ms
[junit4] 2> 443462 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.u.UpdateHandler Using
UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 443462 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.u.UpdateLog Initializing
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 443462 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.u.UpdateHandler Using
UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 443462 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.u.UpdateLog Initializing
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 443463 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.u.CommitTracker Hard
AutoCommit: disabled
[junit4] 2> 443463 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.u.CommitTracker Soft
AutoCommit: disabled
[junit4] 2> 443463 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.u.CommitTracker Hard
AutoCommit: disabled
[junit4] 2> 443464 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.u.CommitTracker Soft
AutoCommit: disabled
[junit4] 2> 443466 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.u.UpdateHandler Using
UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 443466 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.u.UpdateLog Initializing
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 443467 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.s.SolrIndexSearcher
Opening [Searcher@7add75ac[hdfsbackuprestore_testok_shard1_replica_n2] main]
[junit4] 2> 443467 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.u.CommitTracker Hard
AutoCommit: disabled
[junit4] 2> 443467 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.u.CommitTracker Soft
AutoCommit: disabled
[junit4] 2> 443470 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.s.SolrIndexSearcher
Opening [Searcher@abe7412[hdfsbackuprestore_testok_shard1_replica_t4] main]
[junit4] 2> 443473 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.r.ManagedResourceStorage
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 443475 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.r.ManagedResourceStorage
Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 443475 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.h.ReplicationHandler
Commits will be reserved for 10000ms.
[junit4] 2> 443476 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.u.UpdateLog Could not find
max version in index or recent updates, using new clock 1643180686136311808
[junit4] 2> 443484 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.r.ManagedResourceStorage
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 443484 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.s.SolrIndexSearcher
Opening [Searcher@40bbb2b8[hdfsbackuprestore_testok_shard2_replica_n8] main]
[junit4] 2> 443485 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.r.ManagedResourceStorage
Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 443485 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.h.ReplicationHandler
Commits will be reserved for 10000ms.
[junit4] 2> 443485 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.u.UpdateLog Could not find
max version in index or recent updates, using new clock 1643180686145748992
[junit4] 2> 443487 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.r.ManagedResourceStorage
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 443488 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.r.ManagedResourceStorage
Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 443488 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.h.ReplicationHandler
Commits will be reserved for 10000ms.
[junit4] 2> 443489 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.u.UpdateLog Could not find
max version in index or recent updates, using new clock 1643180686148894720
[junit4] 2> 443494 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.u.UpdateHandler Using
UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 443494 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.u.UpdateLog Initializing
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 443494 INFO
(searcherExecutor-1691-thread-1-processing-n:127.0.0.1:36659_solr
x:hdfsbackuprestore_testok_shard1_replica_n2 c:hdfsbackuprestore_testok
s:shard1 r:core_node5) [n:127.0.0.1:36659_solr c:hdfsbackuprestore_testok
s:shard1 r:core_node5 x:hdfsbackuprestore_testok_shard1_replica_n2 ]
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard1_replica_n2] Registered new
searcher Searcher@7add75ac[hdfsbackuprestore_testok_shard1_replica_n2]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 443495 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.u.CommitTracker Hard
AutoCommit: disabled
[junit4] 2> 443495 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.u.CommitTracker Soft
AutoCommit: disabled
[junit4] 2> 443496 INFO
(searcherExecutor-1693-thread-1-processing-n:127.0.0.1:46735_solr
x:hdfsbackuprestore_testok_shard2_replica_n8 c:hdfsbackuprestore_testok
s:shard2 r:core_node11) [n:127.0.0.1:46735_solr c:hdfsbackuprestore_testok
s:shard2 r:core_node11 x:hdfsbackuprestore_testok_shard2_replica_n8 ]
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard2_replica_n8] Registered new
searcher Searcher@40bbb2b8[hdfsbackuprestore_testok_shard2_replica_n8]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 443497 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ] o.a.s.c.ZkShardTerms Successful
update of terms at /collections/hdfsbackuprestore_testok/terms/shard1 to
Terms{values={core_node6=0, core_node3=0}, version=1}
[junit4] 2> 443498 INFO (qtp2078506737-6920) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node6
x:hdfsbackuprestore_testok_shard1_replica_t4 ]
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created
/collections/hdfsbackuprestore_testok/leaders/shard1
[junit4] 2> 443500 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.s.SolrIndexSearcher
Opening [Searcher@69173d6[hdfsbackuprestore_testok_shard2_replica_t10] main]
[junit4] 2> 443501 INFO
(searcherExecutor-1688-thread-1-processing-n:127.0.0.1:46735_solr
x:hdfsbackuprestore_testok_shard1_replica_t4 c:hdfsbackuprestore_testok
s:shard1 r:core_node6) [n:127.0.0.1:46735_solr c:hdfsbackuprestore_testok
s:shard1 r:core_node6 x:hdfsbackuprestore_testok_shard1_replica_t4 ]
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard1_replica_t4] Registered new
searcher Searcher@abe7412[hdfsbackuprestore_testok_shard1_replica_t4]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 443503 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.r.ManagedResourceStorage
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 443504 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.r.ManagedResourceStorage
Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 443505 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.h.ReplicationHandler
Commits will be reserved for 10000ms.
[junit4] 2> 443505 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.u.UpdateLog Could not
find max version in index or recent updates, using new clock 1643180686166720512
[junit4] 2> 443505 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.c.ZkShardTerms Successful
update of terms at /collections/hdfsbackuprestore_testok/terms/shard1 to
Terms{values={core_node6=0, core_node3=0, core_node5=0}, version=2}
[junit4] 2> 443505 INFO (qtp1840676713-6923) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ]
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created
/collections/hdfsbackuprestore_testok/leaders/shard1
[junit4] 2> 443514 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ] o.a.s.c.ZkShardTerms Successful
update of terms at /collections/hdfsbackuprestore_testok/terms/shard2 to
Terms{values={core_node11=0}, version=0}
[junit4] 2> 443514 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ]
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created
/collections/hdfsbackuprestore_testok/leaders/shard2
[junit4] 2> 443520 INFO
(searcherExecutor-1692-thread-1-processing-n:127.0.0.1:36659_solr
x:hdfsbackuprestore_testok_shard2_replica_t10 c:hdfsbackuprestore_testok
s:shard2 r:core_node12) [n:127.0.0.1:36659_solr c:hdfsbackuprestore_testok
s:shard2 r:core_node12 x:hdfsbackuprestore_testok_shard2_replica_t10 ]
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard2_replica_t10] Registered new
searcher Searcher@69173d6[hdfsbackuprestore_testok_shard2_replica_t10]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 443521 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ] o.a.s.c.ZkShardTerms Successful
update of terms at /collections/hdfsbackuprestore_testok/terms/shard2 to
Terms{values={core_node12=0, core_node11=0}, version=1}
[junit4] 2> 443524 INFO (qtp2078506737-6924) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node11
x:hdfsbackuprestore_testok_shard2_replica_n8 ]
o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for
shard shard2: total=3 found=1 timeoutin=9998ms
[junit4] 2> 443530 INFO (qtp1840676713-6921) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node12
x:hdfsbackuprestore_testok_shard2_replica_t10 ]
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created
/collections/hdfsbackuprestore_testok/leaders/shard2
[junit4] 2> 443530 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.u.UpdateHandler Using
UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 443530 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.u.UpdateLog Initializing
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100
maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 443532 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.u.CommitTracker Hard
AutoCommit: disabled
[junit4] 2> 443532 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.u.CommitTracker Soft
AutoCommit: disabled
[junit4] 2> 443536 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.s.SolrIndexSearcher
Opening [Searcher@50d2c5a[hdfsbackuprestore_testok_shard2_replica_n7] main]
[junit4] 2> 443538 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.r.ManagedResourceStorage
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 443538 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.r.ManagedResourceStorage
Loaded null at path _rest_managed.json using
ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 443539 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.h.ReplicationHandler
Commits will be reserved for 10000ms.
[junit4] 2> 443539 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.u.UpdateLog Could not find
max version in index or recent updates, using new clock 1643180686202372096
[junit4] 2> 443545 INFO
(searcherExecutor-1689-thread-1-processing-n:127.0.0.1:36659_solr
x:hdfsbackuprestore_testok_shard2_replica_n7 c:hdfsbackuprestore_testok
s:shard2 r:core_node9) [n:127.0.0.1:36659_solr c:hdfsbackuprestore_testok
s:shard2 r:core_node9 x:hdfsbackuprestore_testok_shard2_replica_n7 ]
o.a.s.c.SolrCore [hdfsbackuprestore_testok_shard2_replica_n7] Registered new
searcher Searcher@50d2c5a[hdfsbackuprestore_testok_shard2_replica_n7]
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 443546 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ] o.a.s.c.ZkShardTerms Successful
update of terms at /collections/hdfsbackuprestore_testok/terms/shard2 to
Terms{values={core_node12=0, core_node11=0, core_node9=0}, version=2}
[junit4] 2> 443546 INFO (qtp1840676713-6919) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard2 r:core_node9
x:hdfsbackuprestore_testok_shard2_replica_n7 ]
o.a.s.c.ShardLeaderElectionContextBase make sure parent is created
/collections/hdfsbackuprestore_testok/leaders/shard2
[junit4] 2> 443952 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ]
o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 443952 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ]
o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 443952 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.c.SyncStrategy Sync
replicas to
https://127.0.0.1:46735/solr/hdfsbackuprestore_testok_shard1_replica_n1/
[junit4] 2> 443953 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.PeerSync PeerSync:
core=hdfsbackuprestore_testok_shard1_replica_n1
url=https://127.0.0.1:46735/solr START
replicas=[https://127.0.0.1:36659/solr/hdfsbackuprestore_testok_shard1_replica_n2/,
https://127.0.0.1:46735/solr/hdfsbackuprestore_testok_shard1_replica_t4/]
nUpdates=100
[junit4] 2> 443954 INFO (qtp2078506737-6918) [n:127.0.0.1:46735_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node3
x:hdfsbackuprestore_testok_shard1_replica_n1 ] o.a.s.u.PeerSync PeerSync:
core=hdfsbackuprestore_testok_shard1_replica_n1
url=https://127.0.0.1:46735/solr DONE. We have no versions. sync failed.
[junit4] 2> 443961 INFO (qtp1840676713-6925) [n:127.0.0.1:36659_solr
c:hdfsbackuprestore_testok s:shard1 r:core_node5
x:hdfsbackuprestore_testok_shard1_replica_n2 ] o.a.s.c.S.Request [hdfsbac
[...truncated too long message...]
loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.
-ivy-fail-disallowed-ivy-version:
ivy-fail:
ivy-fail:
ivy-configure:
[ivy:configure] :: loading settings :: file =
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/lucene/top-level-ivy-settings.xml
resolve:
jar-checksums:
[mkdir] Created dir:
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/null699052273
[copy] Copying 249 files to
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/null699052273
[delete] Deleting directory
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-8.x/solr/null699052273
check-working-copy:
[ivy:cachepath] :: resolving dependencies :: #;working@lucene1-us-west
[ivy:cachepath] confs: [default]
[ivy:cachepath] found
org.eclipse.jgit#org.eclipse.jgit;5.3.0.201903130848-r in public
[ivy:cachepath] found com.jcraft#jsch;0.1.54 in public
[ivy:cachepath] found com.jcraft#jzlib;1.1.1 in public
[ivy:cachepath] found com.googlecode.javaewah#JavaEWAH;1.1.6 in public
[ivy:cachepath] found org.slf4j#slf4j-api;1.7.2 in public
[ivy:cachepath] found org.bouncycastle#bcpg-jdk15on;1.60 in public
[ivy:cachepath] found org.bouncycastle#bcprov-jdk15on;1.60 in public
[ivy:cachepath] found org.bouncycastle#bcpkix-jdk15on;1.60 in public
[ivy:cachepath] found org.slf4j#slf4j-nop;1.7.2 in public
[ivy:cachepath] :: resolution report :: resolve 30ms :: artifacts dl 2ms
---------------------------------------------------------------------
| | modules || artifacts |
| conf | number| search|dwnlded|evicted|| number|dwnlded|
---------------------------------------------------------------------
| default | 9 | 0 | 0 | 0 || 9 | 0 |
---------------------------------------------------------------------
[wc-checker] Initializing working copy...
[wc-checker] Checking working copy status...
-jenkins-base:
BUILD SUCCESSFUL
Total time: 118 minutes 58 seconds
Archiving artifacts
java.lang.InterruptedException: no matches found within 10000
at hudson.FilePath$ValidateAntFileMask.hasMatch(FilePath.java:2847)
at hudson.FilePath$ValidateAntFileMask.invoke(FilePath.java:2726)
at hudson.FilePath$ValidateAntFileMask.invoke(FilePath.java:2707)
at hudson.FilePath$FileCallableWrapper.call(FilePath.java:3086)
Also: hudson.remoting.Channel$CallSiteStackTrace: Remote call to lucene
at
hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1741)
at
hudson.remoting.UserRequest$ExceptionResponse.retrieve(UserRequest.java:357)
at hudson.remoting.Channel.call(Channel.java:955)
at hudson.FilePath.act(FilePath.java:1072)
at hudson.FilePath.act(FilePath.java:1061)
at hudson.FilePath.validateAntFileMask(FilePath.java:2705)
at
hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
at
hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
at
hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
at
hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
at hudson.model.Build$BuildExecution.post2(Build.java:186)
at
hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
at hudson.model.Run.execute(Run.java:1835)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
at
hudson.model.ResourceController.execute(ResourceController.java:97)
at hudson.model.Executor.run(Executor.java:429)
Caused: hudson.FilePath$TunneledInterruptedException
at hudson.FilePath$FileCallableWrapper.call(FilePath.java:3088)
at hudson.remoting.UserRequest.perform(UserRequest.java:212)
at hudson.remoting.UserRequest.perform(UserRequest.java:54)
at hudson.remoting.Request$2.run(Request.java:369)
at
hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:744)
Caused: java.lang.InterruptedException: java.lang.InterruptedException: no
matches found within 10000
at hudson.FilePath.act(FilePath.java:1074)
at hudson.FilePath.act(FilePath.java:1061)
at hudson.FilePath.validateAntFileMask(FilePath.java:2705)
at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
at
hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
at
hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
at hudson.model.Build$BuildExecution.post2(Build.java:186)
at
hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
at hudson.model.Run.execute(Run.java:1835)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
at hudson.model.ResourceController.execute(ResourceController.java:97)
at hudson.model.Executor.run(Executor.java:429)
No artifacts found that match the file pattern
"**/*.events,heapdumps/**,**/hs_err_pid*". Configuration error?
Recording test results
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Email was triggered for: Unstable (Test Failures)
Sending email for trigger: Unstable (Test Failures)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]