Build: https://builds.apache.org/job/Lucene-Solr-Tests-7.x/343/

4 tests failed.
FAILED:  
org.apache.solr.cloud.ConcurrentCreateRoutedAliasTest.testConcurrentCreateRoutedAliasMinimal

Error Message:
concurrent alias creation failed java.lang.IllegalStateException: Connection 
pool shut down

Stack Trace:
java.lang.AssertionError: concurrent alias creation failed 
java.lang.IllegalStateException: Connection pool shut down
        at 
__randomizedtesting.SeedInfo.seed([963C9F53888D1D35:52A793012592AA0F]:0)
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.junit.Assert.assertNull(Assert.java:551)
        at 
org.apache.solr.cloud.ConcurrentCreateRoutedAliasTest.testConcurrentCreateRoutedAliasMinimal(ConcurrentCreateRoutedAliasTest.java:108)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at java.lang.Thread.run(Thread.java:748)


FAILED:  org.apache.solr.cloud.api.collections.TestHdfsCloudBackupRestore.test

Error Message:
expected:<COMPLETED> but was:<RUNNING>

Stack Trace:
java.lang.AssertionError: expected:<COMPLETED> but was:<RUNNING>
        at 
__randomizedtesting.SeedInfo.seed([963C9F53888D1D35:1E68A089267170CD]:0)
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.failNotEquals(Assert.java:647)
        at org.junit.Assert.assertEquals(Assert.java:128)
        at org.junit.Assert.assertEquals(Assert.java:147)
        at 
org.apache.solr.cloud.api.collections.AbstractCloudBackupRestoreTestCase.testBackupAndRestore(AbstractCloudBackupRestoreTestCase.java:289)
        at 
org.apache.solr.cloud.api.collections.AbstractCloudBackupRestoreTestCase.test(AbstractCloudBackupRestoreTestCase.java:142)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at java.lang.Thread.run(Thread.java:748)


FAILED:  
org.apache.solr.cloud.autoscaling.TriggerIntegrationTest.testMetricTrigger

Error Message:


Stack Trace:
java.lang.AssertionError
        at 
__randomizedtesting.SeedInfo.seed([963C9F53888D1D35:2C30A8DCD765CB7A]:0)
        at org.junit.Assert.fail(Assert.java:92)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.junit.Assert.assertNull(Assert.java:551)
        at org.junit.Assert.assertNull(Assert.java:562)
        at 
org.apache.solr.cloud.autoscaling.TriggerIntegrationTest.testMetricTrigger(TriggerIntegrationTest.java:1575)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at java.lang.Thread.run(Thread.java:748)


FAILED:  org.apache.solr.cloud.autoscaling.sim.TestLargeCluster.testSearchRate

Error Message:
The trigger did not fire at all

Stack Trace:
java.lang.AssertionError: The trigger did not fire at all
        at 
__randomizedtesting.SeedInfo.seed([963C9F53888D1D35:CB7481DA474BBB7A]:0)
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at 
org.apache.solr.cloud.autoscaling.sim.TestLargeCluster.testSearchRate(TestLargeCluster.java:547)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
        at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
        at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
        at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
        at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
        at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
        at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
        at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
        at java.lang.Thread.run(Thread.java:748)




Build Log:
[...truncated 11947 lines...]
   [junit4] Suite: 
org.apache.solr.cloud.api.collections.TestHdfsCloudBackupRestore
   [junit4]   2> Creating dataDir: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/init-core-data-001
   [junit4]   2> 917264 WARN  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=4 numCloses=4
   [junit4]   2> 917264 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) 
w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 917301 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via: 
@org.apache.solr.util.RandomizeSSL(reason=, ssl=NaN, value=NaN, clientAuth=NaN)
   [junit4]   2> 917302 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: 
test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 919775 WARN  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.h.m.i.MetricsConfig Cannot locate configuration: tried 
hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 919993 WARN  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 920000 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.m.log jetty-6.1.26
   [junit4]   2> 920156 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs
 to ./temp/Jetty_localhost_51411_hdfs____vvszqa/webapp
   [junit4]   2> 926743 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:51411
   [junit4]   2> 929257 WARN  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 929259 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.m.log jetty-6.1.26
   [junit4]   2> 929359 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode
 to ./temp/Jetty_localhost_47051_datanode____.ll8ncg/webapp
   [junit4]   2> 935682 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:47051
   [junit4]   2> 936844 WARN  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 936850 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.m.log jetty-6.1.26
   [junit4]   2> 937262 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.m.log Extract 
jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode
 to ./temp/Jetty_localhost_38677_datanode____.josk94/webapp
   [junit4]   2> 938918 ERROR (DataNode: 
[[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-001/hdfsBaseDir/data/data1/,
 
[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-001/hdfsBaseDir/data/data2/]]
  heartbeating to localhost/127.0.0.1:32939) [    ] 
o.a.h.h.s.d.DirectoryScanner 
dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 
ms/sec. Assuming default value of 1000
   [junit4]   2> 938995 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0x3e857f447dce87: from storage 
DS-85f543fc-ae18-4654-8ea7-457d58eb022b node 
DatanodeRegistration(127.0.0.1:56062, 
datanodeUuid=3efa0368-b7ff-4cc5-9a36-8589c6421d1d, infoPort=59133, 
infoSecurePort=0, ipcPort=40369, 
storageInfo=lv=-56;cid=testClusterID;nsid=1396815523;c=0), blocks: 0, 
hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 938995 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0x3e857f447dce87: from storage 
DS-d2ca3cce-9ac1-4d21-af71-0b436165d624 node 
DatanodeRegistration(127.0.0.1:56062, 
datanodeUuid=3efa0368-b7ff-4cc5-9a36-8589c6421d1d, infoPort=59133, 
infoSecurePort=0, ipcPort=40369, 
storageInfo=lv=-56;cid=testClusterID;nsid=1396815523;c=0), blocks: 0, 
hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 941957 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.m.log Started 
HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:38677
   [junit4]   2> 947319 ERROR (DataNode: 
[[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-001/hdfsBaseDir/data/data3/,
 
[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-001/hdfsBaseDir/data/data4/]]
  heartbeating to localhost/127.0.0.1:32939) [    ] 
o.a.h.h.s.d.DirectoryScanner 
dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 
ms/sec. Assuming default value of 1000
   [junit4]   2> 947477 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0x3e85813cbddc44: from storage 
DS-f018afea-3b42-41f5-8833-bcfa764f588a node 
DatanodeRegistration(127.0.0.1:45986, 
datanodeUuid=e7c0d423-4302-4585-805e-0a6cf4aef7dc, infoPort=48285, 
infoSecurePort=0, ipcPort=42023, 
storageInfo=lv=-56;cid=testClusterID;nsid=1396815523;c=0), blocks: 0, 
hasStaleStorage: true, processing time: 16 msecs
   [junit4]   2> 947477 INFO  (Block report processor) [    ] BlockStateChange 
BLOCK* processReport 0x3e85813cbddc44: from storage 
DS-5e3846f9-fd92-46cd-81a4-2b9391328869 node 
DatanodeRegistration(127.0.0.1:45986, 
datanodeUuid=e7c0d423-4302-4585-805e-0a6cf4aef7dc, infoPort=48285, 
infoSecurePort=0, ipcPort=42023, 
storageInfo=lv=-56;cid=testClusterID;nsid=1396815523;c=0), blocks: 0, 
hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 949656 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.s.c.MiniSolrCloudCluster Starting cluster of 2 servers in 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002
   [junit4]   2> 949656 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 949657 INFO  (Thread-875) [    ] o.a.s.c.ZkTestServer client 
port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 949657 INFO  (Thread-875) [    ] o.a.s.c.ZkTestServer Starting 
server
   [junit4]   2> 949674 ERROR (Thread-875) [    ] o.a.z.s.ZooKeeperServer 
ZKShutdownHandler is not registered, so ZooKeeper server won't take any action 
on ERROR or SHUTDOWN server state changes
   [junit4]   2> 949758 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.s.c.ZkTestServer start zk server on port:46976
   [junit4]   2> 950061 INFO  (zkConnectionManagerCallback-158-thread-1) [    ] 
o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 950134 INFO  (jetty-launcher-155-thread-1) [    ] 
o.e.j.s.Server jetty-9.4.8.v20171121, build timestamp: 
2017-11-22T00:27:37+03:00, git hash: 82b8fb23f757335bb3329d540ce37a2a2615f0a8
   [junit4]   2> 950146 INFO  (jetty-launcher-155-thread-2) [    ] 
o.e.j.s.Server jetty-9.4.8.v20171121, build timestamp: 
2017-11-22T00:27:37+03:00, git hash: 82b8fb23f757335bb3329d540ce37a2a2615f0a8
   [junit4]   2> 950178 INFO  (jetty-launcher-155-thread-2) [    ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 950178 INFO  (jetty-launcher-155-thread-2) [    ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 950178 INFO  (jetty-launcher-155-thread-2) [    ] 
o.e.j.s.session Scavenging every 660000ms
   [junit4]   2> 950179 INFO  (jetty-launcher-155-thread-2) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@3f6699ad{/solr,null,AVAILABLE}
   [junit4]   2> 950179 INFO  (jetty-launcher-155-thread-2) [    ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@d12c520{HTTP/1.1,[http/1.1]}{127.0.0.1:56822}
   [junit4]   2> 950179 INFO  (jetty-launcher-155-thread-2) [    ] 
o.e.j.s.Server Started @984505ms
   [junit4]   2> 950179 INFO  (jetty-launcher-155-thread-2) [    ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, 
hostPort=56822}
   [junit4]   2> 950180 ERROR (jetty-launcher-155-thread-2) [    ] 
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 950180 INFO  (jetty-launcher-155-thread-2) [    ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 
7.3.0
   [junit4]   2> 950180 INFO  (jetty-launcher-155-thread-2) [    ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 950180 INFO  (jetty-launcher-155-thread-2) [    ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 950180 INFO  (jetty-launcher-155-thread-2) [    ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2018-01-30T21:05:27.580Z
   [junit4]   2> 950182 INFO  (jetty-launcher-155-thread-1) [    ] 
o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 950182 INFO  (jetty-launcher-155-thread-1) [    ] 
o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 950182 INFO  (jetty-launcher-155-thread-1) [    ] 
o.e.j.s.session Scavenging every 660000ms
   [junit4]   2> 950254 INFO  (jetty-launcher-155-thread-1) [    ] 
o.e.j.s.h.ContextHandler Started 
o.e.j.s.ServletContextHandler@75df0b21{/solr,null,AVAILABLE}
   [junit4]   2> 950256 INFO  (jetty-launcher-155-thread-1) [    ] 
o.e.j.s.AbstractConnector Started 
ServerConnector@49915d65{HTTP/1.1,[http/1.1]}{127.0.0.1:45076}
   [junit4]   2> 950256 INFO  (jetty-launcher-155-thread-1) [    ] 
o.e.j.s.Server Started @984582ms
   [junit4]   2> 950256 INFO  (jetty-launcher-155-thread-1) [    ] 
o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, 
hostPort=45076}
   [junit4]   2> 950257 ERROR (jetty-launcher-155-thread-1) [    ] 
o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be 
missing or incomplete.
   [junit4]   2> 950292 INFO  (jetty-launcher-155-thread-1) [    ] 
o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 
7.3.0
   [junit4]   2> 950292 INFO  (jetty-launcher-155-thread-1) [    ] 
o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 950292 INFO  (jetty-launcher-155-thread-1) [    ] 
o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 950292 INFO  (jetty-launcher-155-thread-1) [    ] 
o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 
2018-01-30T21:05:27.692Z
   [junit4]   2> 950736 INFO  (zkConnectionManagerCallback-162-thread-1) [    ] 
o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 950910 INFO  (jetty-launcher-155-thread-1) [    ] 
o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 950944 INFO  (jetty-launcher-155-thread-1) [    ] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 950962 INFO  (jetty-launcher-155-thread-1) [    ] 
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:46976/solr
   [junit4]   2> 950963 INFO  (zkConnectionManagerCallback-160-thread-1) [    ] 
o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 951022 INFO  (jetty-launcher-155-thread-2) [    ] 
o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 951036 INFO  (jetty-launcher-155-thread-2) [    ] 
o.a.s.c.SolrXmlConfig MBean server found: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c, but no JMX reporters were 
configured - adding default JMX reporter.
   [junit4]   2> 951110 INFO  (jetty-launcher-155-thread-2) [    ] 
o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:46976/solr
   [junit4]   2> 951181 INFO  (zkConnectionManagerCallback-168-thread-1) [    ] 
o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 951216 INFO  (zkConnectionManagerCallback-170-thread-1) [    ] 
o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 951514 INFO  
(zkConnectionManagerCallback-172-thread-1-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 951670 INFO  
(zkConnectionManagerCallback-174-thread-1-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 956631 INFO  (jetty-launcher-155-thread-2) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 956633 INFO  (jetty-launcher-155-thread-2) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.OverseerElectionContext I am going to be 
the leader 127.0.0.1:56822_solr
   [junit4]   2> 956687 INFO  (jetty-launcher-155-thread-2) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.Overseer Overseer 
(id=73210912353026054-127.0.0.1:56822_solr-n_0000000000) starting
   [junit4]   2> 956985 INFO  (jetty-launcher-155-thread-2) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:56822_solr
   [junit4]   2> 957069 INFO  
(zkCallback-173-thread-1-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from 
ZooKeeper... (0) -> (1)
   [junit4]   2> 957633 INFO  (jetty-launcher-155-thread-2) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.b.r.BackupRepositoryFactory Added backup 
repository with configuration params {type = repository,name = hdfs,class = 
org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes = 
{name=hdfs, 
class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args = 
{location=/backup,solr.hdfs.home=hdfs://localhost:32939/solr,solr.hdfs.confdir=}}
   [junit4]   2> 957633 INFO  (jetty-launcher-155-thread-2) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.b.r.BackupRepositoryFactory Default 
configuration for backup repository is with configuration params {type = 
repository,name = hdfs,class = 
org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes = 
{name=hdfs, 
class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args = 
{location=/backup,solr.hdfs.home=hdfs://localhost:32939/solr,solr.hdfs.confdir=}}
   [junit4]   2> 958444 INFO  (jetty-launcher-155-thread-2) 
[n:127.0.0.1:56822_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.node' (registry 'solr.node') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 958710 INFO  (jetty-launcher-155-thread-1) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from 
ZooKeeper... (0) -> (1)
   [junit4]   2> 958720 INFO  (jetty-launcher-155-thread-1) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 958728 INFO  (jetty-launcher-155-thread-2) 
[n:127.0.0.1:56822_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jvm' (registry 'solr.jvm') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 958728 INFO  (jetty-launcher-155-thread-2) 
[n:127.0.0.1:56822_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jetty' (registry 'solr.jetty') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 958741 INFO  (jetty-launcher-155-thread-1) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.ZkController Register node as live in 
ZooKeeper:/live_nodes/127.0.0.1:45076_solr
   [junit4]   2> 958742 INFO  
(zkCallback-173-thread-1-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from 
ZooKeeper... (1) -> (2)
   [junit4]   2> 958808 INFO  (jetty-launcher-155-thread-2) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core 
definitions underneath 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node2/.
   [junit4]   2> 959161 INFO  
(zkCallback-171-thread-1-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from 
ZooKeeper... (1) -> (2)
   [junit4]   2> 960062 INFO  (jetty-launcher-155-thread-1) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.b.r.BackupRepositoryFactory Added backup 
repository with configuration params {type = repository,name = hdfs,class = 
org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes = 
{name=hdfs, 
class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args = 
{location=/backup,solr.hdfs.home=hdfs://localhost:32939/solr,solr.hdfs.confdir=}}
   [junit4]   2> 960065 INFO  (jetty-launcher-155-thread-1) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.b.r.BackupRepositoryFactory Default 
configuration for backup repository is with configuration params {type = 
repository,name = hdfs,class = 
org.apache.solr.core.backup.repository.HdfsBackupRepository,attributes = 
{name=hdfs, 
class=org.apache.solr.core.backup.repository.HdfsBackupRepository},args = 
{location=/backup,solr.hdfs.home=hdfs://localhost:32939/solr,solr.hdfs.confdir=}}
   [junit4]   2> 960734 INFO  (jetty-launcher-155-thread-1) 
[n:127.0.0.1:45076_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.node' (registry 'solr.node') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 961038 INFO  (jetty-launcher-155-thread-1) 
[n:127.0.0.1:45076_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jvm' (registry 'solr.jvm') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 961038 INFO  (jetty-launcher-155-thread-1) 
[n:127.0.0.1:45076_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 
'solr.jetty' (registry 'solr.jetty') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 961069 INFO  (jetty-launcher-155-thread-1) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core 
definitions underneath 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node1/.
   [junit4]   2> 961771 INFO  (zkConnectionManagerCallback-180-thread-1) [    ] 
o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 961940 INFO  (zkConnectionManagerCallback-184-thread-1) [    ] 
o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 961946 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 961949 INFO  
(SUITE-TestHdfsCloudBackupRestore-seed#[963C9F53888D1D35]-worker) [    ] 
o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:46976/solr ready
   [junit4]   2> 962521 INFO  
(TEST-TestHdfsCloudBackupRestore.test-seed#[963C9F53888D1D35]) [    ] 
o.a.s.SolrTestCaseJ4 ###Starting test
   [junit4]   2> 962583 INFO  (qtp527904507-1865) [n:127.0.0.1:56822_solr    ] 
o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params 
replicationFactor=2&collection.configName=conf1&router.name=implicit&version=2&pullReplicas=1&shards=shard1,shard2&property.customKey=customValue&maxShardsPerNode=3&router.field=shard_s&autoAddReplicas=true&name=hdfsbackuprestore&nrtReplicas=2&action=CREATE&tlogReplicas=0&wt=javabin
 and sendToOCPQueue=true
   [junit4]   2> 962618 INFO  
(OverseerThreadFactory-775-thread-1-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.a.c.CreateCollectionCmd Create collection 
hdfsbackuprestore
   [junit4]   2> 962619 WARN  
(OverseerThreadFactory-775-thread-1-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.a.c.CreateCollectionCmd Specified number 
of replicas of 3 on collection hdfsbackuprestore is higher than the number of 
Solr instances currently live or live and part of your createNodeSet(2). It's 
unusual to run two replica of the same slice on the same Solr-instance.
   [junit4]   2> 962993 INFO  
(OverseerStateUpdate-73210912353026054-127.0.0.1:56822_solr-n_0000000000) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"hdfsbackuprestore_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:45076/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 963025 INFO  
(OverseerStateUpdate-73210912353026054-127.0.0.1:56822_solr-n_0000000000) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"hdfsbackuprestore_shard1_replica_n2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:56822/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 963059 INFO  
(OverseerStateUpdate-73210912353026054-127.0.0.1:56822_solr-n_0000000000) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"hdfsbackuprestore_shard1_replica_p4",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:45076/solr";,
   [junit4]   2>   "type":"PULL",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 963125 INFO  
(OverseerStateUpdate-73210912353026054-127.0.0.1:56822_solr-n_0000000000) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"hdfsbackuprestore_shard2_replica_n6",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:56822/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 963193 INFO  
(OverseerStateUpdate-73210912353026054-127.0.0.1:56822_solr-n_0000000000) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"hdfsbackuprestore_shard2_replica_n8",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:45076/solr";,
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 963329 INFO  
(OverseerStateUpdate-73210912353026054-127.0.0.1:56822_solr-n_0000000000) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"hdfsbackuprestore",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"hdfsbackuprestore_shard2_replica_p10",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:56822/solr";,
   [junit4]   2>   "type":"PULL",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 963722 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr    ] 
o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node3&name=hdfsbackuprestore_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&wt=javabin
   [junit4]   2> 963722 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr    ] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 
transient cores
   [junit4]   2> 963770 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr    ] 
o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore&version=2&replicaType=PULL&property.customKey=customValue&coreNodeName=core_node12&name=hdfsbackuprestore_shard2_replica_p10&action=CREATE&numShards=2&shard=shard2&wt=javabin
   [junit4]   2> 963777 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr    ] 
o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 
transient cores
   [junit4]   2> 964183 INFO  
(zkCallback-173-thread-1-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 964203 INFO  
(zkCallback-173-thread-2-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 964204 INFO  
(zkCallback-171-thread-1-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 964330 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr    ] 
o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node11&name=hdfsbackuprestore_shard2_replica_n8&action=CREATE&numShards=2&shard=shard2&wt=javabin
   [junit4]   2> 964335 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr    ] 
o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node5&name=hdfsbackuprestore_shard1_replica_n2&action=CREATE&numShards=2&shard=shard1&wt=javabin
   [junit4]   2> 964357 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr    ] 
o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore&version=2&replicaType=PULL&property.customKey=customValue&coreNodeName=core_node7&name=hdfsbackuprestore_shard1_replica_p4&action=CREATE&numShards=2&shard=shard1&wt=javabin
   [junit4]   2> 964398 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr    ] 
o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node9&name=hdfsbackuprestore_shard2_replica_n6&action=CREATE&numShards=2&shard=shard2&wt=javabin
   [junit4]   2> 964504 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 7.3.0
   [junit4]   2> 964538 INFO  
(zkCallback-173-thread-3-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 964554 INFO  
(zkCallback-173-thread-2-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 964554 INFO  
(zkCallback-171-thread-1-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 964587 INFO  
(zkCallback-171-thread-2-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 964686 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.s.IndexSchema 
[hdfsbackuprestore_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 964823 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema 
minimal/1.1 with uniqueid field id
   [junit4]   2> 964823 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 
'hdfsbackuprestore_shard1_replica_n1' using configuration from collection 
hdfsbackuprestore, trusted=true
   [junit4]   2> 964824 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring 
for 'solr.core.hdfsbackuprestore.shard1.replica_n1' (registry 
'solr.core.hdfsbackuprestore.shard1.replica_n1') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 964824 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.SolrCore 
solr.RecoveryStrategy.Builder
   [junit4]   2> 964824 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.SolrCore 
[[hdfsbackuprestore_shard1_replica_n1] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node1/hdfsbackuprestore_shard1_replica_n1],
 
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node1/./hdfsbackuprestore_shard1_replica_n1/data/]
   [junit4]   2> 965288 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 7.3.0
   [junit4]   2> 965450 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.s.IndexSchema 
[hdfsbackuprestore_shard2_replica_p10] Schema name=minimal
   [junit4]   2> 965497 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.s.IndexSchema Loaded schema 
minimal/1.1 with uniqueid field id
   [junit4]   2> 965497 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.c.CoreContainer Creating SolrCore 
'hdfsbackuprestore_shard2_replica_p10' using configuration from collection 
hdfsbackuprestore, trusted=true
   [junit4]   2> 965498 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.m.r.SolrJmxReporter JMX 
monitoring for 'solr.core.hdfsbackuprestore.shard2.replica_p10' (registry 
'solr.core.hdfsbackuprestore.shard2.replica_p10') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 965499 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.c.SolrCore 
solr.RecoveryStrategy.Builder
   [junit4]   2> 965499 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.c.SolrCore 
[[hdfsbackuprestore_shard2_replica_p10] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node2/hdfsbackuprestore_shard2_replica_p10],
 
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node2/./hdfsbackuprestore_shard2_replica_p10/data/]
   [junit4]   2> 966125 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 7.3.0
   [junit4]   2> 966163 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 7.3.0
   [junit4]   2> 966168 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 7.3.0
   [junit4]   2> 966212 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.c.SolrConfig Using Lucene 
MatchVersion: 7.3.0
   [junit4]   2> 966727 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.s.IndexSchema 
[hdfsbackuprestore_shard2_replica_n6] Schema name=minimal
   [junit4]   2> 966790 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.s.IndexSchema Loaded schema 
minimal/1.1 with uniqueid field id
   [junit4]   2> 966790 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.CoreContainer Creating SolrCore 
'hdfsbackuprestore_shard2_replica_n6' using configuration from collection 
hdfsbackuprestore, trusted=true
   [junit4]   2> 966791 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring 
for 'solr.core.hdfsbackuprestore.shard2.replica_n6' (registry 
'solr.core.hdfsbackuprestore.shard2.replica_n6') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 966791 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.SolrCore 
solr.RecoveryStrategy.Builder
   [junit4]   2> 966792 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.SolrCore 
[[hdfsbackuprestore_shard2_replica_n6] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node2/hdfsbackuprestore_shard2_replica_n6],
 
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node2/./hdfsbackuprestore_shard2_replica_n6/data/]
   [junit4]   2> 966990 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.s.IndexSchema 
[hdfsbackuprestore_shard1_replica_p4] Schema name=minimal
   [junit4]   2> 967014 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.s.IndexSchema 
[hdfsbackuprestore_shard1_replica_n2] Schema name=minimal
   [junit4]   2> 967085 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.s.IndexSchema Loaded schema 
minimal/1.1 with uniqueid field id
   [junit4]   2> 967085 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.c.CoreContainer Creating SolrCore 
'hdfsbackuprestore_shard1_replica_p4' using configuration from collection 
hdfsbackuprestore, trusted=true
   [junit4]   2> 967086 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.s.IndexSchema Loaded schema 
minimal/1.1 with uniqueid field id
   [junit4]   2> 967086 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 
'hdfsbackuprestore_shard1_replica_n2' using configuration from collection 
hdfsbackuprestore, trusted=true
   [junit4]   2> 967097 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.m.r.SolrJmxReporter JMX monitoring 
for 'solr.core.hdfsbackuprestore.shard1.replica_p4' (registry 
'solr.core.hdfsbackuprestore.shard1.replica_p4') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 967097 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.c.SolrCore 
solr.RecoveryStrategy.Builder
   [junit4]   2> 967097 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.c.SolrCore 
[[hdfsbackuprestore_shard1_replica_p4] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node1/hdfsbackuprestore_shard1_replica_p4],
 
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node1/./hdfsbackuprestore_shard1_replica_p4/data/]
   [junit4]   2> 967269 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring 
for 'solr.core.hdfsbackuprestore.shard1.replica_n2' (registry 
'solr.core.hdfsbackuprestore.shard1.replica_n2') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 967269 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.c.SolrCore 
solr.RecoveryStrategy.Builder
   [junit4]   2> 967269 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.c.SolrCore 
[[hdfsbackuprestore_shard1_replica_n2] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node2/hdfsbackuprestore_shard1_replica_n2],
 
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node2/./hdfsbackuprestore_shard1_replica_n2/data/]
   [junit4]   2> 967777 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.s.IndexSchema 
[hdfsbackuprestore_shard2_replica_n8] Schema name=minimal
   [junit4]   2> 968010 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.s.IndexSchema Loaded schema 
minimal/1.1 with uniqueid field id
   [junit4]   2> 968010 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.c.CoreContainer Creating SolrCore 
'hdfsbackuprestore_shard2_replica_n8' using configuration from collection 
hdfsbackuprestore, trusted=true
   [junit4]   2> 968011 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.m.r.SolrJmxReporter JMX monitoring 
for 'solr.core.hdfsbackuprestore.shard2.replica_n8' (registry 
'solr.core.hdfsbackuprestore.shard2.replica_n8') enabled at server: 
com.sun.jmx.mbeanserver.JmxMBeanServer@52a4746c
   [junit4]   2> 968011 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.c.SolrCore 
solr.RecoveryStrategy.Builder
   [junit4]   2> 968011 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.c.SolrCore 
[[hdfsbackuprestore_shard2_replica_n8] ] Opening new SolrCore at 
[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node1/hdfsbackuprestore_shard2_replica_n8],
 
dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J1/temp/solr.cloud.api.collections.TestHdfsCloudBackupRestore_963C9F53888D1D35-001/tempDir-002/node1/./hdfsbackuprestore_shard2_replica_n8/data/]
   [junit4]   2> 969016 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog 
implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 969016 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.u.UpdateLog Initializing 
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 969019 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: 
disabled
   [junit4]   2> 969019 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: 
disabled
   [junit4]   2> 969045 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.u.CommitTracker Hard AutoCommit: 
disabled
   [junit4]   2> 969045 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.u.CommitTracker Soft AutoCommit: 
disabled
   [junit4]   2> 969047 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening 
[Searcher@ab1621a[hdfsbackuprestore_shard1_replica_n1] main]
   [junit4]   2> 969071 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 969081 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded 
null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 969158 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will 
be reserved for 10000ms.
   [junit4]   2> 969161 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.s.SolrIndexSearcher Opening 
[Searcher@6419745b[hdfsbackuprestore_shard2_replica_p10] main]
   [junit4]   2> 969259 INFO  
(searcherExecutor-780-thread-1-processing-n:127.0.0.1:45076_solr 
x:hdfsbackuprestore_shard1_replica_n1 s:shard1 c:hdfsbackuprestore 
r:core_node3) [n:127.0.0.1:45076_solr c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.SolrCore 
[hdfsbackuprestore_shard1_replica_n1] Registered new searcher 
Searcher@ab1621a[hdfsbackuprestore_shard1_replica_n1] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 969318 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max 
version in index or recent updates, using new clock 1591052962856173568
   [junit4]   2> 969341 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 969342 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.r.ManagedResourceStorage Loaded 
null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 969343 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.h.ReplicationHandler Commits will 
be reserved for 10000ms.
   [junit4]   2> 970013 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.c.ZkController 
hdfsbackuprestore_shard2_replica_p10 starting background replication from leader
   [junit4]   2> 970015 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.c.ReplicateFromLeader Will start 
replication from leader with poll interval: 00:00:03
   [junit4]   2> 970079 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.h.ReplicationHandler Poll 
scheduled at an interval of 3000ms
   [junit4]   2> 970079 INFO  (qtp527904507-1861) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.h.ReplicationHandler Commits will 
be reserved for 10000ms.
   [junit4]   2> 970081 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext 
Waiting until we see more replicas up for shard shard1: total=2 found=1 
timeoutin=9999ms
   [junit4]   2> 970129 INFO  
(searcherExecutor-781-thread-1-processing-n:127.0.0.1:56822_solr 
x:hdfsbackuprestore_shard2_replica_p10 s:shard2 c:hdfsbackuprestore 
r:core_node12) [n:127.0.0.1:56822_solr c:hdfsbackuprestore s:shard2 
r:core_node12 x:hdfsbackuprestore_shard2_replica_p10] o.a.s.c.SolrCore 
[hdfsbackuprestore_shard2_replica_p10] Registered new searcher 
Searcher@6419745b[hdfsbackuprestore_shard2_replica_p10] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 970145 INFO  
(zkCallback-173-thread-2-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 970145 INFO  
(zkCallback-173-thread-3-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 970149 INFO  
(zkCallback-171-thread-2-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 970149 INFO  
(zkCallback-171-thread-1-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 970193 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.u.CommitTracker Hard AutoCommit: 
disabled
   [junit4]   2> 970193 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.u.CommitTracker Soft AutoCommit: 
disabled
   [junit4]   2> 970617 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.u.UpdateHandler Using UpdateLog 
implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 970617 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.u.UpdateLog Initializing 
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 970619 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.u.CommitTracker Hard AutoCommit: 
disabled
   [junit4]   2> 970619 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.u.CommitTracker Soft AutoCommit: 
disabled
   [junit4]   2> 970657 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.s.SolrIndexSearcher Opening 
[Searcher@23ae362d[hdfsbackuprestore_shard2_replica_n6] main]
   [junit4]   2> 970659 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 970660 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.r.ManagedResourceStorage Loaded 
null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 970673 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.h.ReplicationHandler Commits will 
be reserved for 10000ms.
   [junit4]   2> 970673 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.u.UpdateLog Could not find max 
version in index or recent updates, using new clock 1591052964276994048
   [junit4]   2> 970677 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.s.SolrIndexSearcher Opening 
[Searcher@7f2f4e9c[hdfsbackuprestore_shard1_replica_p4] main]
   [junit4]   2> 970695 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 970695 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.r.ManagedResourceStorage Loaded 
null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 970697 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.h.ReplicationHandler Commits will 
be reserved for 10000ms.
   [junit4]   2> 970713 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.u.UpdateHandler Using UpdateLog 
implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 970714 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.u.UpdateLog Initializing 
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 970716 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.u.CommitTracker Hard AutoCommit: 
disabled
   [junit4]   2> 970716 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.u.CommitTracker Soft AutoCommit: 
disabled
   [junit4]   2> 970845 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.s.SolrIndexSearcher Opening 
[Searcher@1414bf[hdfsbackuprestore_shard2_replica_n8] main]
   [junit4]   2> 970846 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.u.UpdateHandler Using UpdateLog 
implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 970846 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.u.UpdateLog Initializing 
UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 
maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 970862 INFO  
(searcherExecutor-782-thread-1-processing-n:127.0.0.1:56822_solr 
x:hdfsbackuprestore_shard2_replica_n6 s:shard2 c:hdfsbackuprestore 
r:core_node9) [n:127.0.0.1:56822_solr c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.SolrCore 
[hdfsbackuprestore_shard2_replica_n6] Registered new searcher 
Searcher@23ae362d[hdfsbackuprestore_shard2_replica_n6] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 970863 INFO  
(searcherExecutor-783-thread-1-processing-n:127.0.0.1:45076_solr 
x:hdfsbackuprestore_shard1_replica_p4 s:shard1 c:hdfsbackuprestore 
r:core_node7) [n:127.0.0.1:45076_solr c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.c.SolrCore 
[hdfsbackuprestore_shard1_replica_p4] Registered new searcher 
Searcher@7f2f4e9c[hdfsbackuprestore_shard1_replica_p4] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 971467 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.u.CommitTracker Hard AutoCommit: 
disabled
   [junit4]   2> 971467 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.u.CommitTracker Soft AutoCommit: 
disabled
   [junit4]   2> 971468 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 971469 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.r.ManagedResourceStorage Loaded 
null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 971470 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.h.ReplicationHandler Commits will 
be reserved for 10000ms.
   [junit4]   2> 971470 INFO  (qtp1066405676-1870) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.u.UpdateLog Could not find max 
version in index or recent updates, using new clock 1591052965112709120
   [junit4]   2> 971489 INFO  
(searcherExecutor-787-thread-1-processing-n:127.0.0.1:45076_solr 
x:hdfsbackuprestore_shard2_replica_n8 s:shard2 c:hdfsbackuprestore 
r:core_node11) [n:127.0.0.1:45076_solr c:hdfsbackuprestore s:shard2 
r:core_node11 x:hdfsbackuprestore_shard2_replica_n8] o.a.s.c.SolrCore 
[hdfsbackuprestore_shard2_replica_n8] Registered new searcher 
Searcher@1414bf[hdfsbackuprestore_shard2_replica_n8] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 971620 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.s.SolrIndexSearcher Opening 
[Searcher@554bd268[hdfsbackuprestore_shard1_replica_n2] main]
   [junit4]   2> 971621 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.c.ZkController 
hdfsbackuprestore_shard1_replica_p4 starting background replication from leader
   [junit4]   2> 971622 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.c.ReplicateFromLeader Will start 
replication from leader with poll interval: 00:00:03
   [junit4]   2> 971635 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.h.ReplicationHandler Poll 
scheduled at an interval of 3000ms
   [junit4]   2> 971635 INFO  (qtp1066405676-1877) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node7 
x:hdfsbackuprestore_shard1_replica_p4] o.a.s.h.ReplicationHandler Commits will 
be reserved for 10000ms.
   [junit4]   2> 971665 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.r.ManagedResourceStorage 
Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 971673 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.r.ManagedResourceStorage Loaded 
null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 971674 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.h.ReplicationHandler Commits will 
be reserved for 10000ms.
   [junit4]   2> 971675 INFO  (qtp527904507-1863) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.u.UpdateLog Could not find max 
version in index or recent updates, using new clock 1591052965327667200
   [junit4]   2> 971786 INFO  
(searcherExecutor-784-thread-1-processing-n:127.0.0.1:56822_solr 
x:hdfsbackuprestore_shard1_replica_n2 s:shard1 c:hdfsbackuprestore 
r:core_node5) [n:127.0.0.1:56822_solr c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.c.SolrCore 
[hdfsbackuprestore_shard1_replica_n2] Registered new searcher 
Searcher@554bd268[hdfsbackuprestore_shard1_replica_n2] 
main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 971904 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext 
Enough replicas found to continue.
   [junit4]   2> 971905 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext I may 
be the new leader - try and sync
   [junit4]   2> 971905 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.SyncStrategy Sync replicas to 
http://127.0.0.1:56822/solr/hdfsbackuprestore_shard2_replica_n6/
   [junit4]   2> 971932 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.u.PeerSync PeerSync: 
core=hdfsbackuprestore_shard2_replica_n6 url=http://127.0.0.1:56822/solr START 
replicas=[http://127.0.0.1:45076/solr/hdfsbackuprestore_shard2_replica_n8/] 
nUpdates=100
   [junit4]   2> 971978 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext 
Enough replicas found to continue.
   [junit4]   2> 971978 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may 
be the new leader - try and sync
   [junit4]   2> 971978 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to 
http://127.0.0.1:45076/solr/hdfsbackuprestore_shard1_replica_n1/
   [junit4]   2> 971979 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.u.PeerSync PeerSync: 
core=hdfsbackuprestore_shard1_replica_n1 url=http://127.0.0.1:45076/solr START 
replicas=[http://127.0.0.1:56822/solr/hdfsbackuprestore_shard1_replica_n2/] 
nUpdates=100
   [junit4]   2> 972002 INFO  
(zkCallback-171-thread-1-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 972003 INFO  
(zkCallback-173-thread-3-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 972003 INFO  
(zkCallback-173-thread-2-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 972073 INFO  
(zkCallback-171-thread-3-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 972149 INFO  (qtp527904507-1938) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard1 r:core_node5 
x:hdfsbackuprestore_shard1_replica_n2] o.a.s.c.S.Request 
[hdfsbackuprestore_shard1_replica_n2]  webapp=/solr path=/get 
params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2}
 status=0 QTime=19
   [junit4]   2> 972298 INFO  (qtp1066405676-1872) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard2 r:core_node11 
x:hdfsbackuprestore_shard2_replica_n8] o.a.s.c.S.Request 
[hdfsbackuprestore_shard2_replica_n8]  webapp=/solr path=/get 
params={distrib=false&qt=/get&fingerprint=false&getVersions=100&wt=javabin&version=2}
 status=0 QTime=124
   [junit4]   2> 972301 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.u.PeerSync PeerSync: 
core=hdfsbackuprestore_shard2_replica_n6 url=http://127.0.0.1:56822/solr DONE.  
We have no versions.  sync failed.
   [junit4]   2> 972325 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.SyncStrategy Leader's attempt to 
sync with shard failed, moving to the next candidate
   [junit4]   2> 972325 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext We 
failed sync, but we have no versions - we can't sync in that case - we were 
active before, so become leader anyway
   [junit4]   2> 972325 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext Found 
all replicas participating in election, clear LIR
   [junit4]   2> 972373 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.u.PeerSync PeerSync: 
core=hdfsbackuprestore_shard1_replica_n1 url=http://127.0.0.1:45076/solr DONE.  
We have no versions.  sync failed.
   [junit4]   2> 972373 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.SyncStrategy Leader's attempt to 
sync with shard failed, moving to the next candidate
   [junit4]   2> 972373 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext We 
failed sync, but we have no versions - we can't sync in that case - we were 
active before, so become leader anyway
   [junit4]   2> 972373 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found 
all replicas participating in election, clear LIR
   [junit4]   2> 973709 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am 
the new leader: 
http://127.0.0.1:45076/solr/hdfsbackuprestore_shard1_replica_n1/ shard1
   [junit4]   2> 973763 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext I am 
the new leader: 
http://127.0.0.1:56822/solr/hdfsbackuprestore_shard2_replica_n6/ shard2
   [junit4]   2> 973892 INFO  
(zkCallback-173-thread-1-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 973893 INFO  
(zkCallback-173-thread-3-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 973893 INFO  
(zkCallback-171-thread-3-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 973893 INFO  
(zkCallback-171-thread-1-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 973950 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no 
recovery necessary
   [junit4]   2> 973951 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.c.ZkController I am the leader, no 
recovery necessary
   [junit4]   2> 973951 INFO  (indexFetcher-797-thread-1) 
[n:127.0.0.1:56822_solr c:hdfsbackuprestore s:shard2 r:core_node12 
x:hdfsbackuprestore_shard2_replica_p10] o.a.s.h.IndexFetcher Replica core_node9 
is leader but it's state is down, skipping replication
   [junit4]   2> 973961 INFO  (qtp1066405676-1871) [n:127.0.0.1:45076_solr 
c:hdfsbackuprestore s:shard1 r:core_node3 
x:hdfsbackuprestore_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null 
path=/admin/cores 
params={qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node3&name=hdfsbackuprestore_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&wt=javabin}
 status=0 QTime=10238
   [junit4]   2> 973963 INFO  (qtp527904507-1867) [n:127.0.0.1:56822_solr 
c:hdfsbackuprestore s:shard2 r:core_node9 
x:hdfsbackuprestore_shard2_replica_n6] o.a.s.s.HttpSolrCall [admin] webapp=null 
path=/admin/cores 
params={qt=/admin/cores&collection.configName=conf1&newCollection=true&collection=hdfsbackuprestore&version=2&replicaType=NRT&property.customKey=customValue&coreNodeName=core_node9&name=hdfsbackuprestore_shard2_replica_n6&action=CREATE&numShards=2&shard=shard2&wt=javabin}
 status=0 QTime=9565
   [junit4]   2> 974070 INFO  
(zkCallback-171-thread-3-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 974071 INFO  
(zkCallback-171-thread-1-processing-n:127.0.0.1:45076_solr) 
[n:127.0.0.1:45076_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 974074 INFO  
(zkCallback-173-thread-1-processing-n:127.0.0.1:56822_solr) 
[n:127.0.0.1:56822_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: 
[WatchedEvent state:SyncConnected type:NodeDataChanged 
path:/collections/hdfsbackuprestore/state.json] for collection 
[hdfsbackuprestore] has occurred - updating... 

[...truncated too long message...]

ve/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptor-kerberos-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-admin-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-authn-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-authz-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-changelog-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-collective-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-event-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-exception-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-journal-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-normalization-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-operational-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-referral-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-schema-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-subtree-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-interceptors-trigger-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-jdbm-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-jdbm1-2.0.0-M2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-kerberos-codec-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-ldif-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-mavibot-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-protocol-kerberos-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-protocol-ldap-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-protocol-shared-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/apacheds-xdbm-partition-2.0.0-M15.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/api-all-1.0.0-M20.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/bcprov-jdk15on-1.54.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/byte-buddy-1.6.2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/commons-collections-3.2.2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/commons-math3-3.6.1.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/ehcache-core-2.4.4.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/hadoop-common-2.7.4-tests.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/hadoop-hdfs-2.7.4-tests.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/hadoop-minikdc-2.7.4.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/htrace-core-3.2.0-incubating.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/jersey-core-1.9.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/jersey-server-1.9.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/jetty-6.1.26.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/jetty-sslengine-6.1.26.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/jetty-util-6.1.26.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/mina-core-2.0.0-M5.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/mockito-core-2.6.2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/netty-all-4.0.36.Final.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/test-lib/objenesis-2.5.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/build/analysis/icu/lucene-analyzers-icu-7.3.0-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/contrib/solr-analysis-extras/classes/java:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/build/queryparser/classes/test:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/build/backward-codecs/classes/test:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/contrib/analysis-extras/lib/icu4j-60.2.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-launcher.jar:/x1/jenkins/.ant/lib/ivy-2.4.0.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-junit.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-log4j.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-junit4.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jai.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-javamail.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-bsf.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-commons-net.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-antlr.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jsch.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-oro.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-commons-logging.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-netrexx.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-testutil.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jdepend.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-bcel.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-xalan2.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-resolver.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-jmf.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-apache-regexp.jar:/home/jenkins/tools/ant/apache-ant-1.8.4/lib/ant-swing.jar:/usr/local/asfpackages/java/jdk1.8.0_144/lib/tools.jar:/x1/jenkins/.ivy2/cache/com.carrotsearch.randomizedtesting/junit4-ant/jars/junit4-ant-2.5.3.jar
 com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe -eventsfile 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/temp/junit4-J2-20180130_204902_6387734699034853613405.events
 
@/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/temp/junit4-J2-20180130_204902_638738875226015830346.suites
 -stdin
   [junit4] ERROR: JVM J2 ended with an exception: Forked process returned with 
error code: 1. Very likely a JVM crash.  See process stdout at: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/temp/junit4-J2-20180130_204902_6384939566263072388628.sysout
 See process stderr at: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/temp/junit4-J2-20180130_204902_6381482280639190651496.syserr
   [junit4]     at 
com.carrotsearch.ant.tasks.junit4.JUnit4.executeSlave(JUnit4.java:1519)
   [junit4]     at 
com.carrotsearch.ant.tasks.junit4.JUnit4.access$000(JUnit4.java:126)
   [junit4]     at 
com.carrotsearch.ant.tasks.junit4.JUnit4$2.call(JUnit4.java:982)
   [junit4]     at 
com.carrotsearch.ant.tasks.junit4.JUnit4$2.call(JUnit4.java:979)
   [junit4]     at java.util.concurrent.FutureTask.run(FutureTask.java:266)
   [junit4]     at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
   [junit4]     at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
   [junit4]     at java.lang.Thread.run(Thread.java:748)

BUILD FAILED
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/build.xml:836: The 
following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/build.xml:780: The 
following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/build.xml:59: The 
following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build.xml:262: 
The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/common-build.xml:556:
 The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/common-build.xml:1513:
 The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/common-build.xml:1038:
 At least one slave process threw an exception, first: Forked process returned 
with error code: 1. Very likely a JVM crash.  See process stdout at: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/temp/junit4-J1-20180130_204902_6397246301557099666982.sysout
 See process stderr at: 
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/temp/junit4-J1-20180130_204902_6397477289918787376473.syserr

Total time: 440 minutes 33 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
[Fast Archiver] No artifacts from Lucene-Solr-Tests-7.x #342 to compare, so 
performing full copy of artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

Reply via email to