[JENKINS] Lucene-Solr-7.x-Linux (32bit/jdk1.8.0_144) - Build # 607 - Still Unstable!

2017-10-12 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/607/
Java: 32bit/jdk1.8.0_144 -server -XX:+UseSerialGC

1 tests failed.
FAILED:  
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testCollectionsAPI

Error Message:
Error from server at 
https://127.0.0.1:43073/solr/awhollynewcollection_0_shard3_replica_n4: 
ClusterState says we are the leader 
(https://127.0.0.1:43073/solr/awhollynewcollection_0_shard3_replica_n4), but 
locally we don't think so. Request came from null

Stack Trace:
org.apache.solr.client.solrj.impl.CloudSolrClient$RouteException: Error from 
server at 
https://127.0.0.1:43073/solr/awhollynewcollection_0_shard3_replica_n4: 
ClusterState says we are the leader 
(https://127.0.0.1:43073/solr/awhollynewcollection_0_shard3_replica_n4), but 
locally we don't think so. Request came from null
at 
__randomizedtesting.SeedInfo.seed([D20747C58601919D:9A7233718032BE08]:0)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.directUpdate(CloudSolrClient.java:539)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:993)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:862)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:793)
at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:178)
at 
org.apache.solr.client.solrj.request.UpdateRequest.commit(UpdateRequest.java:233)
at 
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testCollectionsAPI(CollectionsAPIDistributedZkTest.java:459)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Statement

[jira] [Commented] (LUCENE-4100) Maxscore - Efficient Scoring

2017-10-12 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4100?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16203051#comment-16203051
 ] 

Robert Muir commented on LUCENE-4100:
-

Can we avoid the ScoreMode.merge? This seems really, really confusing. In 
general I don't think we should support such merging in MultiCollector or 
anywhere else, we should simply throw exception if things are different.

I think the enum should be further revisited/simplified: essentially at the 
minimum it must capture 2 booleans from the user: whether scores are needed, 
and whether exact total hit count is needed. Perhaps instead of the enum two 
booleans would be easier for now.

I don't understand why we should set the totalHitCount to -1, vs setting to a 
useful approximation, like google. The user said they didn't need the exact 
total hit count, so it should be no surprise, and its a hell of a lot more 
useful than a negative number.

> Maxscore - Efficient Scoring
> 
>
> Key: LUCENE-4100
> URL: https://issues.apache.org/jira/browse/LUCENE-4100
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/codecs, core/query/scoring, core/search
>Affects Versions: 4.0-ALPHA
>Reporter: Stefan Pohl
>  Labels: api-change, gsoc2014, patch, performance
> Fix For: 4.9, 6.0
>
> Attachments: LUCENE-4100.patch, LUCENE-4100.patch, 
> contrib_maxscore.tgz, maxscore.patch
>
>
> At Berlin Buzzwords 2012, I will be presenting 'maxscore', an efficient 
> algorithm first published in the IR domain in 1995 by H. Turtle & J. Flood, 
> that I find deserves more attention among Lucene users (and developers).
> I implemented a proof of concept and did some performance measurements with 
> example queries and lucenebench, the package of Mike McCandless, resulting in 
> very significant speedups.
> This ticket is to get started the discussion on including the implementation 
> into Lucene's codebase. Because the technique requires awareness about it 
> from the Lucene user/developer, it seems best to become a contrib/module 
> package so that it consciously can be chosen to be used.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-7.x-Linux (32bit/jdk1.8.0_144) - Build # 606 - Unstable!

2017-10-12 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/606/
Java: 32bit/jdk1.8.0_144 -client -XX:+UseG1GC

1 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.TestSolrCloudWithSecureImpersonation

Error Message:
1 thread leaked from SUITE scope at 
org.apache.solr.cloud.TestSolrCloudWithSecureImpersonation: 1) 
Thread[id=17287, name=jetty-launcher-2946-thread-2-EventThread, 
state=TIMED_WAITING, group=TGRP-TestSolrCloudWithSecureImpersonation] 
at sun.misc.Unsafe.park(Native Method) at 
java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215) 
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedNanos(AbstractQueuedSynchronizer.java:1037)
 at 
java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1328)
 at java.util.concurrent.CountDownLatch.await(CountDownLatch.java:277)  
   at 
org.apache.curator.CuratorZookeeperClient.internalBlockUntilConnectedOrTimedOut(CuratorZookeeperClient.java:323)
 at org.apache.curator.RetryLoop.callWithRetry(RetryLoop.java:105)  
   at 
org.apache.curator.framework.imps.GetDataBuilderImpl.pathInForeground(GetDataBuilderImpl.java:288)
 at 
org.apache.curator.framework.imps.GetDataBuilderImpl.forPath(GetDataBuilderImpl.java:279)
 at 
org.apache.curator.framework.imps.GetDataBuilderImpl.forPath(GetDataBuilderImpl.java:41)
 at 
org.apache.curator.framework.recipes.shared.SharedValue.readValue(SharedValue.java:244)
 at 
org.apache.curator.framework.recipes.shared.SharedValue.access$100(SharedValue.java:44)
 at 
org.apache.curator.framework.recipes.shared.SharedValue$1.process(SharedValue.java:61)
 at 
org.apache.curator.framework.imps.NamespaceWatcher.process(NamespaceWatcher.java:67)
 at 
org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:530)   
  at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:505)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: 1 thread leaked from SUITE 
scope at org.apache.solr.cloud.TestSolrCloudWithSecureImpersonation: 
   1) Thread[id=17287, name=jetty-launcher-2946-thread-2-EventThread, 
state=TIMED_WAITING, group=TGRP-TestSolrCloudWithSecureImpersonation]
at sun.misc.Unsafe.park(Native Method)
at 
java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedNanos(AbstractQueuedSynchronizer.java:1037)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1328)
at java.util.concurrent.CountDownLatch.await(CountDownLatch.java:277)
at 
org.apache.curator.CuratorZookeeperClient.internalBlockUntilConnectedOrTimedOut(CuratorZookeeperClient.java:323)
at org.apache.curator.RetryLoop.callWithRetry(RetryLoop.java:105)
at 
org.apache.curator.framework.imps.GetDataBuilderImpl.pathInForeground(GetDataBuilderImpl.java:288)
at 
org.apache.curator.framework.imps.GetDataBuilderImpl.forPath(GetDataBuilderImpl.java:279)
at 
org.apache.curator.framework.imps.GetDataBuilderImpl.forPath(GetDataBuilderImpl.java:41)
at 
org.apache.curator.framework.recipes.shared.SharedValue.readValue(SharedValue.java:244)
at 
org.apache.curator.framework.recipes.shared.SharedValue.access$100(SharedValue.java:44)
at 
org.apache.curator.framework.recipes.shared.SharedValue$1.process(SharedValue.java:61)
at 
org.apache.curator.framework.imps.NamespaceWatcher.process(NamespaceWatcher.java:67)
at 
org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:530)
at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:505)
at __randomizedtesting.SeedInfo.seed([F5F67A40B9C4CA76]:0)




Build Log:
[...truncated 13373 lines...]
   [junit4] Suite: org.apache.solr.cloud.TestSolrCloudWithSecureImpersonation
   [junit4]   2> 2159021 INFO  
(SUITE-TestSolrCloudWithSecureImpersonation-seed#[F5F67A40B9C4CA76]-worker) [   
 ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: 
test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> Creating dataDir: 
/home/jenkins/workspace/Lucene-Solr-7.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.TestSolrCloudWithSecureImpersonation_F5F67A40B9C4CA76-001/init-core-data-001
   [junit4]   2> 2159022 INFO  
(SUITE-TestSolrCloudWithSecureImpersonation-seed#[F5F67A40B9C4CA76]-worker) [   
 ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) 
w/NUMERIC_DOCVALUES_SYSPROP=false
   [junit4]   2> 2159023 INFO  
(SUITE-TestSolrCloudWithSecureImpersonation-seed#[F5F67A40B9C4CA76]-worker) [   
 ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (true) via: 
@org.apache.solr.util.RandomizeSSL(reason=, ssl=NaN, value=NaN, clientAuth=NaN)

[jira] [Updated] (SOLR-11467) CdcrBootstrapTest::testBootstrapWithContinousIndexingOnSourceCluster Failure

2017-10-12 Thread Varun Thacker (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Varun Thacker updated SOLR-11467:
-
Affects Version/s: (was: master (8.0))
   7.1
   6.6
   6.6.1
   7.0
   7.0.1

> CdcrBootstrapTest::testBootstrapWithContinousIndexingOnSourceCluster Failure
> 
>
> Key: SOLR-11467
> URL: https://issues.apache.org/jira/browse/SOLR-11467
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: CDCR
>Affects Versions: 6.6, 6.6.1, 7.0, 7.0.1, 7.1
>Reporter: Amrit Sarkar
> Attachments: SOLR-11467-debug-code.log
>
>
> CdcrBootstrapTest is still failing in master and other branches with:
> {code}
> [junit4] FAILURE  130s J1 | 
> CdcrBootstrapTest.testBootstrapWithContinousIndexingOnSourceCluster <<<
>[junit4]> Throwable #1: java.lang.AssertionError: Document mismatch on 
> target after sync expected:<2000> but was:<1901>
>[junit4]>  at 
> __randomizedtesting.SeedInfo.seed([41753A7BCCA7C778:953071222BF17483]:0)
>[junit4]>  at 
> org.apache.solr.cloud.CdcrBootstrapTest.testBootstrapWithContinousIndexingOnSourceCluster(CdcrBootstrapTest.java:309)
>[junit4]>  at java.lang.Thread.run(Thread.java:748)
> {code}
> ref: https://jenkins.thetaphi.de/job/Lucene-Solr-7.0-Linux/423/
> From one of the failed Solr jenkins log:
>[junit4]   2> 1143166 INFO  
> (cdcr-replicator-4421-thread-1-processing-n:127.0.0.1:62832_solr 
> x:cdcr-source_shard1_replica_n1 s:shard1 c:cdcr-source r:core_node2) 
> [n:127.0.0.1:62832_solr c:cdcr-source s:shard1 r:core_node2 
> x:cdcr-source_shard1_replica_n1] o.a.s.h.CdcrReplicator Forwarded 991 updates 
> to target cdcr-target
>[junit4]   2> 1144176 INFO  
> (cdcr-replicator-4421-thread-1-processing-n:127.0.0.1:62832_solr 
> x:cdcr-source_shard1_replica_n1 s:shard1 c:cdcr-source r:core_node2) 
> [n:127.0.0.1:62832_solr c:cdcr-source s:shard1 r:core_node2 
> x:cdcr-source_shard1_replica_n1] o.a.s.h.CdcrReplicator Forwarded 909 updates 
> to target cdcr-target
>[junit4]   2> 1145118 INFO  
> (cdcr-replicator-4421-thread-1-processing-n:127.0.0.1:62832_solr 
> x:cdcr-source_shard1_replica_n1 s:shard1 c:cdcr-source r:core_node2) 
> [n:127.0.0.1:62832_solr c:cdcr-source s:shard1 r:core_node2 
> x:cdcr-source_shard1_replica_n1] o.a.s.h.CdcrReplicator Forwarded 0 updates 
> to target cdcr-target
> Total 1900 updates were sent, instead of 2000. Ideally the bootstrap process 
> is responsible for 1000, and normal cdc replication is responsble for 1000. 
> On looking closely, the bootstrap is completed successfully. We are 100% sure 
> here, bootstrap worked w/o any issue. And still 1900 updates were sent via 
> replicator, instead of 1000. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11467) CdcrBootstrapTest::testBootstrapWithContinousIndexingOnSourceCluster Failure

2017-10-12 Thread Varun Thacker (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16203024#comment-16203024
 ] 

Varun Thacker commented on SOLR-11467:
--

Hi Amrit,

I'll commit the patch with additional logging shortly. 

Making a couple of changes before committing :

This probably doesn't need to change?
{code}
- System.out.println("Adding " + docs + " docs with commit=true, numDocs=" + 
numDocs);
+ System.out.println("Adding 100 docs with commit=true, numDocs=" + numDocs);
{code}

Adding "nocommit" will get precommit. So removing that comment and adding 
another placeholder

> CdcrBootstrapTest::testBootstrapWithContinousIndexingOnSourceCluster Failure
> 
>
> Key: SOLR-11467
> URL: https://issues.apache.org/jira/browse/SOLR-11467
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: CDCR
>Affects Versions: 6.6, 6.6.1, 7.0, 7.0.1, 7.1
>Reporter: Amrit Sarkar
> Attachments: SOLR-11467-debug-code.log
>
>
> CdcrBootstrapTest is still failing in master and other branches with:
> {code}
> [junit4] FAILURE  130s J1 | 
> CdcrBootstrapTest.testBootstrapWithContinousIndexingOnSourceCluster <<<
>[junit4]> Throwable #1: java.lang.AssertionError: Document mismatch on 
> target after sync expected:<2000> but was:<1901>
>[junit4]>  at 
> __randomizedtesting.SeedInfo.seed([41753A7BCCA7C778:953071222BF17483]:0)
>[junit4]>  at 
> org.apache.solr.cloud.CdcrBootstrapTest.testBootstrapWithContinousIndexingOnSourceCluster(CdcrBootstrapTest.java:309)
>[junit4]>  at java.lang.Thread.run(Thread.java:748)
> {code}
> ref: https://jenkins.thetaphi.de/job/Lucene-Solr-7.0-Linux/423/
> From one of the failed Solr jenkins log:
>[junit4]   2> 1143166 INFO  
> (cdcr-replicator-4421-thread-1-processing-n:127.0.0.1:62832_solr 
> x:cdcr-source_shard1_replica_n1 s:shard1 c:cdcr-source r:core_node2) 
> [n:127.0.0.1:62832_solr c:cdcr-source s:shard1 r:core_node2 
> x:cdcr-source_shard1_replica_n1] o.a.s.h.CdcrReplicator Forwarded 991 updates 
> to target cdcr-target
>[junit4]   2> 1144176 INFO  
> (cdcr-replicator-4421-thread-1-processing-n:127.0.0.1:62832_solr 
> x:cdcr-source_shard1_replica_n1 s:shard1 c:cdcr-source r:core_node2) 
> [n:127.0.0.1:62832_solr c:cdcr-source s:shard1 r:core_node2 
> x:cdcr-source_shard1_replica_n1] o.a.s.h.CdcrReplicator Forwarded 909 updates 
> to target cdcr-target
>[junit4]   2> 1145118 INFO  
> (cdcr-replicator-4421-thread-1-processing-n:127.0.0.1:62832_solr 
> x:cdcr-source_shard1_replica_n1 s:shard1 c:cdcr-source r:core_node2) 
> [n:127.0.0.1:62832_solr c:cdcr-source s:shard1 r:core_node2 
> x:cdcr-source_shard1_replica_n1] o.a.s.h.CdcrReplicator Forwarded 0 updates 
> to target cdcr-target
> Total 1900 updates were sent, instead of 2000. Ideally the bootstrap process 
> is responsible for 1000, and normal cdc replication is responsble for 1000. 
> On looking closely, the bootstrap is completed successfully. We are 100% sure 
> here, bootstrap worked w/o any issue. And still 1900 updates were sent via 
> replicator, instead of 1000. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-SmokeRelease-7.1 - Build # 1 - Failure

2017-10-12 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-SmokeRelease-7.1/1/

No tests ran.

Build Log:
[...truncated 28184 lines...]
ERROR: command execution failed.
ERROR: lucene is offline; cannot locate JDK 1.8 (latest)
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
ERROR: lucene is offline; cannot locate JDK 1.8 (latest)
ERROR: lucene is offline; cannot locate JDK 1.8 (latest)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (SOLR-11392) StreamExpressionTest.testParallelExecutorStream fails too frequently

2017-10-12 Thread Shalin Shekhar Mangar (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16203010#comment-16203010
 ] 

Shalin Shekhar Mangar commented on SOLR-11392:
--

bq. Notice this is looking for the "_n3" replica. What's odd about this is that 
only two replicas where created for this collection

Joel, sorry for the late reply. Solr does not create replica numbers 
sequentially anymore. This was done in SOLR-11011

> StreamExpressionTest.testParallelExecutorStream fails too frequently
> 
>
> Key: SOLR-11392
> URL: https://issues.apache.org/jira/browse/SOLR-11392
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Joel Bernstein
>
> I've never been able to reproduce the failure but jenkins fails frequently 
> with the following error:
> {code}
> Stack Trace:
> org.apache.solr.client.solrj.impl.CloudSolrClient$RouteException: Error from 
> server at http://127.0.0.1:38180/solr/workQueue_shard2_replica_n3: Expected 
> mime type application/octet-stream but got text/html. 
> 
> 
> Error 404 
> 
> 
> HTTP ERROR: 404
> Problem accessing /solr/workQueue_shard2_replica_n3/update. Reason:
> Can not find: /solr/workQueue_shard2_replica_n3/update
> http://eclipse.org/jetty";>Powered by Jetty:// 
> 9.3.20.v20170531
> 
> 
> {code}
> What appears to be happening is that the test framework is having trouble 
> setting up the collection.
> Here is the test code:
> {code}
> @Test
>   public void testParallelExecutorStream() throws Exception {
> CollectionAdminRequest.createCollection("workQueue", "conf", 2, 
> 1).process(cluster.getSolrClient());
> AbstractDistribZkTestBase.waitForRecoveriesToFinish("workQueue", 
> cluster.getSolrClient().getZkStateReader(),
> false, true, TIMEOUT);
> CollectionAdminRequest.createCollection("mainCorpus", "conf", 2, 
> 1).process(cluster.getSolrClient());
> AbstractDistribZkTestBase.waitForRecoveriesToFinish("mainCorpus", 
> cluster.getSolrClient().getZkStateReader(),
> false, true, TIMEOUT);
> CollectionAdminRequest.createCollection("destination", "conf", 2, 
> 1).process(cluster.getSolrClient());
> AbstractDistribZkTestBase.waitForRecoveriesToFinish("destination", 
> cluster.getSolrClient().getZkStateReader(),
> false, true, TIMEOUT);
> UpdateRequest workRequest = new UpdateRequest();
> UpdateRequest dataRequest = new UpdateRequest();
> for (int i = 0; i < 500; i++) {
>   workRequest.add(id, String.valueOf(i), "expr_s", "update(destination, 
> batchSize=50, search(mainCorpus, q=id:"+i+", rows=1, sort=\"id asc\", 
> fl=\"id, body_t, field_i\"))");
>   dataRequest.add(id, String.valueOf(i), "body_t", "hello world "+i, 
> "field_i", Integer.toString(i));
> }
> workRequest.commit(cluster.getSolrClient(), "workQueue");
> dataRequest.commit(cluster.getSolrClient(), "mainCorpus");
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-8389) Convert CDCR peer cluster and other configurations into collection properties modifiable via APIs

2017-10-12 Thread Varun Thacker (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-8389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16203007#comment-16203007
 ] 

Varun Thacker commented on SOLR-8389:
-

bq. ZkStateReader keeps a set of watchers for each of them (one per 
collection), in a similar way to state.json. But on top of caching the values, 
it provides a method where you can add listeners, that will be called each time 
the znode is changed.

Will it suffer from the same problem as putting the configs in state.json? 
Unrelated updates in the collection properties api might trigger a bootstrap ? 

 

> Convert CDCR peer cluster and other configurations into collection properties 
> modifiable via APIs
> -
>
> Key: SOLR-8389
> URL: https://issues.apache.org/jira/browse/SOLR-8389
> Project: Solr
>  Issue Type: Improvement
>  Components: CDCR, SolrCloud
>Reporter: Shalin Shekhar Mangar
>
> CDCR configuration is kept inside solrconfig.xml which makes it difficult to 
> add or change peer cluster configuration.
> I propose to move all CDCR config to collection level properties in cluster 
> state so that they can be modified using the existing modify collection API.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-11485) Add olsRegress, olsPredict and olsResiduals Stream Evaluators

2017-10-12 Thread Joel Bernstein (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11485?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joel Bernstein updated SOLR-11485:
--
Fix Version/s: master (8.0)
   7.2

> Add olsRegress, olsPredict and olsResiduals Stream Evaluators
> -
>
> Key: SOLR-11485
> URL: https://issues.apache.org/jira/browse/SOLR-11485
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Joel Bernstein
>Assignee: Joel Bernstein
> Fix For: 7.2, master (8.0)
>
>
> This ticket adds support for OLS (ordinary least squares)  multiple 
> regression, prediction and residuals.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (SOLR-11485) Add olsRegress, olsPredict and olsResiduals Stream Evaluators

2017-10-12 Thread Joel Bernstein (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11485?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joel Bernstein reassigned SOLR-11485:
-

Assignee: Joel Bernstein

> Add olsRegress, olsPredict and olsResiduals Stream Evaluators
> -
>
> Key: SOLR-11485
> URL: https://issues.apache.org/jira/browse/SOLR-11485
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Joel Bernstein
>Assignee: Joel Bernstein
> Fix For: 7.2, master (8.0)
>
>
> This ticket adds support for OLS (ordinary least squares)  multiple 
> regression, prediction and residuals.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-11485) Add olsRegress, olsPredict and olsResiduals Stream Evaluators

2017-10-12 Thread Joel Bernstein (JIRA)
Joel Bernstein created SOLR-11485:
-

 Summary: Add olsRegress, olsPredict and olsResiduals Stream 
Evaluators
 Key: SOLR-11485
 URL: https://issues.apache.org/jira/browse/SOLR-11485
 Project: Solr
  Issue Type: New Feature
  Security Level: Public (Default Security Level. Issues are Public)
Reporter: Joel Bernstein


This ticket adds support for OLS (ordinary least squares)  multiple regression, 
prediction and residuals.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-11445) Overseer should not hang when process bad message

2017-10-12 Thread Cao Manh Dat (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Cao Manh Dat resolved SOLR-11445.
-
   Resolution: Fixed
 Assignee: Cao Manh Dat
Fix Version/s: master (8.0)
   7.2

> Overseer should not hang when process bad message
> -
>
> Key: SOLR-11445
> URL: https://issues.apache.org/jira/browse/SOLR-11445
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Affects Versions: 6.6.1, 7.0, master (8.0)
>Reporter: Greg Harris
>Assignee: Cao Manh Dat
> Fix For: 7.2, master (8.0)
>
> Attachments: SOLR-11445.patch, SOLR-11445.patch
>
>
> So we had the following stack trace with a customer:
> 2017-10-04 11:25:30.339 ERROR () [ ] o.a.s.c.Overseer Exception in 
> Overseer main queue loop
> org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = 
> NoNode for /collections//state.json
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
> at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:391)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:388)
> at 
> org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60)
> at org.apache.solr.common.cloud.SolrZkClient.create(SolrZkClient.java:388)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.writePendingUpdates(ZkStateWriter.java:235)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.run(Overseer.java:199)
> at java.lang.Thread.run(Thread.java:748)
> I want to highlight: 
>   at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> This ends up coming from Overseer:
> while (data != null)  {
> final ZkNodeProps message = ZkNodeProps.load(data);
> log.debug("processMessage: workQueueSize: {}, message = {}", 
> workQueue.getStats().getQueueLength(), message);
> // force flush to ZK after each message because there is no 
> fallback if workQueue items
> // are removed from workQueue but fail to be written to ZK
> *clusterState = processQueueItem(message, clusterState, 
> zkStateWriter, false, null);
> workQueue.poll(); // poll-ing removes the element we got by 
> peek-ing*
> data = workQueue.peek();
>   }
> Note: The processQueueItem comes before the poll, therefore upon a thrown 
> exception the same node/message that won't process becomes stuck. This made a 
> large cluster unable to come up on it's own without deleting the problem 
> node. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11445) Overseer should not hang when process bad message

2017-10-12 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11445?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202914#comment-16202914
 ] 

ASF subversion and git services commented on SOLR-11445:


Commit fc981dd5acee6962ff1d35e8238af4e47829e4b5 in lucene-solr's branch 
refs/heads/branch_7x from [~caomanhdat]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=fc981dd ]

SOLR-11445: Overseer should not hang when process bad message


> Overseer should not hang when process bad message
> -
>
> Key: SOLR-11445
> URL: https://issues.apache.org/jira/browse/SOLR-11445
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Affects Versions: 6.6.1, 7.0, master (8.0)
>Reporter: Greg Harris
> Attachments: SOLR-11445.patch, SOLR-11445.patch
>
>
> So we had the following stack trace with a customer:
> 2017-10-04 11:25:30.339 ERROR () [ ] o.a.s.c.Overseer Exception in 
> Overseer main queue loop
> org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = 
> NoNode for /collections//state.json
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
> at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:391)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:388)
> at 
> org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60)
> at org.apache.solr.common.cloud.SolrZkClient.create(SolrZkClient.java:388)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.writePendingUpdates(ZkStateWriter.java:235)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.run(Overseer.java:199)
> at java.lang.Thread.run(Thread.java:748)
> I want to highlight: 
>   at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> This ends up coming from Overseer:
> while (data != null)  {
> final ZkNodeProps message = ZkNodeProps.load(data);
> log.debug("processMessage: workQueueSize: {}, message = {}", 
> workQueue.getStats().getQueueLength(), message);
> // force flush to ZK after each message because there is no 
> fallback if workQueue items
> // are removed from workQueue but fail to be written to ZK
> *clusterState = processQueueItem(message, clusterState, 
> zkStateWriter, false, null);
> workQueue.poll(); // poll-ing removes the element we got by 
> peek-ing*
> data = workQueue.peek();
>   }
> Note: The processQueueItem comes before the poll, therefore upon a thrown 
> exception the same node/message that won't process becomes stuck. This made a 
> large cluster unable to come up on it's own without deleting the problem 
> node. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11445) Overseer should not hang when process bad message

2017-10-12 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11445?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202915#comment-16202915
 ] 

ASF subversion and git services commented on SOLR-11445:


Commit 2795287b0e56148102ae802259a59d80907e4d1a in lucene-solr's branch 
refs/heads/branch_7x from [~caomanhdat]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=2795287 ]

SOLR-11445: Upgrade CHANGES.txt


> Overseer should not hang when process bad message
> -
>
> Key: SOLR-11445
> URL: https://issues.apache.org/jira/browse/SOLR-11445
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Affects Versions: 6.6.1, 7.0, master (8.0)
>Reporter: Greg Harris
> Attachments: SOLR-11445.patch, SOLR-11445.patch
>
>
> So we had the following stack trace with a customer:
> 2017-10-04 11:25:30.339 ERROR () [ ] o.a.s.c.Overseer Exception in 
> Overseer main queue loop
> org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = 
> NoNode for /collections//state.json
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
> at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:391)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:388)
> at 
> org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60)
> at org.apache.solr.common.cloud.SolrZkClient.create(SolrZkClient.java:388)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.writePendingUpdates(ZkStateWriter.java:235)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.run(Overseer.java:199)
> at java.lang.Thread.run(Thread.java:748)
> I want to highlight: 
>   at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> This ends up coming from Overseer:
> while (data != null)  {
> final ZkNodeProps message = ZkNodeProps.load(data);
> log.debug("processMessage: workQueueSize: {}, message = {}", 
> workQueue.getStats().getQueueLength(), message);
> // force flush to ZK after each message because there is no 
> fallback if workQueue items
> // are removed from workQueue but fail to be written to ZK
> *clusterState = processQueueItem(message, clusterState, 
> zkStateWriter, false, null);
> workQueue.poll(); // poll-ing removes the element we got by 
> peek-ing*
> data = workQueue.peek();
>   }
> Note: The processQueueItem comes before the poll, therefore upon a thrown 
> exception the same node/message that won't process becomes stuck. This made a 
> large cluster unable to come up on it's own without deleting the problem 
> node. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11445) Overseer should not hang when process bad message

2017-10-12 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11445?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202912#comment-16202912
 ] 

ASF subversion and git services commented on SOLR-11445:


Commit 4bc7fd2bd54c6a06f88c9aa2de9d3b94f5a77efb in lucene-solr's branch 
refs/heads/master from [~caomanhdat]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=4bc7fd2 ]

SOLR-11445: Upgrade CHANGES.txt


> Overseer should not hang when process bad message
> -
>
> Key: SOLR-11445
> URL: https://issues.apache.org/jira/browse/SOLR-11445
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Affects Versions: 6.6.1, 7.0, master (8.0)
>Reporter: Greg Harris
> Attachments: SOLR-11445.patch, SOLR-11445.patch
>
>
> So we had the following stack trace with a customer:
> 2017-10-04 11:25:30.339 ERROR () [ ] o.a.s.c.Overseer Exception in 
> Overseer main queue loop
> org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = 
> NoNode for /collections//state.json
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
> at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:391)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:388)
> at 
> org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60)
> at org.apache.solr.common.cloud.SolrZkClient.create(SolrZkClient.java:388)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.writePendingUpdates(ZkStateWriter.java:235)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.run(Overseer.java:199)
> at java.lang.Thread.run(Thread.java:748)
> I want to highlight: 
>   at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> This ends up coming from Overseer:
> while (data != null)  {
> final ZkNodeProps message = ZkNodeProps.load(data);
> log.debug("processMessage: workQueueSize: {}, message = {}", 
> workQueue.getStats().getQueueLength(), message);
> // force flush to ZK after each message because there is no 
> fallback if workQueue items
> // are removed from workQueue but fail to be written to ZK
> *clusterState = processQueueItem(message, clusterState, 
> zkStateWriter, false, null);
> workQueue.poll(); // poll-ing removes the element we got by 
> peek-ing*
> data = workQueue.peek();
>   }
> Note: The processQueueItem comes before the poll, therefore upon a thrown 
> exception the same node/message that won't process becomes stuck. This made a 
> large cluster unable to come up on it's own without deleting the problem 
> node. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-11484) Possible bug with CloudSolrClient directedUpdates & cached collection state -- TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete

2017-10-12 Thread Joel Bernstein (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202881#comment-16202881
 ] 

Joel Bernstein edited comment on SOLR-11484 at 10/13/17 12:50 AM:
--

[~hossman], [~noble.paul], this is related I believe to SOLR-11392.

[~romseygeek], discusses the issue in the ticket.


was (Author: joel.bernstein):
[~hossman], [~noble.paul], this is related I believe: SOLR-11392.

[~romseygeek], discusses the issue in the ticket.

> Possible bug with CloudSolrClient directedUpdates & cached collection state 
> -- TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete
> -
>
> Key: SOLR-11484
> URL: https://issues.apache.org/jira/browse/SOLR-11484
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Hoss Man
> Attachments: jenkins.thetaphi.20662.txt
>
>
> {{TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete}} 
> seems to fail with non-trivial frequency, so I grabbed the logs from a recent 
> failure and starting trying to follow along with the actions to figure out 
> what exactly is happening
> https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/20662/
> {noformat}
>[junit4] ERROR   20.3s J1 | 
> TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete <<<
>[junit4]> Throwable #1: 
> org.apache.solr.client.solrj.impl.CloudSolrClient$RouteException: Error from 
> server at https://127.0.0.1:42959/solr/testcollection_shard1_replica_n3: 
> Expected mime type a
> pplication/octet-stream but got text/html. 
>[junit4]> 
>[junit4]>  content="text/html;charset=ISO-8859-1"/>
>[junit4]> Error 404 
> {noformat}
> The crux of this failure appears to be a genuine bug in how CloudSolrClient 
> uses it's cached ClusterState info when doing (direct) updates.  The key bits 
> seem to be:
> * CloudSolrClient does _something_ (update,query,etc...) with a collection 
> causing the current cluster state for the collection to be cached
> * The actual collection changes such that a Solr node/core no longer exists 
> as part of the collection
> * CloudSolrClient is asked to process an UpdateRequest which triggers the 
> code paths for the {{directUpdate()}} method -- which attempts to route the 
> updates directly to a replica of the appropriate shard using the (cache) 
> collection state info
> * CloudSolrClient (may) attempt to send that UpdateRequest to a node/core 
> that doesn't exist, getting a 404 -- which does not (seem to) trigger a state 
> refresh, or retry to find a correct URL to resend the update to.
> Details to follow in comment



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11484) Possible bug with CloudSolrClient directedUpdates & cached collection state -- TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete

2017-10-12 Thread Joel Bernstein (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202881#comment-16202881
 ] 

Joel Bernstein commented on SOLR-11484:
---

[~hossman], [~noble.paul], this is related I believe: SOLR-11392.

[~romseygeek], discusses the issue in the ticket.

> Possible bug with CloudSolrClient directedUpdates & cached collection state 
> -- TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete
> -
>
> Key: SOLR-11484
> URL: https://issues.apache.org/jira/browse/SOLR-11484
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Hoss Man
> Attachments: jenkins.thetaphi.20662.txt
>
>
> {{TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete}} 
> seems to fail with non-trivial frequency, so I grabbed the logs from a recent 
> failure and starting trying to follow along with the actions to figure out 
> what exactly is happening
> https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/20662/
> {noformat}
>[junit4] ERROR   20.3s J1 | 
> TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete <<<
>[junit4]> Throwable #1: 
> org.apache.solr.client.solrj.impl.CloudSolrClient$RouteException: Error from 
> server at https://127.0.0.1:42959/solr/testcollection_shard1_replica_n3: 
> Expected mime type a
> pplication/octet-stream but got text/html. 
>[junit4]> 
>[junit4]>  content="text/html;charset=ISO-8859-1"/>
>[junit4]> Error 404 
> {noformat}
> The crux of this failure appears to be a genuine bug in how CloudSolrClient 
> uses it's cached ClusterState info when doing (direct) updates.  The key bits 
> seem to be:
> * CloudSolrClient does _something_ (update,query,etc...) with a collection 
> causing the current cluster state for the collection to be cached
> * The actual collection changes such that a Solr node/core no longer exists 
> as part of the collection
> * CloudSolrClient is asked to process an UpdateRequest which triggers the 
> code paths for the {{directUpdate()}} method -- which attempts to route the 
> updates directly to a replica of the appropriate shard using the (cache) 
> collection state info
> * CloudSolrClient (may) attempt to send that UpdateRequest to a node/core 
> that doesn't exist, getting a 404 -- which does not (seem to) trigger a state 
> refresh, or retry to find a correct URL to resend the update to.
> Details to follow in comment



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-11484) Possible bug with CloudSolrClient directedUpdates & cached collection state -- TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete

2017-10-12 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-11484:

Attachment: jenkins.thetaphi.20662.txt

I've attached the log from 
https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/20662/

A quick walk through of some of the key bits of logging from 
testCollectionCreateSearchDelete ...


Test Creates a 2x2 collection, resulting in 4 SolrCores...

{noformat}   
   [junit4]   2> 334246 INFO  
(parallelCoreAdminExecutor-1592-thread-1-processing-n:127.0.0.1:42959_solr 
aecb57e9-be15-41b4-892b-269c3ef4df623724636310642455 CREATE) 
[n:127.0.0.1:42959_solr] o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&collection.configName=solrCloudCollectionConfig&newCollection=true&collection=testcollection&version=2&replicaType=NRT&async=aecb57e9-be15-41b4-892b-269c3ef4df623724636310642455&coreNodeName=core_node4&name=testcollection_shard1_replica_n3&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 334261 INFO  
(parallelCoreAdminExecutor-1581-thread-1-processing-n:127.0.0.1:34827_solr 
aecb57e9-be15-41b4-892b-269c3ef4df623724636321671098 CREATE) 
[n:127.0.0.1:34827_solr] o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&collection.configName=solrCloudCollectionConfig&newCollection=true&collection=testcollection&version=2&replicaType=NRT&async=aecb57e9-be15-41b4-892b-269c3ef4df623724636321671098&coreNodeName=core_node8&name=testcollection_shard2_replica_n6&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 334272 INFO  
(parallelCoreAdminExecutor-1579-thread-1-processing-n:127.0.0.1:38927_solr 
aecb57e9-be15-41b4-892b-269c3ef4df623724636317202530 CREATE) 
[n:127.0.0.1:38927_solr] o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&collection.configName=solrCloudCollectionConfig&newCollection=true&collection=testcollection&version=2&replicaType=NRT&async=aecb57e9-be15-41b4-892b-269c3ef4df623724636317202530&coreNodeName=core_node7&name=testcollection_shard2_replica_n5&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 334272 INFO  
(parallelCoreAdminExecutor-1577-thread-1-processing-n:127.0.0.1:36233_solr 
aecb57e9-be15-41b4-892b-269c3ef4df623724636304239736 CREATE) 
[n:127.0.0.1:36233_solr] o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&collection.configName=solrCloudCollectionConfig&newCollection=true&collection=testcollection&version=2&replicaType=NRT&async=aecb57e9-be15-41b4-892b-269c3ef4df623724636304239736&coreNodeName=core_node2&name=testcollection_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
{noformat}

Note that these initial Cores are:

* testcollection_shard1_replica_n3
* testcollection_shard2_replica_n6
* testcollection_shard2_replica_n5
* testcollection_shard1_replica_n1

...the test then does some searching of this collection, and a few other 
things, before ultimately deleting this collection, and verifies the delete 
worked correctly...

{noformat}
   
   [junit4]   2> 340530 INFO  
(TEST-TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete-seed#[DC2FFDE51D1E4138])
 [] o.a.s.c.AbstractDistribZkTestBase Wait for collection to disappear - 
collection: testcollection failOnTimeout:true timeout (sec):330
   [junit4]   1> -
   [junit4]   2> 340530 INFO  
(TEST-TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete-seed#[DC2FFDE51D1E4138])
 [] o.a.s.c.AbstractDistribZkTestBase Collection has disappeared - 
collection: testcollection
{noformat}

...at which point it recreates the collection, and we get 4 completely new 
SolrCores...

{noformat}
   [junit4]   2> 340531 INFO  (qtp6784903-4183) [n:127.0.0.1:38927_solr] 
o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params 
replicationFactor=2&collection.configName=solrCloudCollectionConfig&maxShardsPerNode=1&name=testcollection&nrtReplicas=2&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2
 and sendToOCPQueue=true
   [junit4]   2> 340533 INFO  
(OverseerThreadFactory-1587-thread-3-processing-n:127.0.0.1:34827_solr) 
[n:127.0.0.1:34827_solr] o.a.s.c.CreateCollectionCmd Create collection 
testcollection
   
   [junit4]   2> 340948 INFO  (qtp13197426-4184) [n:127.0.0.1:36233_solr] 
o.a.s.h.a.CoreAdminOperation core create command 
qt=/admin/cores&collection.configName=solrCloudCollectionConfig&newCollection=true&collection=testcollection&version=2&replicaType=NRT&coreNodeName=core_node8&name=testcollection_shard2_replica_n6&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDi

[jira] [Created] (SOLR-11484) Possible bug with CloudSolrClient directedUpdates & cached collection state -- TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete

2017-10-12 Thread Hoss Man (JIRA)
Hoss Man created SOLR-11484:
---

 Summary: Possible bug with CloudSolrClient directedUpdates & 
cached collection state -- 
TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete
 Key: SOLR-11484
 URL: https://issues.apache.org/jira/browse/SOLR-11484
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
Reporter: Hoss Man



{{TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete}} 
seems to fail with non-trivial frequency, so I grabbed the logs from a recent 
failure and starting trying to follow along with the actions to figure out what 
exactly is happening

https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/20662/

{noformat}
   [junit4] ERROR   20.3s J1 | 
TestCollectionsAPIViaSolrCloudCluster.testCollectionCreateSearchDelete <<<
   [junit4]> Throwable #1: 
org.apache.solr.client.solrj.impl.CloudSolrClient$RouteException: Error from 
server at https://127.0.0.1:42959/solr/testcollection_shard1_replica_n3: 
Expected mime type a
pplication/octet-stream but got text/html. 
   [junit4]> 
   [junit4]> 
   [junit4]> Error 404 
{noformat}

The crux of this failure appears to be a genuine bug in how CloudSolrClient 
uses it's cached ClusterState info when doing (direct) updates.  The key bits 
seem to be:

* CloudSolrClient does _something_ (update,query,etc...) with a collection 
causing the current cluster state for the collection to be cached
* The actual collection changes such that a Solr node/core no longer exists as 
part of the collection
* CloudSolrClient is asked to process an UpdateRequest which triggers the code 
paths for the {{directUpdate()}} method -- which attempts to route the updates 
directly to a replica of the appropriate shard using the (cache) collection 
state info
* CloudSolrClient (may) attempt to send that UpdateRequest to a node/core that 
doesn't exist, getting a 404 -- which does not (seem to) trigger a state 
refresh, or retry to find a correct URL to resend the update to.

Details to follow in comment




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-NightlyTests-master - Build # 1401 - Still unstable

2017-10-12 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-master/1401/

8 tests failed.
FAILED:  org.apache.solr.cloud.TestLocalFSCloudBackupRestore.test

Error Message:
expected: but was:

Stack Trace:
java.lang.AssertionError: expected: but was:
at 
__randomizedtesting.SeedInfo.seed([317B471B1312A108:B92F78C1BDEECCF0]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.failNotEquals(Assert.java:647)
at org.junit.Assert.assertEquals(Assert.java:128)
at org.junit.Assert.assertEquals(Assert.java:147)
at 
org.apache.solr.cloud.AbstractCloudBackupRestoreTestCase.testBackupAndRestore(AbstractCloudBackupRestoreTestCase.java:277)
at 
org.apache.solr.cloud.AbstractCloudBackupRestoreTestCase.test(AbstractCloudBackupRestoreTestCase.java:136)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)


FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.TriLevelCompositeIdRoutingTest

Error Message:
1 

[jira] [Commented] (SOLR-11481) Ref guide page explaining nuances of the recovery process

2017-10-12 Thread Varun Thacker (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202739#comment-16202739
 ] 

Varun Thacker commented on SOLR-11481:
--

We document the params in 
http://lucene.apache.org/solr/guide/updatehandlers-in-solrconfig.html#transaction-log
 but it doesn't explain how they interplay 

> Ref guide page explaining nuances of the recovery process
> -
>
> Key: SOLR-11481
> URL: https://issues.apache.org/jira/browse/SOLR-11481
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Varun Thacker
>Priority: Minor
>
> The Solr recovery process involves PeerSync , which has configuration 
> parameters to allow the number of records it should keep.
> If this fails we do a index replication where possibly we can throttle 
> replication 
> I think it's worth explaining to users what these configuration parameters 
> are and how does a node actually recover. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-11483) Keep more transaction log files when maxNumLogsToKeep is specified

2017-10-12 Thread Varun Thacker (JIRA)
Varun Thacker created SOLR-11483:


 Summary: Keep more transaction log files when maxNumLogsToKeep is 
specified
 Key: SOLR-11483
 URL: https://issues.apache.org/jira/browse/SOLR-11483
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
Reporter: Varun Thacker


We allow users to control the number of transaction log files the system should 
keep ( maxNumLogsToKeep ) and the number of records to keep ( numRecordsToKeep )

However just by setting numRecordsToKeep we are not guaranteed that the system 
will actually keep that many records.  

This comment explains why - 
https://issues.apache.org/jira/browse/SOLR-6359?focusedCommentId=15214361&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-15214361

A tradeoff would be if a user explicitly specifies  numRecordsToKeep and 
doesn't specify maxNumLogsToKeep then we should use a higher default for 
maxNumLogsToKeep ( say 1000 instead of 10 today )

Off course better documentation will help so if others think this is a bad idea 
then we can close this out and just make sure it's documented and if users know 
about one parameter they will know about the other parameter as well



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11481) Ref guide page explaining nuances of the recovery process

2017-10-12 Thread Varun Thacker (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202731#comment-16202731
 ] 

Varun Thacker commented on SOLR-11481:
--

{code}

  ${solr.ulog.dir:}
   1

{code}

One might assume that if he sets numRecordsToKeep that PeerSync will actually 
try to sync upto these many records. But looking at the code we only keep 10 
transaction log files by default even when when this number is set. 

Yonik explains on SOLR-6359 that this is because we don't want to run out of 
file handles in the corner case of add 1 doc + commit , add 1 doc + commit 
case. 

So we also need to add specify maxNumLogsToKeep and keep it to a high number as 
well . 

Honestly my first reaction is that to accommodate users who are explicitly 
specifying a high numRecordsToKeep and then doing commits after every add we 
are choosing trappy behaviour for the remaining 99% users is a bad tradeoff. 
I'll file a separate Jira to see where that goes. 

So for PeerSync we need to document both maxNumLogsToKeep and numRecordsToKeep 
and how they help improve recovery times.



> Ref guide page explaining nuances of the recovery process
> -
>
> Key: SOLR-11481
> URL: https://issues.apache.org/jira/browse/SOLR-11481
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Varun Thacker
>Priority: Minor
>
> The Solr recovery process involves PeerSync , which has configuration 
> parameters to allow the number of records it should keep.
> If this fails we do a index replication where possibly we can throttle 
> replication 
> I think it's worth explaining to users what these configuration parameters 
> are and how does a node actually recover. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (LUCENE-7538) Uploading large text file to a field causes "this IndexWriter is closed" error

2017-10-12 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202705#comment-16202705
 ] 

Michael McCandless edited comment on LUCENE-7538 at 10/12/17 10:14 PM:
---

[~ashah301182] see {{IndexWriter.MAX_STORED_STRING_LENGTH}} for the limit; 
you'll have to either not store that huge file, or break it up across several 
documents.


was (Author: mikemccand):
[~nkeet] see {{IndexWriter.MAX_STORED_STRING_LENGTH}} for the limit; you'll 
have to either not store that huge file, or break it up across several 
documents.

> Uploading large text file to a field causes "this IndexWriter is closed" error
> --
>
> Key: LUCENE-7538
> URL: https://issues.apache.org/jira/browse/LUCENE-7538
> Project: Lucene - Core
>  Issue Type: Bug
>Affects Versions: 5.5.1
>Reporter: Steve Chen
> Fix For: 6.4, 7.0
>
> Attachments: LUCENE-7538.patch, LUCENE-7538.patch
>
>
> We have seen "this IndexWriter is closed" error after we tried to upload a 
> large text file to a single Solr text field. The field definition in the 
> schema.xml is:
> {noformat}
>  termVectors="true" termPositions="true" termOffsets="true"/>
> {noformat}
> After that, the IndexWriter remained closed and couldn't be recovered until 
> we reloaded the Solr core.  The text file had size of 800MB, containing only 
> numbers and English characters.
> Stack trace is shown below:
> {noformat}
> 2016-11-02 23:00:17,913 [http-nio-19082-exec-3] ERROR 
> org.apache.solr.handler.RequestHandlerBase - 
> org.apache.solr.common.SolrException: Exception writing document id 1487_0_1 
> to the index; possible analysis error.
> at 
> org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:180)
> at 
> org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:68)
> at 
> org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:48)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalAdd(DistributedUpdateProcessor.java:934)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:1089)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:712)
> at 
> org.apache.solr.update.processor.LogUpdateProcessorFactory$LogUpdateProcessor.processAdd(LogUpdateProcessorFactory.java:103)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.doAdd(ExtractingDocumentLoader.java:126)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.addDoc(ExtractingDocumentLoader.java:131)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.load(ExtractingDocumentLoader.java:237)
> at 
> org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:69)
> at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:155)
> at org.apache.solr.core.SolrCore.execute(SolrCore.java:2082)
> at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:651)
> at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:458)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:229)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:184)
> at 
> veeva.ecm.common.interfaces.web.SolrDispatchOverride.doFilter(SolrDispatchOverride.java:43)
> at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
> at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
> at 
> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:212)
> at 
> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:106)
> at 
> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:141)
> at 
> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79)
> at 
> org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:616)
> at 
> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88)
> at 
> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:521)
> at 
> org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1096)
> at 
> org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:674)
> at 
> org.apache.tomcat.util.

[jira] [Comment Edited] (LUCENE-7538) Uploading large text file to a field causes "this IndexWriter is closed" error

2017-10-12 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202705#comment-16202705
 ] 

Michael McCandless edited comment on LUCENE-7538 at 10/12/17 10:13 PM:
---

[~nkeet] see {{IndexWriter.MAX_STORED_STRING_LENGTH}} for the limit; you'll 
have to either not store that huge file, or break it up across several 
documents.


was (Author: mikemccand):
@nkeet see {{IndexWriter.MAX_STORED_STRING_LENGTH}} for the limit; you'll have 
to either not store that huge file, or break it up across several documents.

> Uploading large text file to a field causes "this IndexWriter is closed" error
> --
>
> Key: LUCENE-7538
> URL: https://issues.apache.org/jira/browse/LUCENE-7538
> Project: Lucene - Core
>  Issue Type: Bug
>Affects Versions: 5.5.1
>Reporter: Steve Chen
> Fix For: 6.4, 7.0
>
> Attachments: LUCENE-7538.patch, LUCENE-7538.patch
>
>
> We have seen "this IndexWriter is closed" error after we tried to upload a 
> large text file to a single Solr text field. The field definition in the 
> schema.xml is:
> {noformat}
>  termVectors="true" termPositions="true" termOffsets="true"/>
> {noformat}
> After that, the IndexWriter remained closed and couldn't be recovered until 
> we reloaded the Solr core.  The text file had size of 800MB, containing only 
> numbers and English characters.
> Stack trace is shown below:
> {noformat}
> 2016-11-02 23:00:17,913 [http-nio-19082-exec-3] ERROR 
> org.apache.solr.handler.RequestHandlerBase - 
> org.apache.solr.common.SolrException: Exception writing document id 1487_0_1 
> to the index; possible analysis error.
> at 
> org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:180)
> at 
> org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:68)
> at 
> org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:48)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalAdd(DistributedUpdateProcessor.java:934)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:1089)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:712)
> at 
> org.apache.solr.update.processor.LogUpdateProcessorFactory$LogUpdateProcessor.processAdd(LogUpdateProcessorFactory.java:103)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.doAdd(ExtractingDocumentLoader.java:126)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.addDoc(ExtractingDocumentLoader.java:131)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.load(ExtractingDocumentLoader.java:237)
> at 
> org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:69)
> at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:155)
> at org.apache.solr.core.SolrCore.execute(SolrCore.java:2082)
> at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:651)
> at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:458)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:229)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:184)
> at 
> veeva.ecm.common.interfaces.web.SolrDispatchOverride.doFilter(SolrDispatchOverride.java:43)
> at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
> at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
> at 
> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:212)
> at 
> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:106)
> at 
> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:141)
> at 
> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79)
> at 
> org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:616)
> at 
> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88)
> at 
> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:521)
> at 
> org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1096)
> at 
> org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:674)
> at 
> org.apache.tomcat.util.net.NioEn

[jira] [Commented] (LUCENE-7538) Uploading large text file to a field causes "this IndexWriter is closed" error

2017-10-12 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202707#comment-16202707
 ] 

Erick Erickson commented on LUCENE-7538:


There are probably some a-priori limits, somewhere  someplace lots of data has 
to be kept. The document has to be indexed in memory-resident segments 
(essentially) before being flushed. I question whether it's reasonable to even 
try to index documents this large.

>From a user's perspective, what's the value in indexing such a huge document? 
>Assuming it's text, it's quite likely that it'll be found for almost every 
>search... and have such a low score that it won't be seen by the users. It 
>can't be displayed in a browser, it's too big. Even transmitting it to the 
>client is prohibitive.

So I'd ask a little different question: "What's the largest document it makes 
sense to index?". Personally I think it's far, far less than a gigabyte


> Uploading large text file to a field causes "this IndexWriter is closed" error
> --
>
> Key: LUCENE-7538
> URL: https://issues.apache.org/jira/browse/LUCENE-7538
> Project: Lucene - Core
>  Issue Type: Bug
>Affects Versions: 5.5.1
>Reporter: Steve Chen
> Fix For: 6.4, 7.0
>
> Attachments: LUCENE-7538.patch, LUCENE-7538.patch
>
>
> We have seen "this IndexWriter is closed" error after we tried to upload a 
> large text file to a single Solr text field. The field definition in the 
> schema.xml is:
> {noformat}
>  termVectors="true" termPositions="true" termOffsets="true"/>
> {noformat}
> After that, the IndexWriter remained closed and couldn't be recovered until 
> we reloaded the Solr core.  The text file had size of 800MB, containing only 
> numbers and English characters.
> Stack trace is shown below:
> {noformat}
> 2016-11-02 23:00:17,913 [http-nio-19082-exec-3] ERROR 
> org.apache.solr.handler.RequestHandlerBase - 
> org.apache.solr.common.SolrException: Exception writing document id 1487_0_1 
> to the index; possible analysis error.
> at 
> org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:180)
> at 
> org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:68)
> at 
> org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:48)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalAdd(DistributedUpdateProcessor.java:934)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:1089)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:712)
> at 
> org.apache.solr.update.processor.LogUpdateProcessorFactory$LogUpdateProcessor.processAdd(LogUpdateProcessorFactory.java:103)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.doAdd(ExtractingDocumentLoader.java:126)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.addDoc(ExtractingDocumentLoader.java:131)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.load(ExtractingDocumentLoader.java:237)
> at 
> org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:69)
> at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:155)
> at org.apache.solr.core.SolrCore.execute(SolrCore.java:2082)
> at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:651)
> at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:458)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:229)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:184)
> at 
> veeva.ecm.common.interfaces.web.SolrDispatchOverride.doFilter(SolrDispatchOverride.java:43)
> at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
> at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
> at 
> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:212)
> at 
> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:106)
> at 
> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:141)
> at 
> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79)
> at 
> org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:616)
> at 
> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88)
> 

[jira] [Commented] (LUCENE-7538) Uploading large text file to a field causes "this IndexWriter is closed" error

2017-10-12 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202705#comment-16202705
 ] 

Michael McCandless commented on LUCENE-7538:


@nkeet see {{IndexWriter.MAX_STORED_STRING_LENGTH}} for the limit; you'll have 
to either not store that huge file, or break it up across several documents.

> Uploading large text file to a field causes "this IndexWriter is closed" error
> --
>
> Key: LUCENE-7538
> URL: https://issues.apache.org/jira/browse/LUCENE-7538
> Project: Lucene - Core
>  Issue Type: Bug
>Affects Versions: 5.5.1
>Reporter: Steve Chen
> Fix For: 6.4, 7.0
>
> Attachments: LUCENE-7538.patch, LUCENE-7538.patch
>
>
> We have seen "this IndexWriter is closed" error after we tried to upload a 
> large text file to a single Solr text field. The field definition in the 
> schema.xml is:
> {noformat}
>  termVectors="true" termPositions="true" termOffsets="true"/>
> {noformat}
> After that, the IndexWriter remained closed and couldn't be recovered until 
> we reloaded the Solr core.  The text file had size of 800MB, containing only 
> numbers and English characters.
> Stack trace is shown below:
> {noformat}
> 2016-11-02 23:00:17,913 [http-nio-19082-exec-3] ERROR 
> org.apache.solr.handler.RequestHandlerBase - 
> org.apache.solr.common.SolrException: Exception writing document id 1487_0_1 
> to the index; possible analysis error.
> at 
> org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:180)
> at 
> org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:68)
> at 
> org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:48)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalAdd(DistributedUpdateProcessor.java:934)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:1089)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:712)
> at 
> org.apache.solr.update.processor.LogUpdateProcessorFactory$LogUpdateProcessor.processAdd(LogUpdateProcessorFactory.java:103)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.doAdd(ExtractingDocumentLoader.java:126)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.addDoc(ExtractingDocumentLoader.java:131)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.load(ExtractingDocumentLoader.java:237)
> at 
> org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:69)
> at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:155)
> at org.apache.solr.core.SolrCore.execute(SolrCore.java:2082)
> at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:651)
> at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:458)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:229)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:184)
> at 
> veeva.ecm.common.interfaces.web.SolrDispatchOverride.doFilter(SolrDispatchOverride.java:43)
> at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
> at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
> at 
> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:212)
> at 
> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:106)
> at 
> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:141)
> at 
> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79)
> at 
> org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:616)
> at 
> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88)
> at 
> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:521)
> at 
> org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1096)
> at 
> org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:674)
> at 
> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1500)
> at 
> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1456)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1

[jira] [Updated] (SOLR-11481) Ref guide page explaining nuances of the recovery process

2017-10-12 Thread Varun Thacker (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11481?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Varun Thacker updated SOLR-11481:
-
Priority: Minor  (was: Major)

> Ref guide page explaining nuances of the recovery process
> -
>
> Key: SOLR-11481
> URL: https://issues.apache.org/jira/browse/SOLR-11481
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Varun Thacker
>Priority: Minor
>
> The Solr recovery process involves PeerSync , which has configuration 
> parameters to allow the number of records it should keep.
> If this fails we do a index replication where possibly we can throttle 
> replication 
> I think it's worth explaining to users what these configuration parameters 
> are and how does a node actually recover. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-11481) Ref guide page explaining nuances of the recovery process

2017-10-12 Thread Varun Thacker (JIRA)
Varun Thacker created SOLR-11481:


 Summary: Ref guide page explaining nuances of the recovery process
 Key: SOLR-11481
 URL: https://issues.apache.org/jira/browse/SOLR-11481
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
Reporter: Varun Thacker


The Solr recovery process involves PeerSync , which has configuration 
parameters to allow the number of records it should keep.

If this fails we do a index replication where possibly we can throttle 
replication 

I think it's worth explaining to users what these configuration parameters are 
and how does a node actually recover. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-7538) Uploading large text file to a field causes "this IndexWriter is closed" error

2017-10-12 Thread nkeet (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202681#comment-16202681
 ] 

nkeet commented on LUCENE-7538:
---

Thanks for having this fixed. I have been using SOLR 6.1.0 for just about a 
year now, and i recently stated seeing this issue. On ivestigation, i found 
that, at times i have files which are 1GB huge. Now i do unsterstand that this 
fix will prevent the IndexWriter from closing. but how do you index such a huge 
file, are we saying that there is a limit to the file size that can be 
indexed?, if yes what is the limit?

> Uploading large text file to a field causes "this IndexWriter is closed" error
> --
>
> Key: LUCENE-7538
> URL: https://issues.apache.org/jira/browse/LUCENE-7538
> Project: Lucene - Core
>  Issue Type: Bug
>Affects Versions: 5.5.1
>Reporter: Steve Chen
> Fix For: 6.4, 7.0
>
> Attachments: LUCENE-7538.patch, LUCENE-7538.patch
>
>
> We have seen "this IndexWriter is closed" error after we tried to upload a 
> large text file to a single Solr text field. The field definition in the 
> schema.xml is:
> {noformat}
>  termVectors="true" termPositions="true" termOffsets="true"/>
> {noformat}
> After that, the IndexWriter remained closed and couldn't be recovered until 
> we reloaded the Solr core.  The text file had size of 800MB, containing only 
> numbers and English characters.
> Stack trace is shown below:
> {noformat}
> 2016-11-02 23:00:17,913 [http-nio-19082-exec-3] ERROR 
> org.apache.solr.handler.RequestHandlerBase - 
> org.apache.solr.common.SolrException: Exception writing document id 1487_0_1 
> to the index; possible analysis error.
> at 
> org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:180)
> at 
> org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:68)
> at 
> org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:48)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalAdd(DistributedUpdateProcessor.java:934)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:1089)
> at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:712)
> at 
> org.apache.solr.update.processor.LogUpdateProcessorFactory$LogUpdateProcessor.processAdd(LogUpdateProcessorFactory.java:103)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.doAdd(ExtractingDocumentLoader.java:126)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.addDoc(ExtractingDocumentLoader.java:131)
> at 
> org.apache.solr.handler.extraction.ExtractingDocumentLoader.load(ExtractingDocumentLoader.java:237)
> at 
> org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:69)
> at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:155)
> at org.apache.solr.core.SolrCore.execute(SolrCore.java:2082)
> at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:651)
> at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:458)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:229)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:184)
> at 
> veeva.ecm.common.interfaces.web.SolrDispatchOverride.doFilter(SolrDispatchOverride.java:43)
> at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
> at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
> at 
> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:212)
> at 
> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:106)
> at 
> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:141)
> at 
> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79)
> at 
> org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:616)
> at 
> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88)
> at 
> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:521)
> at 
> org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1096)
> at 
> org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:674)
> at 
> org.apache.tomcat.util.net.NioEnd

[JENKINS] Lucene-Solr-master-Linux (32bit/jdk1.8.0_144) - Build # 20665 - Unstable!

2017-10-12 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/20665/
Java: 32bit/jdk1.8.0_144 -server -XX:+UseConcMarkSweepGC

1 tests failed.
FAILED:  org.apache.solr.cloud.OverseerRolesTest.testOverseerRole

Error Message:
Timed out waiting for overseer state change

Stack Trace:
java.lang.AssertionError: Timed out waiting for overseer state change
at 
__randomizedtesting.SeedInfo.seed([99A71852907370ED:786CE5C6ABC0463C]:0)
at org.junit.Assert.fail(Assert.java:93)
at 
org.apache.solr.cloud.OverseerRolesTest.waitForNewOverseer(OverseerRolesTest.java:62)
at 
org.apache.solr.cloud.OverseerRolesTest.testOverseerRole(OverseerRolesTest.java:140)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)




Build Log:
[...truncated 13222 lines...]
   [junit4] Suite: org.apache.solr.cloud.OverseerRolesTest
   [junit4]   2> Creating dataDir: 
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.Overse

[jira] [Commented] (SOLR-11469) LeaderElectionContextKeyTest has flawed logic: 50% of the time it checks the wrong shard's elections

2017-10-12 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11469?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202597#comment-16202597
 ] 

ASF subversion and git services commented on SOLR-11469:


Commit 75498a6132ecd6c9afc01b8f12722eebf9b764df in lucene-solr's branch 
refs/heads/branch_7x from Chris Hostetter
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=75498a6 ]

SOLR-11469: disable LeaderElectionContextKeyTest since we know it's logically 
flawed

(cherry picked from commit cd1a635898e2d37b723ae648a270242f9fc80323)


> LeaderElectionContextKeyTest has flawed logic: 50% of the time it checks the 
> wrong shard's elections
> 
>
> Key: SOLR-11469
> URL: https://issues.apache.org/jira/browse/SOLR-11469
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Hoss Man
> Attachments: SOLR-11469.patch, SOLR-11469_incomplete_and_broken.patch
>
>
> LeaderElectionContextKeyTest is very flaky -- and on millers beastit reports 
> it shows a suspiciously close to "50%" failure rate.
> Digging into the test i realized that it creates a 2 shard index, then picks 
> "a leader" to kill (arbitrarily) and then asserts that the leader election 
> nodes for *shard1* are affected ... so ~50% of the time it kills the shard2 
> leader and then fails because it doesn't see an election in shard1.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11469) LeaderElectionContextKeyTest has flawed logic: 50% of the time it checks the wrong shard's elections

2017-10-12 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11469?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202598#comment-16202598
 ] 

ASF subversion and git services commented on SOLR-11469:


Commit cd1a635898e2d37b723ae648a270242f9fc80323 in lucene-solr's branch 
refs/heads/master from Chris Hostetter
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=cd1a635 ]

SOLR-11469: disable LeaderElectionContextKeyTest since we know it's logically 
flawed


> LeaderElectionContextKeyTest has flawed logic: 50% of the time it checks the 
> wrong shard's elections
> 
>
> Key: SOLR-11469
> URL: https://issues.apache.org/jira/browse/SOLR-11469
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Hoss Man
> Attachments: SOLR-11469.patch, SOLR-11469_incomplete_and_broken.patch
>
>
> LeaderElectionContextKeyTest is very flaky -- and on millers beastit reports 
> it shows a suspiciously close to "50%" failure rate.
> Digging into the test i realized that it creates a 2 shard index, then picks 
> "a leader" to kill (arbitrarily) and then asserts that the leader election 
> nodes for *shard1* are affected ... so ~50% of the time it kills the shard2 
> leader and then fails because it doesn't see an election in shard1.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11450) ComplexPhraseQParserPlugin not running charfilter for some multiterm queries in 6.x

2017-10-12 Thread Tim Allison (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11450?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202585#comment-16202585
 ] 

Tim Allison commented on SOLR-11450:


To get the directionality right...sorry.  The issue I opened is a duplicate of 
LUCENE-7687...my error.

> ComplexPhraseQParserPlugin not running charfilter for some multiterm queries 
> in 6.x 
> 
>
> Key: SOLR-11450
> URL: https://issues.apache.org/jira/browse/SOLR-11450
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Affects Versions: 6.6.1
>Reporter: Tim Allison
>Priority: Minor
> Attachments: SOLR-11450-unit-test.patch, SOLR-11450.patch
>
>
> On the user list, [~bjarkebm] reported that the charfilter is not being 
> applied in PrefixQueries in the ComplexPhraseQParserPlugin in 6.x.  Bjarke 
> fixed my proposed unit tests to prove this failure. All appears to work in 
> 7.x and trunk. If there are plans to release a 6.6.2, let's fold this in.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11450) ComplexPhraseQParserPlugin not running charfilter for some multiterm queries in 6.x

2017-10-12 Thread Tim Allison (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11450?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202573#comment-16202573
 ] 

Tim Allison commented on SOLR-11450:


[~mikemccand] [~jpountz], any chance you'd be willing to review and push this 
into 6.6.2?

> ComplexPhraseQParserPlugin not running charfilter for some multiterm queries 
> in 6.x 
> 
>
> Key: SOLR-11450
> URL: https://issues.apache.org/jira/browse/SOLR-11450
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Affects Versions: 6.6.1
>Reporter: Tim Allison
>Priority: Minor
> Attachments: SOLR-11450-unit-test.patch, SOLR-11450.patch
>
>
> On the user list, [~bjarkebm] reported that the charfilter is not being 
> applied in PrefixQueries in the ComplexPhraseQParserPlugin in 6.x.  Bjarke 
> fixed my proposed unit tests to prove this failure. All appears to work in 
> 7.x and trunk. If there are plans to release a 6.6.2, let's fold this in.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-11480) Remove lingering admin-extra.html files and references

2017-10-12 Thread Eric Pugh (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Pugh updated SOLR-11480:
-
Attachment: SOLR-11480.patch

Here is a patch version, however there is some new lines etc that showed up 
that aren't in the GitHub PR version... argh.

> Remove lingering admin-extra.html files and references
> --
>
> Key: SOLR-11480
> URL: https://issues.apache.org/jira/browse/SOLR-11480
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Admin UI, examples
>Affects Versions: 6.6
>Reporter: Eric Pugh
>Priority: Minor
> Attachments: SOLR-11480.patch
>
>
> There are still some admin-extra related files, most confusingly, in the 
> techproducts configset!   I started up solr with -e techproducts, and saw the 
> files listed, but couldn't use them.   
> While sad to see it go, I think it's better to have it be completely removed, 
> versus still lingering about.
> Related to https://issues.apache.org/jira/browse/SOLR-8140.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-7687) ComplexPhraseQueryParser with AsciiFoldingFilterFactory (SOLR)

2017-10-12 Thread Tim Allison (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7687?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202561#comment-16202561
 ] 

Tim Allison commented on LUCENE-7687:
-

There's a patch available on SOLR-11450.  This seems to have been fixed in 7.x

> ComplexPhraseQueryParser with AsciiFoldingFilterFactory (SOLR)
> --
>
> Key: LUCENE-7687
> URL: https://issues.apache.org/jira/browse/LUCENE-7687
> Project: Lucene - Core
>  Issue Type: Bug
>Affects Versions: 6.4.1
> Environment: solr-6.4.1 (yes, solr, but I don't know where the bug 
> exactly is)
>Reporter: Jochen Barth
>
> I modified generic *_txt-Field type to use AsciiFoldingFilterFactory on query 
> & index.
> When quering with
> \{!complexphrase}text_txt:"König*" -- there are 0 results
> \{!complexphrase}text_txt:"Konig*" -- there are >0 results
> \{!complexphrase}text_txt:"König" -- there are >0 results (but less than the 
> line above)
> and without \{!complexphrase} everything works o.k.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-11480) Remove lingering admin-extra.html files and references

2017-10-12 Thread Eric Pugh (JIRA)
Eric Pugh created SOLR-11480:


 Summary: Remove lingering admin-extra.html files and references
 Key: SOLR-11480
 URL: https://issues.apache.org/jira/browse/SOLR-11480
 Project: Solr
  Issue Type: Improvement
  Security Level: Public (Default Security Level. Issues are Public)
  Components: Admin UI, examples
Affects Versions: 6.6
Reporter: Eric Pugh
Priority: Minor


There are still some admin-extra related files, most confusingly, in the 
techproducts configset!   I started up solr with -e techproducts, and saw the 
files listed, but couldn't use them.   

While sad to see it go, I think it's better to have it be completely removed, 
versus still lingering about.

Related to https://issues.apache.org/jira/browse/SOLR-8140.




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[GitHub] lucene-solr pull request #261: Cleanup admin extra

2017-10-12 Thread epugh
GitHub user epugh opened a pull request:

https://github.com/apache/lucene-solr/pull/261

Cleanup admin extra



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/epugh/lucene-solr cleanup_admin_extra

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/lucene-solr/pull/261.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #261


commit 2fd6e25986b81c914f3972f5de9211fe4e272447
Author: epugh 
Date:   2017-10-12T19:50:50Z

remove from list of gettable files the no longer used admin-extra.html

commit 1cfc7525348663057aeb86fce7fc7fe10b63889e
Author: epugh 
Date:   2017-10-12T19:50:50Z

Revert "remove from list of gettable files the no longer used 
admin-extra.html"

This reverts commit 2fd6e25986b81c914f3972f5de9211fe4e272447.

commit e8bd7dc966f30e6fffb8adeb471c01b6cddc04b7
Author: epugh 
Date:   2017-10-12T20:00:30Z

remove deprecated admin-extra.html from list of gettable files

commit 6f78fc67e8c008efb587e2995e798c2cdbcd4792
Author: epugh 
Date:   2017-10-12T20:02:50Z

removed now deprecated admin files from example configurations.

commit fabf60501465c79571112b18ed0672b4c9966c05
Author: epugh 
Date:   2017-10-12T20:04:20Z

remove no longer used admin files from -e techproducts demo




---

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-7.x-Linux (32bit/jdk1.8.0_144) - Build # 603 - Unstable!

2017-10-12 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/603/
Java: 32bit/jdk1.8.0_144 -server -XX:+UseG1GC

1 tests failed.
FAILED:  
org.apache.solr.handler.TestReplicationHandler.doTestIndexAndConfigReplication

Error Message:
Index: 0, Size: 0

Stack Trace:
java.lang.IndexOutOfBoundsException: Index: 0, Size: 0
at 
__randomizedtesting.SeedInfo.seed([65E8B336A3F734D0:71A0E86380F089CE]:0)
at java.util.ArrayList.rangeCheck(ArrayList.java:653)
at java.util.ArrayList.get(ArrayList.java:429)
at 
org.apache.solr.handler.TestReplicationHandler.doTestIndexAndConfigReplication(TestReplicationHandler.java:561)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)




Build Log:
[...truncated 12240 lines...]
   [junit4] Suite: org.apache.solr.handler.TestReplicationHandler
   [junit4]   2> 730155 INFO  
(SUITE-TestReplicationHandler-seed#[65E8B336A3F734D0]-worker) [] 
o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: 
test.solr.allow

[jira] [Commented] (SOLR-11479) Collections API: ADDREPLICA fails if you try to specify property.coreNodeName=foo

2017-10-12 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11479?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202400#comment-16202400
 ] 

Hoss Man commented on SOLR-11479:
-

discovered while trying to fix the tests in SOLR-11469

> Collections API: ADDREPLICA fails if you try to specify 
> property.coreNodeName=foo
> -
>
> Key: SOLR-11479
> URL: https://issues.apache.org/jira/browse/SOLR-11479
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Hoss Man
> Attachments: SOLR-11479.patch
>
>
> if you attempt to specify a {{property.coreNodeName}} param when using the 
> Collection API's {{action=ADDREPLICA}} the request will fail -- the 
> underlying cause of the failure (in the logs for the node where the 
> underlying Core Admin CREATECORE request is routed) seems like a perplexing 
> chicken/egg problem...
> (the various names in the error below are from a test patch i'll attach 
> shortly)
> {noformat}
>[junit4]   2> Caused by: org.apache.solr.common.SolrException: Unable to 
> create core [solrj_replica_explicit_coreNodeName_shard1_replica_n5]
>[junit4]   2>  at 
> org.apache.solr.core.CoreContainer.createFromDescriptor(CoreContainer.java:1049)
>[junit4]   2>  at 
> org.apache.solr.core.CoreContainer.create(CoreContainer.java:947)
>[junit4]   2>  ... 34 more
>[junit4]   2> Caused by: org.apache.solr.common.SolrException: 
> coreNodeName user_assigned_node_name does not exist in shard shard1: 
> DocCollection(solrj_replica_explicit_coreNodeName//collections/solrj_replica_explicit_coreNodeName/state.json/7)={
> ...
>[junit4]   2> 11157 INFO  (qtp1977346252-47) [n:127.0.0.1:51786_solr 
> c:solrj_replica_explicit_coreNodeName s:shard1 r:user_assigned_node_name 
> x:solrj_replica_explicit_coreNodeName_shard1_replica_n5] o.a.s.s.HttpSolrCall 
> [admin] webapp=null path=/admin/cores 
> params={qt=/admin/cores&coreNodeName=core_node6&collection.configName=conf&name=solrj_replica_explicit_coreNodeName_shard1_replica_n5&property.coreNodeName=user_assigned_node_name&action=CREATE&collection=solrj_replica_explicit_coreNodeName&shard=shard1&wt=javabin&version=2&replicaType=NRT}
>  status=400 QTime=3009
> {noformat}
> ...the CREATE core action is failing because the cloud shard we want to 
> ADDREPLICA to says that a replica with that coreNodeName doesn't exist



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-11469) LeaderElectionContextKeyTest has flawed logic: 50% of the time it checks the wrong shard's elections

2017-10-12 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-11469:

Attachment: SOLR-11469_incomplete_and_broken.patch

bq. I'm not really sure how to make this test work reliably? ... unless maybe 
we manually add replicas with explicitly specified coreNodeName and force them 
to be the leader

FWIW, the attached {{SOLR-11469_incomplete_and_broken.patch}} attempts this, 
but i ran into SOLR-11479 which currently makes this impossible.

> LeaderElectionContextKeyTest has flawed logic: 50% of the time it checks the 
> wrong shard's elections
> 
>
> Key: SOLR-11469
> URL: https://issues.apache.org/jira/browse/SOLR-11469
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Hoss Man
> Attachments: SOLR-11469.patch, SOLR-11469_incomplete_and_broken.patch
>
>
> LeaderElectionContextKeyTest is very flaky -- and on millers beastit reports 
> it shows a suspiciously close to "50%" failure rate.
> Digging into the test i realized that it creates a 2 shard index, then picks 
> "a leader" to kill (arbitrarily) and then asserts that the leader election 
> nodes for *shard1* are affected ... so ~50% of the time it kills the shard2 
> leader and then fails because it doesn't see an election in shard1.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-11479) Collections API: ADDREPLICA fails if you try to specify property.coreNodeName=foo

2017-10-12 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11479?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-11479:

Attachment: SOLR-11479.patch

here's a (test only) patch demonstrating the problem ... really straight 
forward addition to CollectionsAPISolrJTest...

{code}
  @Test
  public void testAddReplicaWithCoreNodeNameProp() throws Exception {

final String collectionName = "solrj_replica_explicit_coreNodeName";
CollectionAdminRequest.createCollection(collectionName, "conf", 1, 2)
  .process(cluster.getSolrClient());

final String coreNodeName = "user_assigned_node_name";

CollectionAdminResponse response = 
CollectionAdminRequest.addReplicaToShard(collectionName, "shard1")
  .withProperty(CoreDescriptor.CORE_NODE_NAME, coreNodeName)
  .process(cluster.getSolrClient());

assertEquals(0, response.getStatus());
assertTrue(response.isSuccess());

Replica newReplica = grabNewReplica(response, 
getCollectionState(collectionName));
assertTrue(newReplica.getName().equals(coreNodeName));
  }

{code}

> Collections API: ADDREPLICA fails if you try to specify 
> property.coreNodeName=foo
> -
>
> Key: SOLR-11479
> URL: https://issues.apache.org/jira/browse/SOLR-11479
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Hoss Man
> Attachments: SOLR-11479.patch
>
>
> if you attempt to specify a {{property.coreNodeName}} param when using the 
> Collection API's {{action=ADDREPLICA}} the request will fail -- the 
> underlying cause of the failure (in the logs for the node where the 
> underlying Core Admin CREATECORE request is routed) seems like a perplexing 
> chicken/egg problem...
> (the various names in the error below are from a test patch i'll attach 
> shortly)
> {noformat}
>[junit4]   2> Caused by: org.apache.solr.common.SolrException: Unable to 
> create core [solrj_replica_explicit_coreNodeName_shard1_replica_n5]
>[junit4]   2>  at 
> org.apache.solr.core.CoreContainer.createFromDescriptor(CoreContainer.java:1049)
>[junit4]   2>  at 
> org.apache.solr.core.CoreContainer.create(CoreContainer.java:947)
>[junit4]   2>  ... 34 more
>[junit4]   2> Caused by: org.apache.solr.common.SolrException: 
> coreNodeName user_assigned_node_name does not exist in shard shard1: 
> DocCollection(solrj_replica_explicit_coreNodeName//collections/solrj_replica_explicit_coreNodeName/state.json/7)={
> ...
>[junit4]   2> 11157 INFO  (qtp1977346252-47) [n:127.0.0.1:51786_solr 
> c:solrj_replica_explicit_coreNodeName s:shard1 r:user_assigned_node_name 
> x:solrj_replica_explicit_coreNodeName_shard1_replica_n5] o.a.s.s.HttpSolrCall 
> [admin] webapp=null path=/admin/cores 
> params={qt=/admin/cores&coreNodeName=core_node6&collection.configName=conf&name=solrj_replica_explicit_coreNodeName_shard1_replica_n5&property.coreNodeName=user_assigned_node_name&action=CREATE&collection=solrj_replica_explicit_coreNodeName&shard=shard1&wt=javabin&version=2&replicaType=NRT}
>  status=400 QTime=3009
> {noformat}
> ...the CREATE core action is failing because the cloud shard we want to 
> ADDREPLICA to says that a replica with that coreNodeName doesn't exist



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-11479) Collections API: ADDREPLICA fails if you try to specify property.coreNodeName=foo

2017-10-12 Thread Hoss Man (JIRA)
Hoss Man created SOLR-11479:
---

 Summary: Collections API: ADDREPLICA fails if you try to specify 
property.coreNodeName=foo
 Key: SOLR-11479
 URL: https://issues.apache.org/jira/browse/SOLR-11479
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
Reporter: Hoss Man


if you attempt to specify a {{property.coreNodeName}} param when using the 
Collection API's {{action=ADDREPLICA}} the request will fail -- the underlying 
cause of the failure (in the logs for the node where the underlying Core Admin 
CREATECORE request is routed) seems like a perplexing chicken/egg problem...

(the various names in the error below are from a test patch i'll attach shortly)

{noformat}
   [junit4]   2> Caused by: org.apache.solr.common.SolrException: Unable to 
create core [solrj_replica_explicit_coreNodeName_shard1_replica_n5]
   [junit4]   2>at 
org.apache.solr.core.CoreContainer.createFromDescriptor(CoreContainer.java:1049)
   [junit4]   2>at 
org.apache.solr.core.CoreContainer.create(CoreContainer.java:947)
   [junit4]   2>... 34 more
   [junit4]   2> Caused by: org.apache.solr.common.SolrException: coreNodeName 
user_assigned_node_name does not exist in shard shard1: 
DocCollection(solrj_replica_explicit_coreNodeName//collections/solrj_replica_explicit_coreNodeName/state.json/7)={
...
   [junit4]   2> 11157 INFO  (qtp1977346252-47) [n:127.0.0.1:51786_solr 
c:solrj_replica_explicit_coreNodeName s:shard1 r:user_assigned_node_name 
x:solrj_replica_explicit_coreNodeName_shard1_replica_n5] o.a.s.s.HttpSolrCall 
[admin] webapp=null path=/admin/cores 
params={qt=/admin/cores&coreNodeName=core_node6&collection.configName=conf&name=solrj_replica_explicit_coreNodeName_shard1_replica_n5&property.coreNodeName=user_assigned_node_name&action=CREATE&collection=solrj_replica_explicit_coreNodeName&shard=shard1&wt=javabin&version=2&replicaType=NRT}
 status=400 QTime=3009

{noformat}


...the CREATE core action is failing because the cloud shard we want to 
ADDREPLICA to says that a replica with that coreNodeName doesn't exist



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-11478) Solr should remove it's entry from live_nodes in zk immediately on shutdown and add after solr has loaded it's cores and is ready to serve requests.

2017-10-12 Thread Binoy Dalal (JIRA)
Binoy Dalal created SOLR-11478:
--

 Summary: Solr should remove it's entry from live_nodes in zk 
immediately on shutdown and add after solr has loaded it's cores and is ready 
to serve requests.
 Key: SOLR-11478
 URL: https://issues.apache.org/jira/browse/SOLR-11478
 Project: Solr
  Issue Type: Improvement
  Security Level: Public (Default Security Level. Issues are Public)
  Components: SolrCloud
Reporter: Binoy Dalal
Priority: Minor


Solr currently, upon receiving the stop command, removes its entry from the zk 
/live_nodes znode after it has finished processing all inflight requests, just 
before finally shutting down.
In this case, any applications that depend on a solr node's live_node entry to 
decide whether or not to query it fail once the stop command has been executed 
but solr has not yet fully shut down.
Something similar occurs during startup of a solr node. The solr node seems to 
add it's entry to the /live_nodes in zk once it is up but before it has started 
accepting requests and once again, this causes dependent applications to fail 
in a similar fashion.
Hence, removal of the node entry and addition of the same to the zk live_nodes 
immediately upon shutting down and at the very end upon coming up respectively 
will greatly benefit applications that depend the live_nodes znode.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-master-MacOSX (64bit/jdk-9) - Build # 4222 - Unstable!

2017-10-12 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-MacOSX/4222/
Java: 64bit/jdk-9 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC 
--illegal-access=deny

1 tests failed.
FAILED:  junit.framework.TestSuite.org.apache.solr.util.TestSolrCLIRunExample

Error Message:
ObjectTracker found 5 object(s) that were not released!!! [TransactionLog, 
MDCAwareThreadPoolExecutor, NRTCachingDirectory, NRTCachingDirectory, 
NRTCachingDirectory] 
org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: 
org.apache.solr.update.TransactionLog  at 
org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42)
  at org.apache.solr.update.TransactionLog.(TransactionLog.java:190)  at 
org.apache.solr.update.UpdateLog.newTransactionLog(UpdateLog.java:448)  at 
org.apache.solr.update.UpdateLog.ensureLog(UpdateLog.java:1261)  at 
org.apache.solr.update.UpdateLog.add(UpdateLog.java:534)  at 
org.apache.solr.update.UpdateLog.add(UpdateLog.java:519)  at 
org.apache.solr.update.DirectUpdateHandler2.doNormalUpdate(DirectUpdateHandler2.java:352)
  at 
org.apache.solr.update.DirectUpdateHandler2.addDoc0(DirectUpdateHandler2.java:271)
  at 
org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:221)
  at 
org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:67)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalAdd(DistributedUpdateProcessor.java:991)
  at 
org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:1207)
  at 
org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:753)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.AddSchemaFieldsUpdateProcessorFactory$AddSchemaFieldsUpdateProcessor.processAdd(AddSchemaFieldsUpdateProcessorFactory.java:474)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldMutatingUpdateProcessor.processAdd(FieldMutatingUpdateProcessor.java:118)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldMutatingUpdateProcessor.processAdd(FieldMutatingUpdateProcessor.java:118)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldMutatingUpdateProcessor.processAdd(FieldMutatingUpdateProcessor.java:118)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldMutatingUpdateProcessor.processAdd(FieldMutatingUpdateProcessor.java:118)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldNameMutatingUpdateProcessorFactory$1.processAdd(FieldNameMutatingUpdateProcessorFactory.java:74)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldMutatingUpdateProcessor.processAdd(FieldMutatingUpdateProcessor.java:118)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.AbstractDefaultValueUpdateProcessorFactory$DefaultValueUpdateProcessor.processAdd(AbstractDefaultValueUpdateProcessorFactory.java:91)
  at 
org.apache.solr.handler.loader.JavabinLoader$1.update(JavabinLoader.java:98)  
at 
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readOuterMostDocIterator(JavaBinUpdateRequestCodec.java:188)
  at 
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readIterator(JavaBinUpdateRequestCodec.java:144)
  at org.apache.solr.common.util.JavaBinCodec.readObject(JavaBinCodec.java:311) 
 at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:256)  at 
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readNamedList(JavaBinUpdateRequestCodec.java:130)
  at org.apache.solr.common.util.JavaBinCodec.readObject(JavaBinCodec.java:276) 
 at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:256)  at 
org.apache.solr.common.util.JavaBinCodec.unmarshal(JavaBinCodec.java:178)  at 
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec.unmarshal(JavaBinUpdateRequestCodec.java:195)
  at 
org.apache.solr.handler.loader.JavabinLoader.parseAndLoadDocs(JavabinLoader.java:108)
  at org.apache.solr.handler.loader.JavabinLoader.load(JavabinLoader.java:55)  
at 
org.apache.solr.handler.UpdateRequestHandler$1.load(UpdateRequestHandler.java:97)
  at 
org.apache.solr.handler.ContentStreamHandlerBase.handleRequ

[jira] [Commented] (SOLR-11452) TestTlogReplica.testOnlyLeaderIndexes() failure

2017-10-12 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11452?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202253#comment-16202253
 ] 

Hoss Man commented on SOLR-11452:
-

Dat: we're still seeing jenkins failures (even on master) from these asserts...

https://builds.apache.org/job/Lucene-Solr-Tests-master/2119/
{noformat}
expected:<1> but was:<0>

Stack Trace:
java.lang.AssertionError: expected:<1> but was:<0>
at 
__randomizedtesting.SeedInfo.seed([1F0F4531096B9C4:1DF189DE6533C757]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.failNotEquals(Assert.java:647)
at org.junit.Assert.assertEquals(Assert.java:128)
at org.junit.Assert.assertEquals(Assert.java:472)
at org.junit.Assert.assertEquals(Assert.java:456)
at 
org.apache.solr.cloud.TestTlogReplica.assertCopyOverOldUpdates(TestTlogReplica.java:916)
at 
org.apache.solr.cloud.TestTlogReplica.testOnlyLeaderIndexes(TestTlogReplica.java:490)

{noformat}

> TestTlogReplica.testOnlyLeaderIndexes() failure
> ---
>
> Key: SOLR-11452
> URL: https://issues.apache.org/jira/browse/SOLR-11452
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Steve Rowe
>Assignee: Cao Manh Dat
>
> Reproduces for me, from 
> [https://builds.apache.org/job/Lucene-Solr-NightlyTests-master/1398]:
> {noformat}
> Checking out Revision f0a4b2dafe13e2b372e33ce13d552f169187a44e 
> (refs/remotes/origin/master)
> [...]
>[junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=TestTlogReplica 
> -Dtests.method=testOnlyLeaderIndexes -Dtests.seed=CCAC87827208491B 
> -Dtests.multiplier=2 -Dtests.nightly=true -Dtests.slow=true 
> -Dtests.linedocsfile=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-master/test-data/enwiki.random.lines.txt
>  -Dtests.locale=el -Dtests.timezone=Australia/LHI -Dtests.asserts=true 
> -Dtests.file.encoding=ISO-8859-1
>[junit4] FAILURE 29.5s J2 | TestTlogReplica.testOnlyLeaderIndexes <<<
>[junit4]> Throwable #1: java.lang.AssertionError: expected:<2> but 
> was:<5>
>[junit4]>  at 
> __randomizedtesting.SeedInfo.seed([CCAC87827208491B:D0ADFA0F07AD3788]:0)
>[junit4]>  at 
> org.apache.solr.cloud.TestTlogReplica.assertCopyOverOldUpdates(TestTlogReplica.java:909)
>[junit4]>  at 
> org.apache.solr.cloud.TestTlogReplica.testOnlyLeaderIndexes(TestTlogReplica.java:501)
>[junit4]>  at java.lang.Thread.run(Thread.java:748)
> [...]
>[junit4]   2> NOTE: test params are: codec=CheapBastard, 
> sim=RandomSimilarity(queryNorm=false): {}, locale=el, timezone=Australia/LHI
>[junit4]   2> NOTE: Linux 3.13.0-88-generic amd64/Oracle Corporation 
> 1.8.0_144 (64-bit)/cpus=4,threads=1,free=137513712,total=520093696
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-7.x-Windows (32bit/jdk1.8.0_144) - Build # 246 - Still Unstable!

2017-10-12 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Windows/246/
Java: 32bit/jdk1.8.0_144 -client -XX:+UseSerialGC

2 tests failed.
FAILED:  org.apache.solr.cloud.LeaderElectionContextKeyTest.test

Error Message:


Stack Trace:
java.lang.AssertionError
at 
__randomizedtesting.SeedInfo.seed([3684889586F2CF57:BED0B74F280EA2AF]:0)
at org.junit.Assert.fail(Assert.java:92)
at org.junit.Assert.assertTrue(Assert.java:43)
at org.junit.Assert.assertTrue(Assert.java:54)
at 
org.apache.solr.cloud.LeaderElectionContextKeyTest.test(LeaderElectionContextKeyTest.java:88)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)


FAILED:  
org.apache.solr.cloud.autoscaling.TriggerIntegrationTest.testTriggerThrottling

Error Message:
Both triggers should have fired by now

Stack Trace:
java.lang.AssertionError: Both triggers should have fired by now
at 
__randomizedtesting.SeedInfo.seed([3684889586F2CF57:CDA620B054582CC5]:0)

[jira] [Updated] (LUCENE-4100) Maxscore - Efficient Scoring

2017-10-12 Thread Adrien Grand (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4100?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Adrien Grand updated LUCENE-4100:
-
Attachment: LUCENE-4100.patch

Here is a patch:
 - more docs and tests
 - replaces needsScores with a SearchMode enum as suggested by Robert
 - the MAXSCORE optimization work with top-level disjunctions and filtered 
disjunctions (FILTER or MUST_NOT)
 - TopScoreDocsCollector sets the totalHitCount to -1 when the optimization is 
used since the total hit count is unknown
 - MaxScoreScorer was changed to reason on integers rather than doubles to 
avoid floating-point arithmetic issues. To do that it scales all max scores 
into 0..2^16, rounding up when working on the max scores of sub clauses, and 
down when rounding the min competitive score in order to make sure to not miss 
matches (at the cost of potentially more false positives, but this is fine)

The patch is alreay huge (due to the needsScore/searchMode change mostly) so I 
wanted to do the strict minimum here for this feature to be useful, but we'll 
need follow-ups to make the optimization work with the paging collector, 
conjunctions that have more than one scoring clause, TopFieldCollector when the 
first sort field is the score, integrate it with IndexSearcher (currently you 
need to create the collector manually to use it), etc.

> Maxscore - Efficient Scoring
> 
>
> Key: LUCENE-4100
> URL: https://issues.apache.org/jira/browse/LUCENE-4100
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/codecs, core/query/scoring, core/search
>Affects Versions: 4.0-ALPHA
>Reporter: Stefan Pohl
>  Labels: api-change, gsoc2014, patch, performance
> Fix For: 4.9, 6.0
>
> Attachments: LUCENE-4100.patch, LUCENE-4100.patch, 
> contrib_maxscore.tgz, maxscore.patch
>
>
> At Berlin Buzzwords 2012, I will be presenting 'maxscore', an efficient 
> algorithm first published in the IR domain in 1995 by H. Turtle & J. Flood, 
> that I find deserves more attention among Lucene users (and developers).
> I implemented a proof of concept and did some performance measurements with 
> example queries and lucenebench, the package of Mike McCandless, resulting in 
> very significant speedups.
> This ticket is to get started the discussion on including the implementation 
> into Lucene's codebase. Because the technique requires awareness about it 
> from the Lucene user/developer, it seems best to become a contrib/module 
> package so that it consciously can be chosen to be used.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (LUCENE-7992) Kuromoji fails with UnsupportedOperationException in case of duplicate keys in the user dictionary

2017-10-12 Thread Christian Moen (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-7992?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Christian Moen reassigned LUCENE-7992:
--

Assignee: Christian Moen

> Kuromoji fails with UnsupportedOperationException in case of duplicate keys 
> in the user dictionary
> --
>
> Key: LUCENE-7992
> URL: https://issues.apache.org/jira/browse/LUCENE-7992
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Adrien Grand
>Assignee: Christian Moen
>Priority: Minor
>
> Failing is the right thing to do but the exception could clarify the source 
> of the problem. Today it just throws an UnsupportedOperationException with no 
> error message because of a call to PositiveIntOutputs.merge.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-7992) Kuromoji fails with UnsupportedOperationException in case of duplicate keys in the user dictionary

2017-10-12 Thread Christian Moen (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7992?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16202098#comment-16202098
 ] 

Christian Moen commented on LUCENE-7992:


Thanks, Adrien.  I'll have a look. 

> Kuromoji fails with UnsupportedOperationException in case of duplicate keys 
> in the user dictionary
> --
>
> Key: LUCENE-7992
> URL: https://issues.apache.org/jira/browse/LUCENE-7992
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Adrien Grand
>Priority: Minor
>
> Failing is the right thing to do but the exception could clarify the source 
> of the problem. Today it just throws an UnsupportedOperationException with no 
> error message because of a call to PositiveIntOutputs.merge.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-7992) Kuromoji fails with UnsupportedOperationException in case of duplicate keys in the user dictionary

2017-10-12 Thread Adrien Grand (JIRA)
Adrien Grand created LUCENE-7992:


 Summary: Kuromoji fails with UnsupportedOperationException in case 
of duplicate keys in the user dictionary
 Key: LUCENE-7992
 URL: https://issues.apache.org/jira/browse/LUCENE-7992
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Adrien Grand
Priority: Minor


Failing is the right thing to do but the exception could clarify the source of 
the problem. Today it just throws an UnsupportedOperationException with no 
error message because of a call to PositiveIntOutputs.merge.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-master-Linux (64bit/jdk-9) - Build # 20663 - Still Unstable!

2017-10-12 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/20663/
Java: 64bit/jdk-9 -XX:-UseCompressedOops -XX:+UseParallelGC 
--illegal-access=deny

1 tests failed.
FAILED:  org.apache.solr.cloud.LeaderElectionContextKeyTest.test

Error Message:


Stack Trace:
java.lang.AssertionError
at 
__randomizedtesting.SeedInfo.seed([E5E30D1D8F479F6C:6DB732C721BBF294]:0)
at org.junit.Assert.fail(Assert.java:92)
at org.junit.Assert.assertTrue(Assert.java:43)
at org.junit.Assert.assertTrue(Assert.java:54)
at 
org.apache.solr.cloud.LeaderElectionContextKeyTest.test(LeaderElectionContextKeyTest.java:88)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:564)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.base/java.lang.Thread.run(Thread.java:844)




Build Log:
[...truncated 13266 lines...]
   [junit4] Suite: org.apache.solr.cloud.LeaderElectionContextKeyTest
   [junit4]   2> Creating dataDir: 
/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.LeaderElectionContextKeyTest_E5E30D1D8F479F6C-001/init-core-data

[JENKINS] Lucene-Solr-Tests-master - Build # 2119 - Still Unstable

2017-10-12 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-master/2119/

5 tests failed.
FAILED:  org.apache.solr.cloud.TestTlogReplica.testOnlyLeaderIndexes

Error Message:
expected:<1> but was:<0>

Stack Trace:
java.lang.AssertionError: expected:<1> but was:<0>
at 
__randomizedtesting.SeedInfo.seed([1F0F4531096B9C4:1DF189DE6533C757]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.failNotEquals(Assert.java:647)
at org.junit.Assert.assertEquals(Assert.java:128)
at org.junit.Assert.assertEquals(Assert.java:472)
at org.junit.Assert.assertEquals(Assert.java:456)
at 
org.apache.solr.cloud.TestTlogReplica.assertCopyOverOldUpdates(TestTlogReplica.java:916)
at 
org.apache.solr.cloud.TestTlogReplica.testOnlyLeaderIndexes(TestTlogReplica.java:490)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)


FAILED:  org.apache.solr.core.snapshots.TestSolrCloudSnapshots.testSnapshots

Error Message:
Coul

[JENKINS] Lucene-Solr-master-Windows (64bit/jdk1.8.0_144) - Build # 6956 - Still Unstable!

2017-10-12 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Windows/6956/
Java: 64bit/jdk1.8.0_144 -XX:-UseCompressedOops -XX:+UseG1GC

1 tests failed.
FAILED:  junit.framework.TestSuite.org.apache.solr.core.TestSolrIndexConfig

Error Message:
Could not remove the following files (in the order of attempts):
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-core\test\J1\temp\solr.core.TestSolrIndexConfig_8C6FF671EFE9C986-001\init-core-data-001:
 java.nio.file.AccessDeniedException: 
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-core\test\J1\temp\solr.core.TestSolrIndexConfig_8C6FF671EFE9C986-001\init-core-data-001

C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-core\test\J1\temp\solr.core.TestSolrIndexConfig_8C6FF671EFE9C986-001:
 java.nio.file.DirectoryNotEmptyException: 
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-core\test\J1\temp\solr.core.TestSolrIndexConfig_8C6FF671EFE9C986-001
 

Stack Trace:
java.io.IOException: Could not remove the following files (in the order of 
attempts):
   
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-core\test\J1\temp\solr.core.TestSolrIndexConfig_8C6FF671EFE9C986-001\init-core-data-001:
 java.nio.file.AccessDeniedException: 
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-core\test\J1\temp\solr.core.TestSolrIndexConfig_8C6FF671EFE9C986-001\init-core-data-001
   
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-core\test\J1\temp\solr.core.TestSolrIndexConfig_8C6FF671EFE9C986-001:
 java.nio.file.DirectoryNotEmptyException: 
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-core\test\J1\temp\solr.core.TestSolrIndexConfig_8C6FF671EFE9C986-001

at __randomizedtesting.SeedInfo.seed([8C6FF671EFE9C986]:0)
at org.apache.lucene.util.IOUtils.rm(IOUtils.java:329)
at 
org.apache.lucene.util.TestRuleTemporaryFilesCleanup.afterAlways(TestRuleTemporaryFilesCleanup.java:216)
at 
com.carrotsearch.randomizedtesting.rules.TestRuleAdapter$1.afterAlways(TestRuleAdapter.java:31)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:43)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.lang.Thread.run(Thread.java:748)




Build Log:
[...truncated 11946 lines...]
   [junit4] Suite: org.apache.solr.core.TestSolrIndexConfig
   [junit4]   2> Creating dataDir: 
C:\Users\jenkins\workspace\Lucene-Solr-master-Windows\solr\build\solr-core\test\J1\temp\solr.core.TestSolrIndexConfig_8C6FF671EFE9C986-001\init-core-data-001
   [junit4]   2> 533159 WARN  
(SUITE-TestSolrIndexConfig-seed#[8C6FF671EFE9C986]-worker) [] 
o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=10 numCloses=10
   [junit4]   2> 533159 INFO  
(SUITE-TestSolrIndexConfig-seed#[8C6FF671EFE9C986]-worker) [] 
o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) 
w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 533163 INFO  
(SUITE-TestSolrIndexConfig-seed#[8C6FF671EFE9C986]-worker) [] 
o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via: 
@org.apache.solr.util.RandomizeSSL(reason=, value=NaN, ssl=NaN, clientAuth=NaN)
   [junit4]   2> 533163 INFO  
(SUITE-TestSolrIndexConfig-seed#[8C6FF671EFE9C986]-worker) [] 
o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: 
test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> 533164 INFO  
(SUITE-TestSolrIndexConfig-seed#[8C6FF671EFE9C986]-worker) [] 
o.a.s.SolrTestCaseJ4 initCore
   [junit4]   2> 533164 INFO  
(SUITE-TestSolrIndexConfig-seed#[8C6FF671EFE9C986]-worker) [] 
o.a.s.c.SolrResourceLoader [null] Added 2 libs to classloader, from paths: 
[/C:/Users/jenkins/workspace/Lucene-Solr-master-Windows/solr/core/src/test-files/solr/collection1/lib,
 
/C:/Users/jenkins/workspace/Lucene-Solr-master-Windows/solr/core/src/test-files/solr/collection1/lib/classes]
   [junit4]   2> 533191 INFO  
(SUITE-TestSolrIndexConfig-seed#[8C6FF671EFE9C986]-worker) [] 
o.a.s.u.SolrIndexConfig IndexWriter infoStream solr logging is enabled
   [junit4]   2> 533191 INFO  
(SUITE-TestSolrIndexConfig-seed#[8C6FF671EFE9C986]-worker) [] 
o.a.s.c.SolrConfig Using

Making "routing" strategies for writing segments explicit ?

2017-10-12 Thread Tommaso Teofili
Hi all,

having been involved in such kind of challenge and having seen a few more
similar enquiries on the dev list, I was wondering if it may be time to
think about making it possible to have an explicit (customizable and
therefore pluggable) policy which allows people to chime into where
documents and / or segments get written (on write or on merge).
Recently there was someone asking about possibly having segments sorted by
a field using SortingMergePolicy, but as Uwe noted it's currently an
implementation detail. Personally I have tried (and failed because it was
too costly) to make sure docs belonging to certain clusters (identified by
a field) being written within same segments (for data locality / memory
footprint concerns when "loading" docs from a certain cluster).

As of today that'd be *really* hard, but I just wanted to share my feeling
that such topic might be something to keep an eye on.

My 2 cents,
Tommaso


[jira] [Resolved] (SOLR-11451) ComputePlanActionTest.testNodeLost() failure

2017-10-12 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul resolved SOLR-11451.
---
   Resolution: Fixed
Fix Version/s: 7.2

> ComputePlanActionTest.testNodeLost() failure
> 
>
> Key: SOLR-11451
> URL: https://issues.apache.org/jira/browse/SOLR-11451
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Steve Rowe
>Assignee: Noble Paul
> Fix For: 7.2
>
>
> Reproduces 100% for me (on Linux) if I remove {{-Dtests.method=testNodeLost}} 
> from the repro line, from 
> [https://jenkins.thetaphi.de/job/Lucene-Solr-master-Windows/6946]:
> {noformat}
> Checking out Revision e92bde1e7ece020d581638f4c59c3840bf75d183 
> (refs/remotes/origin/master)
> [...]
>[junit4]   2> NOTE: reproduce with: ant test  
> -Dtestcase=ComputePlanActionTest -Dtests.method=testNodeLost 
> -Dtests.seed=B1DFD0131B7C99E4 -Dtests.slow=true -Dtests.locale=ja 
> -Dtests.timezone=Asia/Katmandu -Dtests.asserts=true 
> -Dtests.file.encoding=UTF-8
>[junit4] FAILURE 8.09s J1 | ComputePlanActionTest.testNodeLost <<<
>[junit4]> Throwable #1: java.lang.AssertionError: The operations 
> computed by ComputePlanAction should not be null
>[junit4]>  at 
> __randomizedtesting.SeedInfo.seed([B1DFD0131B7C99E4:ECA1EED9896FC62]:0)
>[junit4]>  at 
> org.apache.solr.cloud.autoscaling.ComputePlanActionTest.testNodeLost(ComputePlanActionTest.java:193)
>[junit4]>  at java.lang.Thread.run(Thread.java:748)
> [...]
>[junit4]   2> NOTE: test params are: codec=Asserting(Lucene70): {}, 
> docValues:{}, maxPointsInLeafNode=491, maxMBSortInHeap=6.600294616233929, 
> sim=RandomSimilarity(queryNorm=true): {}, locale=ja, timezone=Asia/Katmandu
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11451) ComputePlanActionTest.testNodeLost() failure

2017-10-12 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16201968#comment-16201968
 ] 

ASF subversion and git services commented on SOLR-11451:


Commit d3ef1bff41e781ad077ff0dbf314c4ed8e33f80f in lucene-solr's branch 
refs/heads/branch_7_1 from [~noble.paul]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=d3ef1bf ]

SOLR-11451: ComputePlanActionTest.testNodeLost() failure


> ComputePlanActionTest.testNodeLost() failure
> 
>
> Key: SOLR-11451
> URL: https://issues.apache.org/jira/browse/SOLR-11451
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Steve Rowe
>Assignee: Noble Paul
> Fix For: 7.2
>
>
> Reproduces 100% for me (on Linux) if I remove {{-Dtests.method=testNodeLost}} 
> from the repro line, from 
> [https://jenkins.thetaphi.de/job/Lucene-Solr-master-Windows/6946]:
> {noformat}
> Checking out Revision e92bde1e7ece020d581638f4c59c3840bf75d183 
> (refs/remotes/origin/master)
> [...]
>[junit4]   2> NOTE: reproduce with: ant test  
> -Dtestcase=ComputePlanActionTest -Dtests.method=testNodeLost 
> -Dtests.seed=B1DFD0131B7C99E4 -Dtests.slow=true -Dtests.locale=ja 
> -Dtests.timezone=Asia/Katmandu -Dtests.asserts=true 
> -Dtests.file.encoding=UTF-8
>[junit4] FAILURE 8.09s J1 | ComputePlanActionTest.testNodeLost <<<
>[junit4]> Throwable #1: java.lang.AssertionError: The operations 
> computed by ComputePlanAction should not be null
>[junit4]>  at 
> __randomizedtesting.SeedInfo.seed([B1DFD0131B7C99E4:ECA1EED9896FC62]:0)
>[junit4]>  at 
> org.apache.solr.cloud.autoscaling.ComputePlanActionTest.testNodeLost(ComputePlanActionTest.java:193)
>[junit4]>  at java.lang.Thread.run(Thread.java:748)
> [...]
>[junit4]   2> NOTE: test params are: codec=Asserting(Lucene70): {}, 
> docValues:{}, maxPointsInLeafNode=491, maxMBSortInHeap=6.600294616233929, 
> sim=RandomSimilarity(queryNorm=true): {}, locale=ja, timezone=Asia/Katmandu
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11451) ComputePlanActionTest.testNodeLost() failure

2017-10-12 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16201967#comment-16201967
 ] 

ASF subversion and git services commented on SOLR-11451:


Commit 2345b087bcd0840ae491fabc01a9cbf669e0a29b in lucene-solr's branch 
refs/heads/branch_7_1 from [~noble.paul]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=2345b08 ]

SOLR-11451: added logging to track the failures


> ComputePlanActionTest.testNodeLost() failure
> 
>
> Key: SOLR-11451
> URL: https://issues.apache.org/jira/browse/SOLR-11451
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Steve Rowe
>Assignee: Noble Paul
> Fix For: 7.2
>
>
> Reproduces 100% for me (on Linux) if I remove {{-Dtests.method=testNodeLost}} 
> from the repro line, from 
> [https://jenkins.thetaphi.de/job/Lucene-Solr-master-Windows/6946]:
> {noformat}
> Checking out Revision e92bde1e7ece020d581638f4c59c3840bf75d183 
> (refs/remotes/origin/master)
> [...]
>[junit4]   2> NOTE: reproduce with: ant test  
> -Dtestcase=ComputePlanActionTest -Dtests.method=testNodeLost 
> -Dtests.seed=B1DFD0131B7C99E4 -Dtests.slow=true -Dtests.locale=ja 
> -Dtests.timezone=Asia/Katmandu -Dtests.asserts=true 
> -Dtests.file.encoding=UTF-8
>[junit4] FAILURE 8.09s J1 | ComputePlanActionTest.testNodeLost <<<
>[junit4]> Throwable #1: java.lang.AssertionError: The operations 
> computed by ComputePlanAction should not be null
>[junit4]>  at 
> __randomizedtesting.SeedInfo.seed([B1DFD0131B7C99E4:ECA1EED9896FC62]:0)
>[junit4]>  at 
> org.apache.solr.cloud.autoscaling.ComputePlanActionTest.testNodeLost(ComputePlanActionTest.java:193)
>[junit4]>  at java.lang.Thread.run(Thread.java:748)
> [...]
>[junit4]   2> NOTE: test params are: codec=Asserting(Lucene70): {}, 
> docValues:{}, maxPointsInLeafNode=491, maxMBSortInHeap=6.600294616233929, 
> sim=RandomSimilarity(queryNorm=true): {}, locale=ja, timezone=Asia/Katmandu
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11451) ComputePlanActionTest.testNodeLost() failure

2017-10-12 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16201958#comment-16201958
 ] 

ASF subversion and git services commented on SOLR-11451:


Commit 3f3e49d5bc9ab8603386d8544640661bb0a5b2ab in lucene-solr's branch 
refs/heads/branch_7x from [~noble.paul]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=3f3e49d ]

SOLR-11451: ComputePlanActionTest.testNodeLost() failure


> ComputePlanActionTest.testNodeLost() failure
> 
>
> Key: SOLR-11451
> URL: https://issues.apache.org/jira/browse/SOLR-11451
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Steve Rowe
>Assignee: Noble Paul
>
> Reproduces 100% for me (on Linux) if I remove {{-Dtests.method=testNodeLost}} 
> from the repro line, from 
> [https://jenkins.thetaphi.de/job/Lucene-Solr-master-Windows/6946]:
> {noformat}
> Checking out Revision e92bde1e7ece020d581638f4c59c3840bf75d183 
> (refs/remotes/origin/master)
> [...]
>[junit4]   2> NOTE: reproduce with: ant test  
> -Dtestcase=ComputePlanActionTest -Dtests.method=testNodeLost 
> -Dtests.seed=B1DFD0131B7C99E4 -Dtests.slow=true -Dtests.locale=ja 
> -Dtests.timezone=Asia/Katmandu -Dtests.asserts=true 
> -Dtests.file.encoding=UTF-8
>[junit4] FAILURE 8.09s J1 | ComputePlanActionTest.testNodeLost <<<
>[junit4]> Throwable #1: java.lang.AssertionError: The operations 
> computed by ComputePlanAction should not be null
>[junit4]>  at 
> __randomizedtesting.SeedInfo.seed([B1DFD0131B7C99E4:ECA1EED9896FC62]:0)
>[junit4]>  at 
> org.apache.solr.cloud.autoscaling.ComputePlanActionTest.testNodeLost(ComputePlanActionTest.java:193)
>[junit4]>  at java.lang.Thread.run(Thread.java:748)
> [...]
>[junit4]   2> NOTE: test params are: codec=Asserting(Lucene70): {}, 
> docValues:{}, maxPointsInLeafNode=491, maxMBSortInHeap=6.600294616233929, 
> sim=RandomSimilarity(queryNorm=true): {}, locale=ja, timezone=Asia/Katmandu
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11451) ComputePlanActionTest.testNodeLost() failure

2017-10-12 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16201957#comment-16201957
 ] 

ASF subversion and git services commented on SOLR-11451:


Commit c969634b6bd9fc8d73d95ff1a1e9169a26efa38c in lucene-solr's branch 
refs/heads/branch_7x from [~noble.paul]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=c969634 ]

SOLR-11451: added logging to track the failures


> ComputePlanActionTest.testNodeLost() failure
> 
>
> Key: SOLR-11451
> URL: https://issues.apache.org/jira/browse/SOLR-11451
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Steve Rowe
>Assignee: Noble Paul
>
> Reproduces 100% for me (on Linux) if I remove {{-Dtests.method=testNodeLost}} 
> from the repro line, from 
> [https://jenkins.thetaphi.de/job/Lucene-Solr-master-Windows/6946]:
> {noformat}
> Checking out Revision e92bde1e7ece020d581638f4c59c3840bf75d183 
> (refs/remotes/origin/master)
> [...]
>[junit4]   2> NOTE: reproduce with: ant test  
> -Dtestcase=ComputePlanActionTest -Dtests.method=testNodeLost 
> -Dtests.seed=B1DFD0131B7C99E4 -Dtests.slow=true -Dtests.locale=ja 
> -Dtests.timezone=Asia/Katmandu -Dtests.asserts=true 
> -Dtests.file.encoding=UTF-8
>[junit4] FAILURE 8.09s J1 | ComputePlanActionTest.testNodeLost <<<
>[junit4]> Throwable #1: java.lang.AssertionError: The operations 
> computed by ComputePlanAction should not be null
>[junit4]>  at 
> __randomizedtesting.SeedInfo.seed([B1DFD0131B7C99E4:ECA1EED9896FC62]:0)
>[junit4]>  at 
> org.apache.solr.cloud.autoscaling.ComputePlanActionTest.testNodeLost(ComputePlanActionTest.java:193)
>[junit4]>  at java.lang.Thread.run(Thread.java:748)
> [...]
>[junit4]   2> NOTE: test params are: codec=Asserting(Lucene70): {}, 
> docValues:{}, maxPointsInLeafNode=491, maxMBSortInHeap=6.600294616233929, 
> sim=RandomSimilarity(queryNorm=true): {}, locale=ja, timezone=Asia/Katmandu
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [VOTE] Release Lucene/Solr 7.1.0 RC1

2017-10-12 Thread Tommaso Teofili
+1 SUCCESS! [3:35:06.139114]

Il giorno gio 12 ott 2017 alle ore 13:20 Shalin Shekhar Mangar <
sha...@apache.org> ha scritto:

> Please vote for release candidate 1 for Lucene/Solr 7.1.0
>
> The artifacts can be downloaded from:
>
> https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-7.1.0-RC1-reva2c54447f118a5dc70ab0e0ae14bd87b3545254b
>
> You can run the smoke tester directly with this command:
>
> python3 -u dev-tools/scripts/smokeTestRelease.py \
>
> https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-7.1.0-RC1-reva2c54447f118a5dc70ab0e0ae14bd87b3545254b
>
> Smoke tester passed for me (but on the 2nd attempt due to a flaky test
> that's already being tracked in a Jira).
> SUCCESS! [0:55:14.107386]
>
> Here's my +1 to release.
>
>
> --
> Regards,
> Shalin Shekhar Mangar.
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
>
>


[jira] [Commented] (SOLR-11451) ComputePlanActionTest.testNodeLost() failure

2017-10-12 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16201949#comment-16201949
 ] 

ASF subversion and git services commented on SOLR-11451:


Commit 4f52d2db109557c5e6ddf2b6f0bdf7090707f41b in lucene-solr's branch 
refs/heads/master from [~noble.paul]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=4f52d2d ]

SOLR-11451: ComputePlanActionTest.testNodeLost() failure


> ComputePlanActionTest.testNodeLost() failure
> 
>
> Key: SOLR-11451
> URL: https://issues.apache.org/jira/browse/SOLR-11451
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Steve Rowe
>Assignee: Noble Paul
>
> Reproduces 100% for me (on Linux) if I remove {{-Dtests.method=testNodeLost}} 
> from the repro line, from 
> [https://jenkins.thetaphi.de/job/Lucene-Solr-master-Windows/6946]:
> {noformat}
> Checking out Revision e92bde1e7ece020d581638f4c59c3840bf75d183 
> (refs/remotes/origin/master)
> [...]
>[junit4]   2> NOTE: reproduce with: ant test  
> -Dtestcase=ComputePlanActionTest -Dtests.method=testNodeLost 
> -Dtests.seed=B1DFD0131B7C99E4 -Dtests.slow=true -Dtests.locale=ja 
> -Dtests.timezone=Asia/Katmandu -Dtests.asserts=true 
> -Dtests.file.encoding=UTF-8
>[junit4] FAILURE 8.09s J1 | ComputePlanActionTest.testNodeLost <<<
>[junit4]> Throwable #1: java.lang.AssertionError: The operations 
> computed by ComputePlanAction should not be null
>[junit4]>  at 
> __randomizedtesting.SeedInfo.seed([B1DFD0131B7C99E4:ECA1EED9896FC62]:0)
>[junit4]>  at 
> org.apache.solr.cloud.autoscaling.ComputePlanActionTest.testNodeLost(ComputePlanActionTest.java:193)
>[junit4]>  at java.lang.Thread.run(Thread.java:748)
> [...]
>[junit4]   2> NOTE: test params are: codec=Asserting(Lucene70): {}, 
> docValues:{}, maxPointsInLeafNode=491, maxMBSortInHeap=6.600294616233929, 
> sim=RandomSimilarity(queryNorm=true): {}, locale=ja, timezone=Asia/Katmandu
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-7.x-Linux (32bit/jdk1.8.0_144) - Build # 601 - Unstable!

2017-10-12 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-Linux/601/
Java: 32bit/jdk1.8.0_144 -server -XX:+UseSerialGC

3 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.TestSolrCloudWithSecureImpersonation

Error Message:
4 threads leaked from SUITE scope at 
org.apache.solr.cloud.TestSolrCloudWithSecureImpersonation: 1) 
Thread[id=13291, name=jetty-launcher-1531-thread-1-EventThread, state=WAITING, 
group=TGRP-TestSolrCloudWithSecureImpersonation] at 
sun.misc.Unsafe.park(Native Method) at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
 at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) 
at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:501)
2) Thread[id=13288, 
name=jetty-launcher-1531-thread-2-SendThread(127.0.0.1:42127), 
state=TIMED_WAITING, group=TGRP-TestSolrCloudWithSecureImpersonation] 
at java.lang.Thread.sleep(Native Method) at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:101)
 at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:997)
 at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1060)
3) Thread[id=13289, 
name=jetty-launcher-1531-thread-1-SendThread(127.0.0.1:42127), 
state=TIMED_WAITING, group=TGRP-TestSolrCloudWithSecureImpersonation] 
at java.lang.Thread.sleep(Native Method) at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:101)
 at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:997)
 at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1060)
4) Thread[id=13290, name=jetty-launcher-1531-thread-2-EventThread, 
state=WAITING, group=TGRP-TestSolrCloudWithSecureImpersonation] at 
sun.misc.Unsafe.park(Native Method) at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
 at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) 
at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:501)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: 4 threads leaked from SUITE 
scope at org.apache.solr.cloud.TestSolrCloudWithSecureImpersonation: 
   1) Thread[id=13291, name=jetty-launcher-1531-thread-1-EventThread, 
state=WAITING, group=TGRP-TestSolrCloudWithSecureImpersonation]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:501)
   2) Thread[id=13288, 
name=jetty-launcher-1531-thread-2-SendThread(127.0.0.1:42127), 
state=TIMED_WAITING, group=TGRP-TestSolrCloudWithSecureImpersonation]
at java.lang.Thread.sleep(Native Method)
at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:101)
at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:997)
at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1060)
   3) Thread[id=13289, 
name=jetty-launcher-1531-thread-1-SendThread(127.0.0.1:42127), 
state=TIMED_WAITING, group=TGRP-TestSolrCloudWithSecureImpersonation]
at java.lang.Thread.sleep(Native Method)
at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:101)
at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:997)
at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1060)
   4) Thread[id=13290, name=jetty-launcher-1531-thread-2-EventThread, 
state=WAITING, group=TGRP-TestSolrCloudWithSecureImpersonation]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:501)
at __randomizedtesting.SeedInfo.seed([44A114FA7C211A7C]:0)


FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.TestSolrCloudWithSecureImpersonation

Error Message:
There are still zombie threads that couldn't be terminated:1) 
Thread[id=13288, name=jetty-launcher-1531-thread-2-SendThread(127.0.0.1:42127), 
state=TIMED_WAITING, group=TGRP-TestSolrCloudWithSecureIm

[jira] [Commented] (SOLR-11444) Improve Aliases.java and comma delimited collection list handling

2017-10-12 Thread Kevin Risden (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11444?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16201866#comment-16201866
 ] 

Kevin Risden commented on SOLR-11444:
-

Changes for SQL sounds reasonable to me. I think streaming expressions support 
aliases now. When it was originally implemented not sure that was the case.

> Improve Aliases.java and comma delimited collection list handling
> -
>
> Key: SOLR-11444
> URL: https://issues.apache.org/jira/browse/SOLR-11444
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: SolrCloud
>Reporter: David Smiley
>Assignee: David Smiley
> Attachments: SOLR_11444_Aliases.patch, SOLR_11444_Aliases.patch
>
>
> While starting to look at SOLR-11299 I noticed some brittleness in 
> assumptions about Strings that refer to a collection.  Sometimes they are in 
> fact references to comma separated lists, which appears was added with the 
> introduction of collection aliases (an alias can refer to a comma delimited 
> list).  So Java's type system kind of goes out the window when we do this.  
> In one case this leads to a bug -- CloudSolrClient will throw an NPE if you 
> try to write to such an alias.  Sending an update via HTTP will allow it and 
> send it to the first in the list.
> So this issue is about refactoring and some little improvements pertaining to 
> Aliases.java plus certain key spots that deal with collection references.  I 
> don't think I want to go as far as changing the public SolrJ API except to 
> adding documentation on what's possible.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-11476) Solr seems to ignore replica placement rules

2017-10-12 Thread Franz Wimmer (JIRA)
Franz Wimmer created SOLR-11476:
---

 Summary: Solr seems to ignore replica placement rules
 Key: SOLR-11476
 URL: https://issues.apache.org/jira/browse/SOLR-11476
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
  Components: v2 API
Affects Versions: 7.0.1, 7.0, 6.6.1
Reporter: Franz Wimmer


I'm running Solr Cloud in OpenShift. I have set up 5 pods across the cluster, 
each running one solr node. The instances are managed by a ZooKeeper ensemble.

In Solr 6.1, when I created a collection with 5 shards and a replicationFactor 
of 2, it would distribute all 10 replicas evenly on the nodes. With Solr 6.6 
and 7.0, the distribution is random (why would I even want to gather many 
shards on one node and leave another node empty?).

I tried using "Rule-based Replica Placement", but my rules seem to be ignored. 
I tried:


{code}
/solr/admin/collections?action=CREATE&name=wikipedia&numShards=5&maxShardsPerNode=2&replicationFactor=2&replica:<3,node:*&collection.configName=wikipedia
/solr/admin/collections?action=CREATE&name=wikipedia&numShards=5&maxShardsPerNode=2&replicationFactor=2&shard:*,replica:1,node:*&collection.configName=wikipedia
/solr/admin/collections?action=CREATE&name=wikipedia4&numShards=5&maxShardsPerNode=2&replicationFactor=2&replica:2,node:*&collection.configName=wikipedia
{code}

In all three cases, shards are distributed unevenly, often leaving one node 
completely empty.

What seems odd is, when i try a replicationFactor of 1, it still doesn't work, 
but the API response is suggesting otherwise:

{code}
{
  "responseHeader":{
"status":0,
"QTime":3368},
  "success":{
"solr-3.solr:8983_solr":{
  "responseHeader":{
"status":0,
"QTime":1836},
  "core":"wikipedia5_shard5_replica_n1"},
"solr-2.solr:8983_solr":{
  "responseHeader":{
"status":0,
"QTime":1852},
  "core":"wikipedia5_shard4_replica_n1"},
"solr-4.solr:8983_solr":{
  "responseHeader":{
"status":0,
"QTime":1859},
  "core":"wikipedia5_shard3_replica_n1"},
"solr-0.solr:8983_solr":{
  "responseHeader":{
"status":0,
"QTime":1867},
  "core":"wikipedia5_shard2_replica_n1"},
"solr-1.solr:8983_solr":{
  "responseHeader":{
"status":0,
"QTime":1872},
  "core":"wikipedia5_shard1_replica_n1"}}}
{code}

Other than the returned JSON, the collection looks like this: 
!https://i.imgur.com/UwqwQ5e.png!



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11473) Make HDFSDirectoryFactory support other prefixes (besides hdfs:/)

2017-10-12 Thread Mike Drob (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11473?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16201846#comment-16201846
 ] 

Mike Drob commented on SOLR-11473:
--

bq. In general, I would prefer to use the URI class throughout 
HDFSDirectoryFactory for resolving paths, but this is just my personal opinion.
 
IIRC I tried this once before and ran into something thorny. But I can't figure 
out what it was. If you run into problems, feel free to ping me about it 
[~radu0gheorghe]

> Make HDFSDirectoryFactory support other prefixes (besides hdfs:/)
> -
>
> Key: SOLR-11473
> URL: https://issues.apache.org/jira/browse/SOLR-11473
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: hdfs
>Affects Versions: 6.6.1
>Reporter: Radu Gheorghe
>Priority: Minor
>
> Not sure if it's a bug or a missing feature :) I'm trying to make Solr work 
> on Alluxio, as described by [~thelabdude] in 
> https://www.slideshare.net/thelabdude/running-solr-in-the-cloud-at-memory-speed-with-alluxio/1
> The problem I'm facing here is with autoAddReplicas. If I have 
> replicationFactor=1 and the node with that replica dies, the node taking over 
> incorrectly assigns the data directory. For example:
> before
> {code}"dataDir":"alluxio://localhost:19998/solr/test/",{code}
> after
> {code}"dataDir":"alluxio://localhost:19998/solr/test/core_node1/alluxio://localhost:19998/solr/test/",{code}
> The same happens for ulogDir. Apparently, this has to do with this bit from 
> HDFSDirectoryFactory:
> {code}  public boolean isAbsolute(String path) {
> return path.startsWith("hdfs:/");
>   }{code}
> If I add "alluxio:/" in there, the paths are correct and the index is 
> recovered.
> I see a few options here:
> * add "alluxio:/" to the list there
> * add a regular expression in the lines of \[a-z]*:/ I hope that's not too 
> expensive, I'm not sure how often this method is called
> * don't do anything and expect alluxio to work with an "hdfs:/" path? I 
> actually tried that and didn't manage to make it work
> * have a different DirectoryFactory or something else?
> What do you think?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-master-Linux (32bit/jdk1.8.0_144) - Build # 20662 - Still Unstable!

2017-10-12 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/20662/
Java: 32bit/jdk1.8.0_144 -client -XX:+UseParallelGC

4 tests failed.
FAILED:  
org.apache.solr.cloud.TestCloudSearcherWarming.testRepFactor1LeaderStartup

Error Message:
 null Live Nodes: [127.0.0.1:43827_solr] Last available state: 
DocCollection(testRepFactor1LeaderStartup//collections/testRepFactor1LeaderStartup/state.json/4)={
   "pullReplicas":"0",   "replicationFactor":"1",   "shards":{"shard1":{   
"range":"8000-7fff",   "state":"active",   
"replicas":{"core_node2":{   
"core":"testRepFactor1LeaderStartup_shard1_replica_n1",   
"base_url":"https://127.0.0.1:43827/solr";,   
"node_name":"127.0.0.1:43827_solr",   "state":"active",   
"type":"NRT",   "leader":"true",   "router":{"name":"compositeId"}, 
  "maxShardsPerNode":"1",   "autoAddReplicas":"false",   "nrtReplicas":"1",   
"tlogReplicas":"0"}

Stack Trace:
java.lang.AssertionError: 
null
Live Nodes: [127.0.0.1:43827_solr]
Last available state: 
DocCollection(testRepFactor1LeaderStartup//collections/testRepFactor1LeaderStartup/state.json/4)={
  "pullReplicas":"0",
  "replicationFactor":"1",
  "shards":{"shard1":{
  "range":"8000-7fff",
  "state":"active",
  "replicas":{"core_node2":{
  "core":"testRepFactor1LeaderStartup_shard1_replica_n1",
  "base_url":"https://127.0.0.1:43827/solr";,
  "node_name":"127.0.0.1:43827_solr",
  "state":"active",
  "type":"NRT",
  "leader":"true",
  "router":{"name":"compositeId"},
  "maxShardsPerNode":"1",
  "autoAddReplicas":"false",
  "nrtReplicas":"1",
  "tlogReplicas":"0"}
at 
__randomizedtesting.SeedInfo.seed([DC2FFDE51D1E4138:B07B03D0CFE08D6]:0)
at org.junit.Assert.fail(Assert.java:93)
at 
org.apache.solr.cloud.SolrCloudTestCase.waitForState(SolrCloudTestCase.java:269)
at 
org.apache.solr.cloud.TestCloudSearcherWarming.testRepFactor1LeaderStartup(TestCloudSearcherWarming.java:126)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carro

[jira] [Commented] (SOLR-11473) Make HDFSDirectoryFactory support other prefixes (besides hdfs:/)

2017-10-12 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11473?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16201829#comment-16201829
 ] 

Uwe Schindler commented on SOLR-11473:
--

You can propose a patch, if you like! I'd use: 
[https://docs.oracle.com/javase/8/docs/api/java/net/URI.html#isAbsolute--]

I don't think this method is called too often. It is not very expensive. In 
general, I would prefer to use the URI class throughout HDFSDirectoryFactory 
for resolving paths, but this is just my personal opinion.

> Make HDFSDirectoryFactory support other prefixes (besides hdfs:/)
> -
>
> Key: SOLR-11473
> URL: https://issues.apache.org/jira/browse/SOLR-11473
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: hdfs
>Affects Versions: 6.6.1
>Reporter: Radu Gheorghe
>Priority: Minor
>
> Not sure if it's a bug or a missing feature :) I'm trying to make Solr work 
> on Alluxio, as described by [~thelabdude] in 
> https://www.slideshare.net/thelabdude/running-solr-in-the-cloud-at-memory-speed-with-alluxio/1
> The problem I'm facing here is with autoAddReplicas. If I have 
> replicationFactor=1 and the node with that replica dies, the node taking over 
> incorrectly assigns the data directory. For example:
> before
> {code}"dataDir":"alluxio://localhost:19998/solr/test/",{code}
> after
> {code}"dataDir":"alluxio://localhost:19998/solr/test/core_node1/alluxio://localhost:19998/solr/test/",{code}
> The same happens for ulogDir. Apparently, this has to do with this bit from 
> HDFSDirectoryFactory:
> {code}  public boolean isAbsolute(String path) {
> return path.startsWith("hdfs:/");
>   }{code}
> If I add "alluxio:/" in there, the paths are correct and the index is 
> recovered.
> I see a few options here:
> * add "alluxio:/" to the list there
> * add a regular expression in the lines of \[a-z]*:/ I hope that's not too 
> expensive, I'm not sure how often this method is called
> * don't do anything and expect alluxio to work with an "hdfs:/" path? I 
> actually tried that and didn't manage to make it work
> * have a different DirectoryFactory or something else?
> What do you think?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11473) Make HDFSDirectoryFactory support other prefixes (besides hdfs:/)

2017-10-12 Thread Radu Gheorghe (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11473?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16201818#comment-16201818
 ] 

Radu Gheorghe commented on SOLR-11473:
--

Sounds good, thanks Uwe! Should I just attach a patch with that change?

> Make HDFSDirectoryFactory support other prefixes (besides hdfs:/)
> -
>
> Key: SOLR-11473
> URL: https://issues.apache.org/jira/browse/SOLR-11473
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: hdfs
>Affects Versions: 6.6.1
>Reporter: Radu Gheorghe
>Priority: Minor
>
> Not sure if it's a bug or a missing feature :) I'm trying to make Solr work 
> on Alluxio, as described by [~thelabdude] in 
> https://www.slideshare.net/thelabdude/running-solr-in-the-cloud-at-memory-speed-with-alluxio/1
> The problem I'm facing here is with autoAddReplicas. If I have 
> replicationFactor=1 and the node with that replica dies, the node taking over 
> incorrectly assigns the data directory. For example:
> before
> {code}"dataDir":"alluxio://localhost:19998/solr/test/",{code}
> after
> {code}"dataDir":"alluxio://localhost:19998/solr/test/core_node1/alluxio://localhost:19998/solr/test/",{code}
> The same happens for ulogDir. Apparently, this has to do with this bit from 
> HDFSDirectoryFactory:
> {code}  public boolean isAbsolute(String path) {
> return path.startsWith("hdfs:/");
>   }{code}
> If I add "alluxio:/" in there, the paths are correct and the index is 
> recovered.
> I see a few options here:
> * add "alluxio:/" to the list there
> * add a regular expression in the lines of \[a-z]*:/ I hope that's not too 
> expensive, I'm not sure how often this method is called
> * don't do anything and expect alluxio to work with an "hdfs:/" path? I 
> actually tried that and didn't manage to make it work
> * have a different DirectoryFactory or something else?
> What do you think?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-11473) Make HDFSDirectoryFactory support other prefixes (besides hdfs:/)

2017-10-12 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11473?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16201811#comment-16201811
 ] 

Uwe Schindler commented on SOLR-11473:
--

I'd parse this with the URI java class. After parsing it you can use methods 
like isAbsolute() and similar to check the status.

> Make HDFSDirectoryFactory support other prefixes (besides hdfs:/)
> -
>
> Key: SOLR-11473
> URL: https://issues.apache.org/jira/browse/SOLR-11473
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: hdfs
>Affects Versions: 6.6.1
>Reporter: Radu Gheorghe
>Priority: Minor
>
> Not sure if it's a bug or a missing feature :) I'm trying to make Solr work 
> on Alluxio, as described by [~thelabdude] in 
> https://www.slideshare.net/thelabdude/running-solr-in-the-cloud-at-memory-speed-with-alluxio/1
> The problem I'm facing here is with autoAddReplicas. If I have 
> replicationFactor=1 and the node with that replica dies, the node taking over 
> incorrectly assigns the data directory. For example:
> before
> {code}"dataDir":"alluxio://localhost:19998/solr/test/",{code}
> after
> {code}"dataDir":"alluxio://localhost:19998/solr/test/core_node1/alluxio://localhost:19998/solr/test/",{code}
> The same happens for ulogDir. Apparently, this has to do with this bit from 
> HDFSDirectoryFactory:
> {code}  public boolean isAbsolute(String path) {
> return path.startsWith("hdfs:/");
>   }{code}
> If I add "alluxio:/" in there, the paths are correct and the index is 
> recovered.
> I see a few options here:
> * add "alluxio:/" to the list there
> * add a regular expression in the lines of \[a-z]*:/ I hope that's not too 
> expensive, I'm not sure how often this method is called
> * don't do anything and expect alluxio to work with an "hdfs:/" path? I 
> actually tried that and didn't manage to make it work
> * have a different DirectoryFactory or something else?
> What do you think?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[VOTE] Release Lucene/Solr 7.1.0 RC1

2017-10-12 Thread Shalin Shekhar Mangar
Please vote for release candidate 1 for Lucene/Solr 7.1.0

The artifacts can be downloaded from:
https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-7.1.0-RC1-reva2c54447f118a5dc70ab0e0ae14bd87b3545254b

You can run the smoke tester directly with this command:

python3 -u dev-tools/scripts/smokeTestRelease.py \
https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-7.1.0-RC1-reva2c54447f118a5dc70ab0e0ae14bd87b3545254b

Smoke tester passed for me (but on the 2nd attempt due to a flaky test
that's already being tracked in a Jira).
SUCCESS! [0:55:14.107386]

Here's my +1 to release.


-- 
Regards,
Shalin Shekhar Mangar.

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Solr-reference-guide-master - Build # 2537 - Failure

2017-10-12 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Solr-reference-guide-master/2537/

Log: 
Started by timer
[EnvInject] - Loading node environment variables.
Building remotely on H19 (git-websites) in workspace 
/home/jenkins/jenkins-slave/workspace/Solr-reference-guide-master
 > git rev-parse --is-inside-work-tree # timeout=10
Fetching changes from the remote Git repository
 > git config remote.origin.url git://git.apache.org/lucene-solr.git # 
 > timeout=10
Cleaning workspace
 > git rev-parse --verify HEAD # timeout=10
Resetting working tree
 > git reset --hard # timeout=10
 > git clean -fdx # timeout=10
Fetching upstream changes from git://git.apache.org/lucene-solr.git
 > git --version # timeout=10
 > git fetch --tags --progress git://git.apache.org/lucene-solr.git 
 > +refs/heads/*:refs/remotes/origin/*
 > git rev-parse refs/remotes/origin/master^{commit} # timeout=10
 > git rev-parse refs/remotes/origin/origin/master^{commit} # timeout=10
Checking out Revision d8c132129ed44d8a26e7a31064e0f36b50ba8f6c 
(refs/remotes/origin/master)
Commit message: "Some details about .system trigger listener, other minor 
edits."
 > git config core.sparsecheckout # timeout=10
 > git checkout -f d8c132129ed44d8a26e7a31064e0f36b50ba8f6c
 > git rev-list bac40493174b95f6eb92aacb065e6b308b8d364c # timeout=10
No emails were triggered.
[Solr-reference-guide-master] $ /usr/bin/env bash 
/tmp/jenkins7929466548214265962.sh
+ set -e
+ RVM_PATH=/home/jenkins/.rvm
+ RUBY_VERSION=ruby-2.3.3
+ GEMSET=solr-refguide-gemset
+ curl -sSL https://get.rvm.io
+ bash -s -- --ignore-dotfiles stable
Turning on ignore dotfiles mode.
Downloading https://github.com/rvm/rvm/archive/1.29.3.tar.gz
Downloading 
https://github.com/rvm/rvm/releases/download/1.29.3/1.29.3.tar.gz.asc
gpg: Signature made Sun 10 Sep 2017 08:59:21 PM UTC using RSA key ID BF04FF17
gpg: Good signature from "Michal Papis (RVM signing) "
gpg: aka "Michal Papis "
gpg: aka "[jpeg image of size 5015]"
gpg: WARNING: This key is not certified with a trusted signature!
gpg:  There is no indication that the signature belongs to the owner.
Primary key fingerprint: 409B 6B17 96C2 7546 2A17  0311 3804 BB82 D39D C0E3
 Subkey fingerprint: 62C9 E5F4 DA30 0D94 AC36  166B E206 C29F BF04 FF17
GPG verified '/home/jenkins/.rvm/archives/rvm-1.29.3.tgz'

Upgrading the RVM installation in /home/jenkins/.rvm/
Upgrade of RVM in /home/jenkins/.rvm/ is complete.

Upgrade Notes:

  * No new notes to display.

+ set +x
Running 'source /home/jenkins/.rvm/scripts/rvm'
Running 'rvm autolibs disable'
Running 'rvm install ruby-2.3.3'
Already installed ruby-2.3.3.
To reinstall use:

rvm reinstall ruby-2.3.3

Running 'rvm gemset create solr-refguide-gemset'
ruby-2.3.3 - #gemset created 
/home/jenkins/.rvm/gems/ruby-2.3.3@solr-refguide-gemset
ruby-2.3.3 - #generating solr-refguide-gemset wrappers
Running 'rvm ruby-2.3.3@solr-refguide-gemset'
Using /home/jenkins/.rvm/gems/ruby-2.3.3 with gemset solr-refguide-gemset
Running 'gem install --force --version 3.5.0 jekyll'
Successfully installed jekyll-3.5.0
Parsing documentation for jekyll-3.5.0
Done installing documentation for jekyll after 2 seconds
1 gem installed
Running 'gem install --force --version 2.1.0 jekyll-asciidoc'
Successfully installed jekyll-asciidoc-2.1.0
Parsing documentation for jekyll-asciidoc-2.1.0
Done installing documentation for jekyll-asciidoc after 0 seconds
1 gem installed
Running 'gem install --force --version 1.1.2 pygments.rb'
Successfully installed pygments.rb-1.1.2
Parsing documentation for pygments.rb-1.1.2
Done installing documentation for pygments.rb after 0 seconds
1 gem installed
+ ant clean build-site build-pdf
Buildfile: 
/home/jenkins/jenkins-slave/workspace/Solr-reference-guide-master/solr/solr-ref-guide/build.xml

clean:

build-init:
[mkdir] Created dir: 
/home/jenkins/jenkins-slave/workspace/Solr-reference-guide-master/solr/build/solr-ref-guide/content
 [echo] Copying all non template files from src ...
 [copy] Copying 370 files to 
/home/jenkins/jenkins-slave/workspace/Solr-reference-guide-master/solr/build/solr-ref-guide/content
 [echo] Copy (w/prop replacement) any template files from src...
 [copy] Copying 1 file to 
/home/jenkins/jenkins-slave/workspace/Solr-reference-guide-master/solr/build/solr-ref-guide/content

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-configure:
[ivy:configure] :: Apache Ivy 2.3.0 - 20130110142753 :: 
http://ant.apache.org/ivy/ ::
[ivy:configure] :: loading settings :: file = 
/home/jenkins/jenkins-slave/workspace/Solr-reference-guide-master/lucene/top-level-ivy-settings.xml

resolve:
[ivy:retrieve] 
[ivy:retrieve] :: problems summary ::
[ivy:retrieve]  ERRORS
[ivy:retrieve]  unknown resolver null
[ivy:retrieve]  unknown resolver null
[ivy:retrieve]  unknown resolver null
[ivy:retrieve]  unknown resolver null
[ivy:retrieve]

[jira] [Updated] (SOLR-11475) Endless loop and OOM in PeerSync

2017-10-12 Thread Andrey Kudryavtsev (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrey Kudryavtsev updated SOLR-11475:
--
Description: 
After problem described in SOLR-11459, I restarted cluster and got OOM on 
start. 

[PeerSync#handleVersionsWithRanges|https://github.com/apache/lucene-solr/blob/68bda0be421ce18811e03b229781fd6152fcc04a/solr/core/src/java/org/apache/solr/update/PeerSync.java#L539]
 contains this logic: 

{code}
while (otherUpdatesIndex >= 0) {
  // we have run out of ourUpdates, pick up all the remaining versions from 
the other versions
  if (ourUpdatesIndex < 0) {
String range = otherVersions.get(otherUpdatesIndex) + "..." + 
otherVersions.get(0);
rangesToRequest.add(range);
totalRequestedVersions += otherUpdatesIndex + 1;
break;
  }

  // stop when the entries get old enough that reorders may lead us to see 
updates we don't need
  if (!completeList && Math.abs(otherVersions.get(otherUpdatesIndex)) < 
ourLowThreshold) break;

  if (ourUpdates.get(ourUpdatesIndex).longValue() == 
otherVersions.get(otherUpdatesIndex).longValue()) {
ourUpdatesIndex--;
otherUpdatesIndex--;
  } else if (Math.abs(ourUpdates.get(ourUpdatesIndex)) < 
Math.abs(otherVersions.get(otherUpdatesIndex))) {
ourUpdatesIndex--;
  } else {
long rangeStart = otherVersions.get(otherUpdatesIndex);
while ((otherUpdatesIndex < otherVersions.size())
&& (Math.abs(otherVersions.get(otherUpdatesIndex)) < 
Math.abs(ourUpdates.get(ourUpdatesIndex {
  otherUpdatesIndex--;
  totalRequestedVersions++;
}
// construct range here
rangesToRequest.add(rangeStart + "..." + 
otherVersions.get(otherUpdatesIndex + 1));
  }
}
{code}

If at some point there will be
{code} ourUpdates.get(ourUpdatesIndex) = -otherVersions.get(otherUpdatesIndex) 
{code}
loop will never end. It will same string again and again into 
{{rangesToRequest}} until process runs out of memory.




  was:
After problem described in SOLR-11459, I restarted cluster and got OOM on 
start. 

PeerSync#handleVersionsWithRanges contains this logic: 

{code}
while (otherUpdatesIndex >= 0) {
  // we have run out of ourUpdates, pick up all the remaining versions from 
the other versions
  if (ourUpdatesIndex < 0) {
String range = otherVersions.get(otherUpdatesIndex) + "..." + 
otherVersions.get(0);
rangesToRequest.add(range);
totalRequestedVersions += otherUpdatesIndex + 1;
break;
  }

  // stop when the entries get old enough that reorders may lead us to see 
updates we don't need
  if (!completeList && Math.abs(otherVersions.get(otherUpdatesIndex)) < 
ourLowThreshold) break;

  if (ourUpdates.get(ourUpdatesIndex).longValue() == 
otherVersions.get(otherUpdatesIndex).longValue()) {
ourUpdatesIndex--;
otherUpdatesIndex--;
  } else if (Math.abs(ourUpdates.get(ourUpdatesIndex)) < 
Math.abs(otherVersions.get(otherUpdatesIndex))) {
ourUpdatesIndex--;
  } else {
long rangeStart = otherVersions.get(otherUpdatesIndex);
while ((otherUpdatesIndex < otherVersions.size())
&& (Math.abs(otherVersions.get(otherUpdatesIndex)) < 
Math.abs(ourUpdates.get(ourUpdatesIndex {
  otherUpdatesIndex--;
  totalRequestedVersions++;
}
// construct range here
rangesToRequest.add(rangeStart + "..." + 
otherVersions.get(otherUpdatesIndex + 1));
  }
}
{code}

If at some point there will be
{code} ourUpdates.get(ourUpdatesIndex) = -otherVersions.get(otherUpdatesIndex) 
{code}
loop will never end. It will same string again and again into 
{{rangesToRequest}} until process runs out of memory.





> Endless loop and OOM in PeerSync
> 
>
> Key: SOLR-11475
> URL: https://issues.apache.org/jira/browse/SOLR-11475
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Andrey Kudryavtsev
>
> After problem described in SOLR-11459, I restarted cluster and got OOM on 
> start. 
> [PeerSync#handleVersionsWithRanges|https://github.com/apache/lucene-solr/blob/68bda0be421ce18811e03b229781fd6152fcc04a/solr/core/src/java/org/apache/solr/update/PeerSync.java#L539]
>  contains this logic: 
> {code}
> while (otherUpdatesIndex >= 0) {
>   // we have run out of ourUpdates, pick up all the remaining versions 
> from the other versions
>   if (ourUpdatesIndex < 0) {
> String range = otherVersions.get(otherUpdatesIndex) + "..." + 
> otherVersions.get(0);
> rangesToRequest.add(range);
> totalRequestedVersions += otherUpdatesIndex + 1;
> break;
>   }
>   // stop when the entries get old enough that reorders may l

[jira] [Updated] (SOLR-11475) Endless loop and OOM in PeerSync

2017-10-12 Thread Andrey Kudryavtsev (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrey Kudryavtsev updated SOLR-11475:
--
Description: 
After problem described in SOLR-11459, I restarted cluster and got OOM on 
start. 

[PeerSync#handleVersionsWithRanges|https://github.com/apache/lucene-solr/blob/68bda0be421ce18811e03b229781fd6152fcc04a/solr/core/src/java/org/apache/solr/update/PeerSync.java#L539]
 contains this logic: 

{code}
while (otherUpdatesIndex >= 0) {
  // we have run out of ourUpdates, pick up all the remaining versions from 
the other versions
  if (ourUpdatesIndex < 0) {
String range = otherVersions.get(otherUpdatesIndex) + "..." + 
otherVersions.get(0);
rangesToRequest.add(range);
totalRequestedVersions += otherUpdatesIndex + 1;
break;
  }

  // stop when the entries get old enough that reorders may lead us to see 
updates we don't need
  if (!completeList && Math.abs(otherVersions.get(otherUpdatesIndex)) < 
ourLowThreshold) break;

  if (ourUpdates.get(ourUpdatesIndex).longValue() == 
otherVersions.get(otherUpdatesIndex).longValue()) {
ourUpdatesIndex--;
otherUpdatesIndex--;
  } else if (Math.abs(ourUpdates.get(ourUpdatesIndex)) < 
Math.abs(otherVersions.get(otherUpdatesIndex))) {
ourUpdatesIndex--;
  } else {
long rangeStart = otherVersions.get(otherUpdatesIndex);
while ((otherUpdatesIndex < otherVersions.size())
&& (Math.abs(otherVersions.get(otherUpdatesIndex)) < 
Math.abs(ourUpdates.get(ourUpdatesIndex {
  otherUpdatesIndex--;
  totalRequestedVersions++;
}
// construct range here
rangesToRequest.add(rangeStart + "..." + 
otherVersions.get(otherUpdatesIndex + 1));
  }
}
{code}

If at some point there will be
{code} ourUpdates.get(ourUpdatesIndex) = -otherVersions.get(otherUpdatesIndex) 
{code}
loop will never end. It will add same string again and again into 
{{rangesToRequest}} until process runs out of memory.




  was:
After problem described in SOLR-11459, I restarted cluster and got OOM on 
start. 

[PeerSync#handleVersionsWithRanges|https://github.com/apache/lucene-solr/blob/68bda0be421ce18811e03b229781fd6152fcc04a/solr/core/src/java/org/apache/solr/update/PeerSync.java#L539]
 contains this logic: 

{code}
while (otherUpdatesIndex >= 0) {
  // we have run out of ourUpdates, pick up all the remaining versions from 
the other versions
  if (ourUpdatesIndex < 0) {
String range = otherVersions.get(otherUpdatesIndex) + "..." + 
otherVersions.get(0);
rangesToRequest.add(range);
totalRequestedVersions += otherUpdatesIndex + 1;
break;
  }

  // stop when the entries get old enough that reorders may lead us to see 
updates we don't need
  if (!completeList && Math.abs(otherVersions.get(otherUpdatesIndex)) < 
ourLowThreshold) break;

  if (ourUpdates.get(ourUpdatesIndex).longValue() == 
otherVersions.get(otherUpdatesIndex).longValue()) {
ourUpdatesIndex--;
otherUpdatesIndex--;
  } else if (Math.abs(ourUpdates.get(ourUpdatesIndex)) < 
Math.abs(otherVersions.get(otherUpdatesIndex))) {
ourUpdatesIndex--;
  } else {
long rangeStart = otherVersions.get(otherUpdatesIndex);
while ((otherUpdatesIndex < otherVersions.size())
&& (Math.abs(otherVersions.get(otherUpdatesIndex)) < 
Math.abs(ourUpdates.get(ourUpdatesIndex {
  otherUpdatesIndex--;
  totalRequestedVersions++;
}
// construct range here
rangesToRequest.add(rangeStart + "..." + 
otherVersions.get(otherUpdatesIndex + 1));
  }
}
{code}

If at some point there will be
{code} ourUpdates.get(ourUpdatesIndex) = -otherVersions.get(otherUpdatesIndex) 
{code}
loop will never end. It will same string again and again into 
{{rangesToRequest}} until process runs out of memory.





> Endless loop and OOM in PeerSync
> 
>
> Key: SOLR-11475
> URL: https://issues.apache.org/jira/browse/SOLR-11475
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Andrey Kudryavtsev
>
> After problem described in SOLR-11459, I restarted cluster and got OOM on 
> start. 
> [PeerSync#handleVersionsWithRanges|https://github.com/apache/lucene-solr/blob/68bda0be421ce18811e03b229781fd6152fcc04a/solr/core/src/java/org/apache/solr/update/PeerSync.java#L539]
>  contains this logic: 
> {code}
> while (otherUpdatesIndex >= 0) {
>   // we have run out of ourUpdates, pick up all the remaining versions 
> from the other versions
>   if (ourUpdatesIndex < 0) {
> String range = otherVersions.get(otherUpdatesIndex) + "..." + 
> otherVersions.get(0);
> rangesToRequest.add(range);
>

[jira] [Updated] (SOLR-11475) Endless loop and OOM in PeerSync

2017-10-12 Thread Andrey Kudryavtsev (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrey Kudryavtsev updated SOLR-11475:
--
Description: 
After problem described in SOLR-11459, I restarted cluster and got OOM on 
start. 

PeerSync#handleVersionsWithRanges contains this logic: 

{code}
while (otherUpdatesIndex >= 0) {
  // we have run out of ourUpdates, pick up all the remaining versions from 
the other versions
  if (ourUpdatesIndex < 0) {
String range = otherVersions.get(otherUpdatesIndex) + "..." + 
otherVersions.get(0);
rangesToRequest.add(range);
totalRequestedVersions += otherUpdatesIndex + 1;
break;
  }

  // stop when the entries get old enough that reorders may lead us to see 
updates we don't need
  if (!completeList && Math.abs(otherVersions.get(otherUpdatesIndex)) < 
ourLowThreshold) break;

  if (ourUpdates.get(ourUpdatesIndex).longValue() == 
otherVersions.get(otherUpdatesIndex).longValue()) {
ourUpdatesIndex--;
otherUpdatesIndex--;
  } else if (Math.abs(ourUpdates.get(ourUpdatesIndex)) < 
Math.abs(otherVersions.get(otherUpdatesIndex))) {
ourUpdatesIndex--;
  } else {
long rangeStart = otherVersions.get(otherUpdatesIndex);
while ((otherUpdatesIndex < otherVersions.size())
&& (Math.abs(otherVersions.get(otherUpdatesIndex)) < 
Math.abs(ourUpdates.get(ourUpdatesIndex {
  otherUpdatesIndex--;
  totalRequestedVersions++;
}
// construct range here
rangesToRequest.add(rangeStart + "..." + 
otherVersions.get(otherUpdatesIndex + 1));
  }
}
{code}

If at some point there will be
{code} ourUpdates.get(ourUpdatesIndex) = -otherVersions.get(otherUpdatesIndex) 
{code}
loop will never end. It will same string again and again into 
{{rangesToRequest}} until process runs out of memory.




  was:
After problem discussed in SOLR-11459, I restarted cluster and got OOM on 
start. 

PeerSync#handleVersionsWithRanges contains this logic: 

{code}
while (otherUpdatesIndex >= 0) {
  // we have run out of ourUpdates, pick up all the remaining versions from 
the other versions
  if (ourUpdatesIndex < 0) {
String range = otherVersions.get(otherUpdatesIndex) + "..." + 
otherVersions.get(0);
rangesToRequest.add(range);
totalRequestedVersions += otherUpdatesIndex + 1;
break;
  }

  // stop when the entries get old enough that reorders may lead us to see 
updates we don't need
  if (!completeList && Math.abs(otherVersions.get(otherUpdatesIndex)) < 
ourLowThreshold) break;

  if (ourUpdates.get(ourUpdatesIndex).longValue() == 
otherVersions.get(otherUpdatesIndex).longValue()) {
ourUpdatesIndex--;
otherUpdatesIndex--;
  } else if (Math.abs(ourUpdates.get(ourUpdatesIndex)) < 
Math.abs(otherVersions.get(otherUpdatesIndex))) {
ourUpdatesIndex--;
  } else {
long rangeStart = otherVersions.get(otherUpdatesIndex);
while ((otherUpdatesIndex < otherVersions.size())
&& (Math.abs(otherVersions.get(otherUpdatesIndex)) < 
Math.abs(ourUpdates.get(ourUpdatesIndex {
  otherUpdatesIndex--;
  totalRequestedVersions++;
}
// construct range here
rangesToRequest.add(rangeStart + "..." + 
otherVersions.get(otherUpdatesIndex + 1));
  }
}
{code}

If at some point there will be
{code} ourUpdates.get(ourUpdatesIndex) = -otherVersions.get(otherUpdatesIndex) 
{code}
loop will never end. It will same string again and again into 
{{rangesToRequest}} until process runs out of memory.





> Endless loop and OOM in PeerSync
> 
>
> Key: SOLR-11475
> URL: https://issues.apache.org/jira/browse/SOLR-11475
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Andrey Kudryavtsev
>
> After problem described in SOLR-11459, I restarted cluster and got OOM on 
> start. 
> PeerSync#handleVersionsWithRanges contains this logic: 
> {code}
> while (otherUpdatesIndex >= 0) {
>   // we have run out of ourUpdates, pick up all the remaining versions 
> from the other versions
>   if (ourUpdatesIndex < 0) {
> String range = otherVersions.get(otherUpdatesIndex) + "..." + 
> otherVersions.get(0);
> rangesToRequest.add(range);
> totalRequestedVersions += otherUpdatesIndex + 1;
> break;
>   }
>   // stop when the entries get old enough that reorders may lead us to 
> see updates we don't need
>   if (!completeList && Math.abs(otherVersions.get(otherUpdatesIndex)) < 
> ourLowThreshold) break;
>   if (ourUpdates.get(ourUpdatesIndex).longValue() == 
> otherVersions.get(otherUpdatesIndex).longValue()) {
> ourUpdatesIndex--;
> o

[jira] [Created] (SOLR-11474) Endless loop and OOM in PeerSync

2017-10-12 Thread Andrey Kudryavtsev (JIRA)
Andrey Kudryavtsev created SOLR-11474:
-

 Summary: Endless loop and OOM in PeerSync
 Key: SOLR-11474
 URL: https://issues.apache.org/jira/browse/SOLR-11474
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
Reporter: Andrey Kudryavtsev


After problem discussed in SOLR-11459, I restarted cluster and got OOM on 
start. 

PeerSync#handleVersionsWithRanges contains this logic: 

{code}
while (otherUpdatesIndex >= 0) {
  // we have run out of ourUpdates, pick up all the remaining versions from 
the other versions
  if (ourUpdatesIndex < 0) {
String range = otherVersions.get(otherUpdatesIndex) + "..." + 
otherVersions.get(0);
rangesToRequest.add(range);
totalRequestedVersions += otherUpdatesIndex + 1;
break;
  }

  // stop when the entries get old enough that reorders may lead us to see 
updates we don't need
  if (!completeList && Math.abs(otherVersions.get(otherUpdatesIndex)) < 
ourLowThreshold) break;

  if (ourUpdates.get(ourUpdatesIndex).longValue() == 
otherVersions.get(otherUpdatesIndex).longValue()) {
ourUpdatesIndex--;
otherUpdatesIndex--;
  } else if (Math.abs(ourUpdates.get(ourUpdatesIndex)) < 
Math.abs(otherVersions.get(otherUpdatesIndex))) {
ourUpdatesIndex--;
  } else {
long rangeStart = otherVersions.get(otherUpdatesIndex);
while ((otherUpdatesIndex < otherVersions.size())
&& (Math.abs(otherVersions.get(otherUpdatesIndex)) < 
Math.abs(ourUpdates.get(ourUpdatesIndex {
  otherUpdatesIndex--;
  totalRequestedVersions++;
}
// construct range here
rangesToRequest.add(rangeStart + "..." + 
otherVersions.get(otherUpdatesIndex + 1));
  }
}
{code}

If at some point there will be
{code} ourUpdates.get(ourUpdatesIndex) = -otherVersions.get(otherUpdatesIndex) 
{code}
loop will never end. It will same string again and again into 
{{rangesToRequest}} until process runs out of memory.






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-11475) Endless loop and OOM in PeerSync

2017-10-12 Thread Andrey Kudryavtsev (JIRA)
Andrey Kudryavtsev created SOLR-11475:
-

 Summary: Endless loop and OOM in PeerSync
 Key: SOLR-11475
 URL: https://issues.apache.org/jira/browse/SOLR-11475
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
Reporter: Andrey Kudryavtsev


After problem discussed in SOLR-11459, I restarted cluster and got OOM on 
start. 

PeerSync#handleVersionsWithRanges contains this logic: 

{code}
while (otherUpdatesIndex >= 0) {
  // we have run out of ourUpdates, pick up all the remaining versions from 
the other versions
  if (ourUpdatesIndex < 0) {
String range = otherVersions.get(otherUpdatesIndex) + "..." + 
otherVersions.get(0);
rangesToRequest.add(range);
totalRequestedVersions += otherUpdatesIndex + 1;
break;
  }

  // stop when the entries get old enough that reorders may lead us to see 
updates we don't need
  if (!completeList && Math.abs(otherVersions.get(otherUpdatesIndex)) < 
ourLowThreshold) break;

  if (ourUpdates.get(ourUpdatesIndex).longValue() == 
otherVersions.get(otherUpdatesIndex).longValue()) {
ourUpdatesIndex--;
otherUpdatesIndex--;
  } else if (Math.abs(ourUpdates.get(ourUpdatesIndex)) < 
Math.abs(otherVersions.get(otherUpdatesIndex))) {
ourUpdatesIndex--;
  } else {
long rangeStart = otherVersions.get(otherUpdatesIndex);
while ((otherUpdatesIndex < otherVersions.size())
&& (Math.abs(otherVersions.get(otherUpdatesIndex)) < 
Math.abs(ourUpdates.get(ourUpdatesIndex {
  otherUpdatesIndex--;
  totalRequestedVersions++;
}
// construct range here
rangesToRequest.add(rangeStart + "..." + 
otherVersions.get(otherUpdatesIndex + 1));
  }
}
{code}

If at some point there will be
{code} ourUpdates.get(ourUpdatesIndex) = -otherVersions.get(otherUpdatesIndex) 
{code}
loop will never end. It will same string again and again into 
{{rangesToRequest}} until process runs out of memory.






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-11473) Make HDFSDirectoryFactory support other prefixes (besides hdfs:/)

2017-10-12 Thread Radu Gheorghe (JIRA)
Radu Gheorghe created SOLR-11473:


 Summary: Make HDFSDirectoryFactory support other prefixes (besides 
hdfs:/)
 Key: SOLR-11473
 URL: https://issues.apache.org/jira/browse/SOLR-11473
 Project: Solr
  Issue Type: Improvement
  Security Level: Public (Default Security Level. Issues are Public)
  Components: hdfs
Affects Versions: 6.6.1
Reporter: Radu Gheorghe
Priority: Minor


Not sure if it's a bug or a missing feature :) I'm trying to make Solr work on 
Alluxio, as described by [~thelabdude] in 
https://www.slideshare.net/thelabdude/running-solr-in-the-cloud-at-memory-speed-with-alluxio/1

The problem I'm facing here is with autoAddReplicas. If I have 
replicationFactor=1 and the node with that replica dies, the node taking over 
incorrectly assigns the data directory. For example:

before
{code}"dataDir":"alluxio://localhost:19998/solr/test/",{code}

after
{code}"dataDir":"alluxio://localhost:19998/solr/test/core_node1/alluxio://localhost:19998/solr/test/",{code}

The same happens for ulogDir. Apparently, this has to do with this bit from 
HDFSDirectoryFactory:
{code}  public boolean isAbsolute(String path) {
return path.startsWith("hdfs:/");
  }{code}

If I add "alluxio:/" in there, the paths are correct and the index is recovered.

I see a few options here:
* add "alluxio:/" to the list there
* add a regular expression in the lines of \[a-z]*:/ I hope that's not too 
expensive, I'm not sure how often this method is called
* don't do anything and expect alluxio to work with an "hdfs:/" path? I 
actually tried that and didn't manage to make it work
* have a different DirectoryFactory or something else?

What do you think?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-SmokeRelease-7.x - Build # 59 - Still Failing

2017-10-12 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-SmokeRelease-7.x/59/

No tests ran.

Build Log:
[...truncated 28019 lines...]
prepare-release-no-sign:
[mkdir] Created dir: 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-7.x/lucene/build/smokeTestRelease/dist
 [copy] Copying 476 files to 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-7.x/lucene/build/smokeTestRelease/dist/lucene
 [copy] Copying 215 files to 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-7.x/lucene/build/smokeTestRelease/dist/solr
   [smoker] Java 1.8 JAVA_HOME=/home/jenkins/tools/java/latest1.8
   [smoker] NOTE: output encoding is UTF-8
   [smoker] 
   [smoker] Load release URL 
"file:/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-7.x/lucene/build/smokeTestRelease/dist/"...
   [smoker] 
   [smoker] Test Lucene...
   [smoker]   test basics...
   [smoker]   get KEYS
   [smoker] 0.2 MB in 0.02 sec (9.4 MB/sec)
   [smoker]   check changes HTML...
   [smoker]   download lucene-7.2.0-src.tgz...
   [smoker] 30.9 MB in 0.09 sec (337.3 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   download lucene-7.2.0.tgz...
   [smoker] 69.5 MB in 0.21 sec (334.3 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   download lucene-7.2.0.zip...
   [smoker] 79.9 MB in 0.25 sec (318.8 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   unpack lucene-7.2.0.tgz...
   [smoker] verify JAR metadata/identity/no javax.* or java.* classes...
   [smoker] test demo with 1.8...
   [smoker]   got 6221 hits for query "lucene"
   [smoker] checkindex with 1.8...
   [smoker] check Lucene's javadoc JAR
   [smoker]   unpack lucene-7.2.0.zip...
   [smoker] verify JAR metadata/identity/no javax.* or java.* classes...
   [smoker] test demo with 1.8...
   [smoker]   got 6221 hits for query "lucene"
   [smoker] checkindex with 1.8...
   [smoker] check Lucene's javadoc JAR
   [smoker]   unpack lucene-7.2.0-src.tgz...
   [smoker] make sure no JARs/WARs in src dist...
   [smoker] run "ant validate"
   [smoker] run tests w/ Java 8 and testArgs='-Dtests.slow=false'...
   [smoker] test demo with 1.8...
   [smoker]   got 213 hits for query "lucene"
   [smoker] checkindex with 1.8...
   [smoker] generate javadocs w/ Java 8...
   [smoker] 
   [smoker] Crawl/parse...
   [smoker] 
   [smoker] Verify...
   [smoker]   confirm all releases have coverage in TestBackwardsCompatibility
   [smoker] find all past Lucene releases...
   [smoker] run TestBackwardsCompatibility..
   [smoker] success!
   [smoker] 
   [smoker] Test Solr...
   [smoker]   test basics...
   [smoker]   get KEYS
   [smoker] 0.2 MB in 0.01 sec (17.8 MB/sec)
   [smoker]   check changes HTML...
   [smoker]   download solr-7.2.0-src.tgz...
   [smoker] 52.6 MB in 0.16 sec (327.7 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   download solr-7.2.0.tgz...
   [smoker] 143.8 MB in 0.43 sec (331.0 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   download solr-7.2.0.zip...
   [smoker] 144.8 MB in 0.43 sec (339.9 MB/sec)
   [smoker] verify md5/sha1 digests
   [smoker]   unpack solr-7.2.0.tgz...
   [smoker] verify JAR metadata/identity/no javax.* or java.* classes...
   [smoker] unpack lucene-7.2.0.tgz...
   [smoker]   **WARNING**: skipping check of 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-7.x/lucene/build/smokeTestRelease/tmp/unpack/solr-7.2.0/contrib/dataimporthandler-extras/lib/javax.mail-1.5.1.jar:
 it has javax.* classes
   [smoker]   **WARNING**: skipping check of 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-7.x/lucene/build/smokeTestRelease/tmp/unpack/solr-7.2.0/contrib/dataimporthandler-extras/lib/activation-1.1.1.jar:
 it has javax.* classes
   [smoker] copying unpacked distribution for Java 8 ...
   [smoker] test solr example w/ Java 8...
   [smoker]   start Solr instance 
(log=/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-7.x/lucene/build/smokeTestRelease/tmp/unpack/solr-7.2.0-java8/solr-example.log)...
   [smoker] No process found for Solr node running on port 8983
   [smoker]   Running techproducts example on port 8983 from 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-7.x/lucene/build/smokeTestRelease/tmp/unpack/solr-7.2.0-java8
   [smoker] Creating Solr home directory 
/home/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-7.x/lucene/build/smokeTestRelease/tmp/unpack/solr-7.2.0-java8/example/techproducts/solr
   [smoker] 
   [smoker] Starting up Solr on port 8983 using command:
   [smoker] "bin/solr" start -p 8983 -s "example/techproducts/solr"
   [smoker] 
   [smoker] Waiting up to 180 seconds to see Solr running on port 8983 [|]  
 [/]   [-]   [\]   [|]   [/]   [-]   
[\]   [|]   [/]   [-]   [\]  

[jira] [Updated] (SOLR-11472) Leader election bug

2017-10-12 Thread Andrzej Bialecki (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrzej Bialecki  updated SOLR-11472:
-
Attachment: Console_output_of_AutoscalingHistoryHandlerTest_failure.txt

> Leader election bug
> ---
>
> Key: SOLR-11472
> URL: https://issues.apache.org/jira/browse/SOLR-11472
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Affects Versions: 7.1, master (8.0)
>Reporter: Andrzej Bialecki 
> Attachments: 
> Console_output_of_AutoscalingHistoryHandlerTest_failure.txt
>
>
> SOLR-11407 uncovered a bug in leader election, where the same failing node is 
> retried indefinitely. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-11472) Leader election bug

2017-10-12 Thread Andrzej Bialecki (JIRA)
Andrzej Bialecki  created SOLR-11472:


 Summary: Leader election bug
 Key: SOLR-11472
 URL: https://issues.apache.org/jira/browse/SOLR-11472
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
Affects Versions: 7.1, master (8.0)
Reporter: Andrzej Bialecki 


SOLR-11407 uncovered a bug in leader election, where the same failing node is 
retried indefinitely. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-11471) MoveReplicaSuggester should suggest to move a leader only if no other replica is available

2017-10-12 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11471?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul updated SOLR-11471:
--
Summary: MoveReplicaSuggester should suggest to move a leader only if no 
other replica is available  (was: MoveReplicaSuggester should suggest to move a 
leader if no other replica is available)

> MoveReplicaSuggester should suggest to move a leader only if no other replica 
> is available
> --
>
> Key: SOLR-11471
> URL: https://issues.apache.org/jira/browse/SOLR-11471
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Noble Paul
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-11471) MoveReplicaSuggester should suggest to move a leader if no other replica is available

2017-10-12 Thread Noble Paul (JIRA)
Noble Paul created SOLR-11471:
-

 Summary: MoveReplicaSuggester should suggest to move a leader if 
no other replica is available
 Key: SOLR-11471
 URL: https://issues.apache.org/jira/browse/SOLR-11471
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
Reporter: Noble Paul






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (SOLR-11451) ComputePlanActionTest.testNodeLost() failure

2017-10-12 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul reassigned SOLR-11451:
-

Assignee: Noble Paul

> ComputePlanActionTest.testNodeLost() failure
> 
>
> Key: SOLR-11451
> URL: https://issues.apache.org/jira/browse/SOLR-11451
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Steve Rowe
>Assignee: Noble Paul
>
> Reproduces 100% for me (on Linux) if I remove {{-Dtests.method=testNodeLost}} 
> from the repro line, from 
> [https://jenkins.thetaphi.de/job/Lucene-Solr-master-Windows/6946]:
> {noformat}
> Checking out Revision e92bde1e7ece020d581638f4c59c3840bf75d183 
> (refs/remotes/origin/master)
> [...]
>[junit4]   2> NOTE: reproduce with: ant test  
> -Dtestcase=ComputePlanActionTest -Dtests.method=testNodeLost 
> -Dtests.seed=B1DFD0131B7C99E4 -Dtests.slow=true -Dtests.locale=ja 
> -Dtests.timezone=Asia/Katmandu -Dtests.asserts=true 
> -Dtests.file.encoding=UTF-8
>[junit4] FAILURE 8.09s J1 | ComputePlanActionTest.testNodeLost <<<
>[junit4]> Throwable #1: java.lang.AssertionError: The operations 
> computed by ComputePlanAction should not be null
>[junit4]>  at 
> __randomizedtesting.SeedInfo.seed([B1DFD0131B7C99E4:ECA1EED9896FC62]:0)
>[junit4]>  at 
> org.apache.solr.cloud.autoscaling.ComputePlanActionTest.testNodeLost(ComputePlanActionTest.java:193)
>[junit4]>  at java.lang.Thread.run(Thread.java:748)
> [...]
>[junit4]   2> NOTE: test params are: codec=Asserting(Lucene70): {}, 
> docValues:{}, maxPointsInLeafNode=491, maxMBSortInHeap=6.600294616233929, 
> sim=RandomSimilarity(queryNorm=true): {}, locale=ja, timezone=Asia/Katmandu
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-master-Linux (64bit/jdk-9) - Build # 20661 - Still Unstable!

2017-10-12 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/20661/
Java: 64bit/jdk-9 -XX:+UseCompressedOops -XX:+UseSerialGC --illegal-access=deny

5 tests failed.
FAILED:  junit.framework.TestSuite.org.apache.solr.cloud.BasicZkTest

Error Message:
SolrCore.getOpenCount()==2

Stack Trace:
java.lang.RuntimeException: SolrCore.getOpenCount()==2
at __randomizedtesting.SeedInfo.seed([E5C1B423C4CDC882]:0)
at org.apache.solr.util.TestHarness.close(TestHarness.java:379)
at org.apache.solr.SolrTestCaseJ4.deleteCore(SolrTestCaseJ4.java:774)
at 
org.apache.solr.cloud.AbstractZkTestCase.azt_afterClass(AbstractZkTestCase.java:147)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:564)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:897)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.base/java.lang.Thread.run(Thread.java:844)


FAILED:  junit.framework.TestSuite.org.apache.solr.cloud.BasicZkTest

Error Message:
SolrCore.getOpenCount()==2

Stack Trace:
java.lang.RuntimeException: SolrCore.getOpenCount()==2
at __randomizedtesting.SeedInfo.seed([E5C1B423C4CDC882]:0)
at org.apache.solr.util.TestHarness.close(TestHarness.java:379)
at org.apache.solr.SolrTestCaseJ4.deleteCore(SolrTestCaseJ4.java:774)
at 
org.apache.solr.SolrTestCaseJ4.teardownTestCases(SolrTestCaseJ4.java:287)
at jdk.internal.reflect.GeneratedMethodAccessor63.invoke(Unknown Source)
at 
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:564)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:897)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)

[JENKINS] Lucene-Solr-NightlyTests-7.x - Build # 60 - Still Unstable

2017-10-12 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-7.x/60/

12 tests failed.
FAILED:  
org.apache.lucene.codecs.lucene70.TestLucene70DocValuesFormat.testTermsEnumVariableWidth

Error Message:
Test abandoned because suite timeout was reached.

Stack Trace:
java.lang.Exception: Test abandoned because suite timeout was reached.
at __randomizedtesting.SeedInfo.seed([9EA1AFB6B4C0022A]:0)


FAILED:  
junit.framework.TestSuite.org.apache.lucene.codecs.lucene70.TestLucene70DocValuesFormat

Error Message:
Suite timeout exceeded (>= 720 msec).

Stack Trace:
java.lang.Exception: Suite timeout exceeded (>= 720 msec).
at __randomizedtesting.SeedInfo.seed([9EA1AFB6B4C0022A]:0)


FAILED:  org.apache.solr.cloud.BasicDistributedZkTest.test

Error Message:
Could not load collection from ZK: collection1

Stack Trace:
org.apache.solr.common.SolrException: Could not load collection from ZK: 
collection1
at 
__randomizedtesting.SeedInfo.seed([1F7247B850B685CC:97267862FE4AE834]:0)
at 
org.apache.solr.common.cloud.ZkStateReader.getCollectionLive(ZkStateReader.java:1170)
at 
org.apache.solr.common.cloud.ZkStateReader$LazyCollectionRef.get(ZkStateReader.java:690)
at 
org.apache.solr.common.cloud.ClusterState.getCollectionOrNull(ClusterState.java:130)
at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.getTotalReplicas(AbstractFullDistribZkTestBase.java:488)
at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.createJettys(AbstractFullDistribZkTestBase.java:441)
at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.createServers(AbstractFullDistribZkTestBase.java:334)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:991)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:968)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.e

[jira] [Commented] (SOLR-11445) Overseer should not hang when process bad message

2017-10-12 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-11445?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16201602#comment-16201602
 ] 

ASF subversion and git services commented on SOLR-11445:


Commit df5fefb0dbfca1783126798df9c7e303d40ace16 in lucene-solr's branch 
refs/heads/master from [~caomanhdat]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=df5fefb ]

SOLR-11445: Overseer should not hang when process bad message


> Overseer should not hang when process bad message
> -
>
> Key: SOLR-11445
> URL: https://issues.apache.org/jira/browse/SOLR-11445
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Affects Versions: 6.6.1, 7.0, master (8.0)
>Reporter: Greg Harris
> Attachments: SOLR-11445.patch, SOLR-11445.patch
>
>
> So we had the following stack trace with a customer:
> 2017-10-04 11:25:30.339 ERROR () [ ] o.a.s.c.Overseer Exception in 
> Overseer main queue loop
> org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = 
> NoNode for /collections//state.json
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
> at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:391)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:388)
> at 
> org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60)
> at org.apache.solr.common.cloud.SolrZkClient.create(SolrZkClient.java:388)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.writePendingUpdates(ZkStateWriter.java:235)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.run(Overseer.java:199)
> at java.lang.Thread.run(Thread.java:748)
> I want to highlight: 
>   at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> This ends up coming from Overseer:
> while (data != null)  {
> final ZkNodeProps message = ZkNodeProps.load(data);
> log.debug("processMessage: workQueueSize: {}, message = {}", 
> workQueue.getStats().getQueueLength(), message);
> // force flush to ZK after each message because there is no 
> fallback if workQueue items
> // are removed from workQueue but fail to be written to ZK
> *clusterState = processQueueItem(message, clusterState, 
> zkStateWriter, false, null);
> workQueue.poll(); // poll-ing removes the element we got by 
> peek-ing*
> data = workQueue.peek();
>   }
> Note: The processQueueItem comes before the poll, therefore upon a thrown 
> exception the same node/message that won't process becomes stuck. This made a 
> large cluster unable to come up on it's own without deleting the problem 
> node. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-11445) Overseer should not hang when process bad message

2017-10-12 Thread Cao Manh Dat (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Cao Manh Dat updated SOLR-11445:

Summary: Overseer should not hang when process bad message  (was: 
Overseer.processQueueItem()  zkStateWriter.enqueueUpdate might ideally have 
a try{}catch{} around it)

> Overseer should not hang when process bad message
> -
>
> Key: SOLR-11445
> URL: https://issues.apache.org/jira/browse/SOLR-11445
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Affects Versions: 6.6.1, 7.0, master (8.0)
>Reporter: Greg Harris
> Attachments: SOLR-11445.patch, SOLR-11445.patch
>
>
> So we had the following stack trace with a customer:
> 2017-10-04 11:25:30.339 ERROR () [ ] o.a.s.c.Overseer Exception in 
> Overseer main queue loop
> org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = 
> NoNode for /collections//state.json
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
> at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:391)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:388)
> at 
> org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60)
> at org.apache.solr.common.cloud.SolrZkClient.create(SolrZkClient.java:388)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.writePendingUpdates(ZkStateWriter.java:235)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.run(Overseer.java:199)
> at java.lang.Thread.run(Thread.java:748)
> I want to highlight: 
>   at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> This ends up coming from Overseer:
> while (data != null)  {
> final ZkNodeProps message = ZkNodeProps.load(data);
> log.debug("processMessage: workQueueSize: {}, message = {}", 
> workQueue.getStats().getQueueLength(), message);
> // force flush to ZK after each message because there is no 
> fallback if workQueue items
> // are removed from workQueue but fail to be written to ZK
> *clusterState = processQueueItem(message, clusterState, 
> zkStateWriter, false, null);
> workQueue.poll(); // poll-ing removes the element we got by 
> peek-ing*
> data = workQueue.peek();
>   }
> Note: The processQueueItem comes before the poll, therefore upon a thrown 
> exception the same node/message that won't process becomes stuck. This made a 
> large cluster unable to come up on it's own without deleting the problem 
> node. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-11445) Overseer.processQueueItem().... zkStateWriter.enqueueUpdate might ideally have a try{}catch{} around it

2017-10-12 Thread Cao Manh Dat (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-11445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Cao Manh Dat updated SOLR-11445:

Attachment: SOLR-11445.patch

Updated patch, including test. These are some changes
- Refactoring logic of how Overseer catching Exception
- The bad message will be polled out from the queue ( message that causes 
NoNodeException || NodeExistException || Any other exceptions - except 
InterruptedException and KeeperException )
- After poll out the bad message, we will continue processing workqueue

Will commit soon.

> Overseer.processQueueItem()  zkStateWriter.enqueueUpdate might ideally 
> have a try{}catch{} around it
> 
>
> Key: SOLR-11445
> URL: https://issues.apache.org/jira/browse/SOLR-11445
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Affects Versions: 6.6.1, 7.0, master (8.0)
>Reporter: Greg Harris
> Attachments: SOLR-11445.patch, SOLR-11445.patch
>
>
> So we had the following stack trace with a customer:
> 2017-10-04 11:25:30.339 ERROR () [ ] o.a.s.c.Overseer Exception in 
> Overseer main queue loop
> org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = 
> NoNode for /collections//state.json
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
> at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:391)
> at 
> org.apache.solr.common.cloud.SolrZkClient$9.execute(SolrZkClient.java:388)
> at 
> org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60)
> at org.apache.solr.common.cloud.SolrZkClient.create(SolrZkClient.java:388)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.writePendingUpdates(ZkStateWriter.java:235)
> at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.run(Overseer.java:199)
> at java.lang.Thread.run(Thread.java:748)
> I want to highlight: 
>   at 
> org.apache.solr.cloud.overseer.ZkStateWriter.enqueueUpdate(ZkStateWriter.java:152)
> at 
> org.apache.solr.cloud.Overseer$ClusterStateUpdater.processQueueItem(Overseer.java:271)
> This ends up coming from Overseer:
> while (data != null)  {
> final ZkNodeProps message = ZkNodeProps.load(data);
> log.debug("processMessage: workQueueSize: {}, message = {}", 
> workQueue.getStats().getQueueLength(), message);
> // force flush to ZK after each message because there is no 
> fallback if workQueue items
> // are removed from workQueue but fail to be written to ZK
> *clusterState = processQueueItem(message, clusterState, 
> zkStateWriter, false, null);
> workQueue.poll(); // poll-ing removes the element we got by 
> peek-ing*
> data = workQueue.peek();
>   }
> Note: The processQueueItem comes before the poll, therefore upon a thrown 
> exception the same node/message that won't process becomes stuck. This made a 
> large cluster unable to come up on it's own without deleting the problem 
> node. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org