[jira] [Commented] (SOLR-3575) solr.xml should default to persist=true

2012-12-04 Thread Magnus Bergmark (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3575?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509569#comment-13509569
 ] 

Magnus Bergmark commented on SOLR-3575:
---

 related: it would be nice if the coming/going of cores didn't necessitate 
 changing solr.xml
 Perhaps every directory under the solr home could implicitly define a 
 collection.

Not a good idea. We unload cores to save memory on some machines, then load 
them on demand. This would make it much harder for us to accomplish.

Also, what would happen if a core was unloaded?

 solr.xml should default to persist=true
 ---

 Key: SOLR-3575
 URL: https://issues.apache.org/jira/browse/SOLR-3575
 Project: Solr
  Issue Type: Improvement
Reporter: Mark Miller
Assignee: Mark Miller
 Fix For: 4.0-BETA, 5.0


 The default of false is kind of silly IMO.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-4586) Change default ResultMode of FacetRequest to PER_NODE_IN_TREE

2012-12-04 Thread Shai Erera (JIRA)
Shai Erera created LUCENE-4586:
--

 Summary: Change default ResultMode of FacetRequest to 
PER_NODE_IN_TREE
 Key: LUCENE-4586
 URL: https://issues.apache.org/jira/browse/LUCENE-4586
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/facet
Reporter: Shai Erera
Assignee: Shai Erera
Priority: Minor
 Fix For: 4.1, 5.0


Today the default ResultMode is GLOBAL_FLAT, but it should be PER_NODE_IN_TREE. 
ResultMode is being used whenever you set the depth of FacetRequest to greater 
than 1. The difference between the two is:

* PER_NODE_IN_TREE would then compute the top-K categories recursively, for 
every top category at every level (up to depth). The results are returned in a 
tree structure as well. For instance:

{noformat}
Date
  2010
March
February
  2011
April
May
{noformat}

* GLOBAL_FLAT computes the top categories among all the nodes up to depth, and 
returns a flat list of categories.

GLOBAL_FLAT is faster to compute than PER_NODE_IN_TREE (it just computes top-K 
among N total categories), however I think that it's less intuitive, and 
therefore should not be used as a default. In fact, I think this is kind of an 
expert usage.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4141) EDismax: Strange combination of subqueries with parentheses

2012-12-04 Thread Leonhard Maylein (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4141?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509590#comment-13509590
 ] 

Leonhard Maylein commented on SOLR-4141:


Oh, what a pity!
Then I have to check to user query in the front end first to
decide if the edismax parser is applicable or not.
If the user types in parentheses, there is no chance to use the
edismax handler because the user would be very surprised by
the the result.
Or, perhaps, I have to prefix every word within the parentheses
with a plus sign (and if there are camel case words I have to
separate the word parts before). But I think this is not the
whole purpose of the edismax query parser.

 EDismax: Strange combination of subqueries with parentheses
 ---

 Key: SOLR-4141
 URL: https://issues.apache.org/jira/browse/SOLR-4141
 Project: Solr
  Issue Type: Bug
  Components: query parsers
Affects Versions: 4.0
Reporter: Leonhard Maylein

 fi = field name, mm=100% (all examples)
 The query 'fi:a fi:b'
 (parsed query: '(+((fi:a fi:b)~2))/no_coord')
 is interpreted as 'fi:a AND fi:b'
 This also applies to the queries '(fi:a fi:b)' respectively 
 'fi:(a b)'.
 But the query '(fi:a fi:b) (fi:a fi:b)'
 (parsed query: '(+(((fi:a fi:b) (fi:a fi:b))~2))/no_coord')
 shows the same result as 'fi:a OR fi:b'.
 I'm not sure but I think this is a bug, isn't it?
 If it's a intended behavior I think it is very difficult
 to explain this to a searcher.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4580) Facet DrillDown should return a ConstantScoreQuery

2012-12-04 Thread Shai Erera (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4580?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509616#comment-13509616
 ] 

Shai Erera commented on LUCENE-4580:


I understand what you mean by 0f, sorry. I will fix it.

Also, I think that I'll simplify SimpleSearcher to create the FacetSearchParams 
up front, so that it's clear where they're created. In addition, I think that 
we can cut down the number of query() methods in DrillDown, to always take a 
baseQuery and handle the case that it's null. Will post an updated patch soon.

 Facet DrillDown should return a ConstantScoreQuery
 --

 Key: LUCENE-4580
 URL: https://issues.apache.org/jira/browse/LUCENE-4580
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/facet
Reporter: Shai Erera
Priority: Minor
 Attachments: LUCENE-4580.patch


 DrillDown is a helper class which the user can use to convert a facet value 
 that a user selected into a Query for performing drill-down or narrowing the 
 results. The API has several static methods that create e.g. a Term or Query.
 Rather than creating a Query, it would make more sense to create a Filter I 
 think. In most cases, the clicked facets should not affect the scoring of 
 documents. Anyway, even if it turns out that it must return a Query (which I 
 doubt), we should at least modify the impl to return a ConstantScoreQuery.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4580) Facet DrillDown should return a ConstantScoreQuery

2012-12-04 Thread Shai Erera (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4580?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shai Erera updated LUCENE-4580:
---

Attachment: LUCENE-4580.patch

* Removed DrillDown.query() which didn't take baseQuery -- DrillDown will 
usually be used for drill-down queries. However, the query() methods handle the 
case where baseQuery is null.

* Fixed SimpleSearcher.searchWithDrillDown to initialize FacetIndexingParams 
and pass it to the other methods.

I think this is ready to go in.

 Facet DrillDown should return a ConstantScoreQuery
 --

 Key: LUCENE-4580
 URL: https://issues.apache.org/jira/browse/LUCENE-4580
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/facet
Reporter: Shai Erera
Priority: Minor
 Attachments: LUCENE-4580.patch, LUCENE-4580.patch


 DrillDown is a helper class which the user can use to convert a facet value 
 that a user selected into a Query for performing drill-down or narrowing the 
 results. The API has several static methods that create e.g. a Term or Query.
 Rather than creating a Query, it would make more sense to create a Filter I 
 think. In most cases, the clicked facets should not affect the scoring of 
 documents. Anyway, even if it turns out that it must return a Query (which I 
 doubt), we should at least modify the impl to return a ConstantScoreQuery.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4580) Facet DrillDown should return a ConstantScoreQuery

2012-12-04 Thread Gilad Barkai (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4580?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509628#comment-13509628
 ] 

Gilad Barkai commented on LUCENE-4580:
--

+1

 Facet DrillDown should return a ConstantScoreQuery
 --

 Key: LUCENE-4580
 URL: https://issues.apache.org/jira/browse/LUCENE-4580
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/facet
Reporter: Shai Erera
Priority: Minor
 Attachments: LUCENE-4580.patch, LUCENE-4580.patch


 DrillDown is a helper class which the user can use to convert a facet value 
 that a user selected into a Query for performing drill-down or narrowing the 
 results. The API has several static methods that create e.g. a Term or Query.
 Rather than creating a Query, it would make more sense to create a Filter I 
 think. In most cases, the clicked facets should not affect the scoring of 
 documents. Anyway, even if it turns out that it must return a Query (which I 
 doubt), we should at least modify the impl to return a ConstantScoreQuery.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4580) Facet DrillDown should return a ConstantScoreQuery

2012-12-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4580?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509632#comment-13509632
 ] 

Commit Tag Bot commented on LUCENE-4580:


[trunk commit] Shai Erera
http://svn.apache.org/viewvc?view=revisionrevision=1416860

LUCENE-4580: Facet DrillDown should return a ConstantScoreQuery



 Facet DrillDown should return a ConstantScoreQuery
 --

 Key: LUCENE-4580
 URL: https://issues.apache.org/jira/browse/LUCENE-4580
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/facet
Reporter: Shai Erera
Priority: Minor
 Attachments: LUCENE-4580.patch, LUCENE-4580.patch


 DrillDown is a helper class which the user can use to convert a facet value 
 that a user selected into a Query for performing drill-down or narrowing the 
 results. The API has several static methods that create e.g. a Term or Query.
 Rather than creating a Query, it would make more sense to create a Filter I 
 think. In most cases, the clicked facets should not affect the scoring of 
 documents. Anyway, even if it turns out that it must return a Query (which I 
 doubt), we should at least modify the impl to return a ConstantScoreQuery.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



dev-tools/eclipse/dot.classpath is broken?

2012-12-04 Thread Shai Erera
Hi

It references lucene/analysis/icu/src/test-files but I don't see such
directory, and eclipse complains.

Is it safe to remove it from dot.classpath or has it been moved elsewhere?

Shai


[jira] [Resolved] (LUCENE-4580) Facet DrillDown should return a ConstantScoreQuery

2012-12-04 Thread Shai Erera (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4580?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shai Erera resolved LUCENE-4580.


   Resolution: Fixed
Fix Version/s: 5.0
   4.1
 Assignee: Shai Erera
Lucene Fields: New,Patch Available  (was: New)

Committed to trunk and 4x.

 Facet DrillDown should return a ConstantScoreQuery
 --

 Key: LUCENE-4580
 URL: https://issues.apache.org/jira/browse/LUCENE-4580
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/facet
Reporter: Shai Erera
Assignee: Shai Erera
Priority: Minor
 Fix For: 4.1, 5.0

 Attachments: LUCENE-4580.patch, LUCENE-4580.patch


 DrillDown is a helper class which the user can use to convert a facet value 
 that a user selected into a Query for performing drill-down or narrowing the 
 results. The API has several static methods that create e.g. a Term or Query.
 Rather than creating a Query, it would make more sense to create a Filter I 
 think. In most cases, the clicked facets should not affect the scoring of 
 documents. Anyway, even if it turns out that it must return a Query (which I 
 doubt), we should at least modify the impl to return a ConstantScoreQuery.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4580) Facet DrillDown should return a ConstantScoreQuery

2012-12-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4580?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509647#comment-13509647
 ] 

Commit Tag Bot commented on LUCENE-4580:


[branch_4x commit] Shai Erera
http://svn.apache.org/viewvc?view=revisionrevision=1416864

LUCENE-4580: Facet DrillDown should return a ConstantScoreQuery



 Facet DrillDown should return a ConstantScoreQuery
 --

 Key: LUCENE-4580
 URL: https://issues.apache.org/jira/browse/LUCENE-4580
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/facet
Reporter: Shai Erera
Assignee: Shai Erera
Priority: Minor
 Fix For: 4.1, 5.0

 Attachments: LUCENE-4580.patch, LUCENE-4580.patch


 DrillDown is a helper class which the user can use to convert a facet value 
 that a user selected into a Query for performing drill-down or narrowing the 
 results. The API has several static methods that create e.g. a Term or Query.
 Rather than creating a Query, it would make more sense to create a Filter I 
 think. In most cases, the clicked facets should not affect the scoring of 
 documents. Anyway, even if it turns out that it must return a Query (which I 
 doubt), we should at least modify the impl to return a ConstantScoreQuery.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-trunk-Linux (64bit/jdk1.7.0_09) - Build # 3055 - Failure!

2012-12-04 Thread Policeman Jenkins Server
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-trunk-Linux/3055/
Java: 64bit/jdk1.7.0_09 -XX:+UseConcMarkSweepGC

1 tests failed.
REGRESSION:  org.apache.lucene.facet.search.DrillDownTest.testScoring

Error Message:
score of doc=4 modified expected:0.804885745048523 but 
was:0.5365905165672302

Stack Trace:
java.lang.AssertionError: score of doc=4 modified expected:0.804885745048523 
but was:0.5365905165672302
at 
__randomizedtesting.SeedInfo.seed([3C984CDCB25F27BB:1EBA4BE3B82B45EA]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.failNotEquals(Assert.java:647)
at org.junit.Assert.assertEquals(Assert.java:443)
at 
org.apache.lucene.facet.search.DrillDownTest.testScoring(DrillDownTest.java:208)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:601)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:43)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
at java.lang.Thread.run(Thread.java:722)




Build Log:
[...truncated 6528 lines...]
[junit4:junit4] Suite: org.apache.lucene.facet.search.DrillDownTest
[junit4:junit4]   2 NOTE: reproduce with: ant test  -Dtestcase=DrillDownTest 
-Dtests.method=testScoring -Dtests.seed=3C984CDCB25F27BB -Dtests.multiplier=3 
-Dtests.slow=true -Dtests.locale=fr_LU -Dtests.timezone=Pacific/Tongatapu 

[jira] [Commented] (SOLR-4114) Collection API: Allow multiple shards from one collection on the same Solr server

2012-12-04 Thread Per Steffensen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509687#comment-13509687
 ] 

Per Steffensen commented on SOLR-4114:
--

bq. We have to do something else other than a one minute sleep in the test.

In checkForCollection we are willing to wait up to 120 secs (used to be 60 
secs) for the collection to be created (correctly). So it is kinda obvious to 
also wait at least 60 secs to verify that nothing related to a collection was 
actually created in cases where you want to be sure that nothing is created. In 
the case of waiting for a collection being created correctly you can skip the 
waiting as soon as you see that is has been created correctly. In the case 
where you want to see that nothing gets created you have no chance to do that. 
I think the 60 sec wait fits well into the way the test already works.

bq. I'd like the default data dir to remain /data. If you are not doing 
anything special that works fine.

Thats also fine, but as soon at you create more than one shard on the same 
node, dont you need different data-dirs? You cant have two shards using the 
same folder for lucene-index-files, right? Did I get the instance-dir and 
data-dir concepts wrong? Or is it the instance-dir you want to vary between 
different shards? I would say that running only one shard on a node IS doing 
something special.

bq. I'd also like the ability to override the data dir on per shard basis, but 
I'm not sure what the best API for that is.

Yes, me too, but that has to be another ticket/issue. For now, until the API 
allowes to control the data-dir-naming, it should be ok for the algorithm to 
decide for something reasonable by itself.

bq. So I'd like to support what you want, but I don't want to change the 
default behavior.

I agree, but if I got the concepts correct you cannot use /data as data-dir 
for more than one shard on each node. So default behaviour about using /data 
as data-dir will not work as soon as you run more than one shard on a node. I 
probably got something wrong - please try to explain.

bq. My latest patch - I'll commit this soon and we can iterate from there.

Well I would prefer you committed my patch, and we can iterate from there :-) 
It will also make it much easier to get SOLR-4120 in, which I hope you will 
also consider.

Had a quick peek at your patch and have the following comments
* I see that you removed the auto-reduce replica-per-shard to never have more 
than one replica of the same shard on the same node-feature and just issue a 
warning instead in OverseerCollectionProcessor (the if (numShardsPerSlice  
nodeList.size())-thingy). It is ok for me, eventhough I believe it is 
pointless to replicate data to the same machine and under the same Solr 
instance. But then you probably need to change the BasicDistrbutedZkTest also - 
in checkCollectionExceptations I believe you'll need to change from
{code}
int expectedShardsPerSlice = (Math.min(clusterState.getLiveNodes().size(), 
numShardsNumReplicaList.get(1) + 1));
{code}
to
{code}
int expectedShardsPerSlice = numShardsNumReplicaList.get(1) + 1;
{code}
* You removed the following lines (because you just want default-values for 
instance-dir and data-dir)
{code}
params.set(CoreAdminParams.INSTANCE_DIR, .);
params.set(CoreAdminParams.DATA_DIR, shardName + _data);
{code}
I still do not understand how that will work, but hopefully you will explain
* You didnt like my rename of variable replica to 
nodeUrlWithoutProtocolPart in OverseerCollectionProcessor.createCollection. 
As I said on mailing-list I dont like the term replica as a replacement for 
what we used to call shards, because I think it will cause misunderstandings, 
as replica is also (in my mind) a role played at runtime. But getting the 
terms right and reflect them correctly in API, variable-names etc. across the 
code must be another issue/ticket. But here in this specific example replica 
is just a very bad name, because the variable is not even containing a 
replica-url, which would require the shard/replica-name to be postfixed to 
the string. So this replica-variable is closest to being an node-url (without 
the protocol part) - NOT a shard/replica-url. I would accept my name-change if 
I where you, but I have a motto of carefully choosing my fights and this is a 
fight I will not fight for very long :-)
* I see that you did not include my changes to HttpShardHandler, making 
shardToUrls-map (renamed) concurrency protected through getURLs-method 
(renamed to getFullURLs), so that you do not have to use the map so carefully 
outside. I understand that it has very little to do with this issue SOLR-4114, 
but it is a nice modification. Please consider committing it - maybe related to 
another issue/ticket. It is little bit of a problem that good refactoring does 
not easy get in as part of issues/tickets not requiring the refactor. If 

Re: [JENKINS] Lucene-Solr-trunk-Linux (64bit/jdk1.7.0_09) - Build # 3055 - Failure!

2012-12-04 Thread Shai Erera
It looks to be caused by the random IndexSearcher that's created. I'm still
checking if it can alter the scoring, since that test verifies that scores
between two invocations of IndexSearcher.search() do not change.

Anyway, this looks like a test issue.

Shai

On Tue, Dec 4, 2012 at 1:45 PM, Policeman Jenkins Server 
jenk...@sd-datasolutions.de wrote:

 Build:
 http://jenkins.sd-datasolutions.de/job/Lucene-Solr-trunk-Linux/3055/
 Java: 64bit/jdk1.7.0_09 -XX:+UseConcMarkSweepGC

 1 tests failed.
 REGRESSION:  org.apache.lucene.facet.search.DrillDownTest.testScoring

 Error Message:
 score of doc=4 modified expected:0.804885745048523 but
 was:0.5365905165672302

 Stack Trace:
 java.lang.AssertionError: score of doc=4 modified
 expected:0.804885745048523 but was:0.5365905165672302
 at
 __randomizedtesting.SeedInfo.seed([3C984CDCB25F27BB:1EBA4BE3B82B45EA]:0)
 at org.junit.Assert.fail(Assert.java:93)
 at org.junit.Assert.failNotEquals(Assert.java:647)
 at org.junit.Assert.assertEquals(Assert.java:443)
 at
 org.apache.lucene.facet.search.DrillDownTest.testScoring(DrillDownTest.java:208)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
 at
 org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
 at
 org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
 at
 org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
 at
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
 at
 org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
 at
 org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
 at
 org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
 at
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
 at
 com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
 at
 com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782)
 at
 com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693)
 at
 org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
 at
 org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
 at
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
 at
 com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
 at
 com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
 at
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
 at
 org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:43)
 at
 org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
 at
 org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
 at
 org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
 at
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
 at
 com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
 

[jira] [Updated] (SOLR-3926) solrj should support better way of finding active sorts

2012-12-04 Thread Eirik Lygre (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3926?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eirik Lygre updated SOLR-3926:
--

Attachment: SOLR-3926.patch

This patch implements the API as discussed above. The patch contains javadoc 
and unit tests, and was built against 
http://svn.apache.org/repos/asf/lucene/dev/branches/branch_4x.

This patch solves the original requirements that led to this issue 
(introspection of active sorts), and hopefully also the requirements of the 
Solr project (as represented by [~yo...@apache.org] and [~hossman]).

 solrj should support better way of finding active sorts
 ---

 Key: SOLR-3926
 URL: https://issues.apache.org/jira/browse/SOLR-3926
 Project: Solr
  Issue Type: Improvement
  Components: clients - java
Affects Versions: 4.0
Reporter: Eirik Lygre
Priority: Minor
 Fix For: 4.1

 Attachments: SOLR-3926.patch, SOLR-3926.patch, SOLR-3926.patch, 
 SOLR-3926.patch


 The Solrj api uses ortogonal concepts for setting/removing and getting sort 
 information. Setting/removing uses a combination of (name,order), while 
 getters return a String name order:
 {code}
 public SolrQuery setSortField(String field, ORDER order);
 public SolrQuery addSortField(String field, ORDER order);
 public SolrQuery removeSortField(String field, ORDER order);
 public String[] getSortFields();
 public String getSortField();
 {code}
 If you want to use the current sort information to present a list of active 
 sorts, with the possibility to remove then, you need to manually parse the 
 string(s) returned from getSortFields, to recreate the information required 
 by removeSortField(). Not difficult, but not convenient either :-)
 Therefore this suggestion: Add a new method {{public MapString,ORDER 
 getSortFieldMap();}} which returns an ordered map of active sort fields. This 
 will make introspection of the current sort setup much easier.
 {code}
   public MapString, ORDER getSortFieldMap() {
 String[] actualSortFields = getSortFields();
 if (actualSortFields == null || actualSortFields.length == 0)
   return Collections.emptyMap();
 MapString, ORDER sortFieldMap = new LinkedHashMapString, ORDER();
 for (String sortField : actualSortFields) {
   String[] fieldSpec = sortField.trim().split( );
   sortFieldMap.put(fieldSpec[0], ORDER.valueOf(fieldSpec[1]));
 }
 return Collections.unmodifiableMap(sortFieldMap);
   }
 {code}
 For what it's worth, this is possible client code:
 {code}
 System.out.println(Active sorts);
 MapString, ORDER fieldMap = getSortFieldMap(query);
 for (String field : fieldMap.keySet()) {
System.out.println(-  + field + ; dir= + fieldMap.get(field));
 }
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-4.x-Linux (32bit/jdk1.8.0-ea-b65) - Build # 3046 - Failure!

2012-12-04 Thread Policeman Jenkins Server
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Linux/3046/
Java: 32bit/jdk1.8.0-ea-b65 -client -XX:+UseSerialGC

1 tests failed.
REGRESSION:  org.apache.lucene.facet.search.DrillDownTest.testScoring

Error Message:
score of doc=4 modified expected:2.304394006729126 but 
was:1.5362627506256104

Stack Trace:
java.lang.AssertionError: score of doc=4 modified expected:2.304394006729126 
but was:1.5362627506256104
at 
__randomizedtesting.SeedInfo.seed([D1495224FFA8AD3F:F36B551BF5DCCF6E]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.failNotEquals(Assert.java:647)
at org.junit.Assert.assertEquals(Assert.java:443)
at 
org.apache.lucene.facet.search.DrillDownTest.testScoring(DrillDownTest.java:208)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:474)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:43)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
at java.lang.Thread.run(Thread.java:722)




Build Log:
[...truncated 6398 lines...]
[junit4:junit4] Suite: org.apache.lucene.facet.search.DrillDownTest
[junit4:junit4]   2 NOTE: reproduce with: ant test  -Dtestcase=DrillDownTest 
-Dtests.method=testScoring -Dtests.seed=D1495224FFA8AD3F -Dtests.multiplier=3 
-Dtests.slow=true -Dtests.locale=de_AT -Dtests.timezone=America/Cayenne 

[jira] [Commented] (SOLR-4141) EDismax: Strange combination of subqueries with parentheses

2012-12-04 Thread Jack Krupansky (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4141?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509725#comment-13509725
 ] 

Jack Krupansky commented on SOLR-4141:
--

Feel free to suggest an improvement to make edismax at least customizable to 
address your needs.

 EDismax: Strange combination of subqueries with parentheses
 ---

 Key: SOLR-4141
 URL: https://issues.apache.org/jira/browse/SOLR-4141
 Project: Solr
  Issue Type: Bug
  Components: query parsers
Affects Versions: 4.0
Reporter: Leonhard Maylein

 fi = field name, mm=100% (all examples)
 The query 'fi:a fi:b'
 (parsed query: '(+((fi:a fi:b)~2))/no_coord')
 is interpreted as 'fi:a AND fi:b'
 This also applies to the queries '(fi:a fi:b)' respectively 
 'fi:(a b)'.
 But the query '(fi:a fi:b) (fi:a fi:b)'
 (parsed query: '(+(((fi:a fi:b) (fi:a fi:b))~2))/no_coord')
 shows the same result as 'fi:a OR fi:b'.
 I'm not sure but I think this is a bug, isn't it?
 If it's a intended behavior I think it is very difficult
 to explain this to a searcher.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (64bit/jdk1.7.0_09) - Build # 3055 - Failure!

2012-12-04 Thread Shai Erera
It was a real bug -- DrilDown.query did not set disableCoord=true on the
result BooleanQuery, and RandomSimProvider caught that !

Committed a fix.

Shai

On Tue, Dec 4, 2012 at 2:05 PM, Shai Erera ser...@gmail.com wrote:

 It looks to be caused by the random IndexSearcher that's created. I'm
 still checking if it can alter the scoring, since that test verifies that
 scores between two invocations of IndexSearcher.search() do not change.

 Anyway, this looks like a test issue.

 Shai

 On Tue, Dec 4, 2012 at 1:45 PM, Policeman Jenkins Server 
 jenk...@sd-datasolutions.de wrote:

 Build:
 http://jenkins.sd-datasolutions.de/job/Lucene-Solr-trunk-Linux/3055/
 Java: 64bit/jdk1.7.0_09 -XX:+UseConcMarkSweepGC

 1 tests failed.
 REGRESSION:  org.apache.lucene.facet.search.DrillDownTest.testScoring

 Error Message:
 score of doc=4 modified expected:0.804885745048523 but
 was:0.5365905165672302

 Stack Trace:
 java.lang.AssertionError: score of doc=4 modified
 expected:0.804885745048523 but was:0.5365905165672302
 at
 __randomizedtesting.SeedInfo.seed([3C984CDCB25F27BB:1EBA4BE3B82B45EA]:0)
 at org.junit.Assert.fail(Assert.java:93)
 at org.junit.Assert.failNotEquals(Assert.java:647)
 at org.junit.Assert.assertEquals(Assert.java:443)
 at
 org.apache.lucene.facet.search.DrillDownTest.testScoring(DrillDownTest.java:208)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:601)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
 at
 org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
 at
 org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
 at
 org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
 at
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
 at
 org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
 at
 org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
 at
 org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
 at
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
 at
 com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
 at
 com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782)
 at
 com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682)
 at
 com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693)
 at
 org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
 at
 org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
 at
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
 at
 com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
 at
 com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
 at
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
 at
 org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:43)
 at
 org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
 at
 org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
 at
 

Re: dev-tools/eclipse/dot.classpath is broken?

2012-12-04 Thread Steve Rowe
Hi Shai - it's safe to remove.
On Dec 4, 2012 5:09 AM, Shai Erera ser...@gmail.com wrote:

 Hi

 It references lucene/analysis/icu/src/test-files but I don't see such
 directory, and eclipse complains.

 Is it safe to remove it from dot.classpath or has it been moved elsewhere?

 Shai



Re: dev-tools/eclipse/dot.classpath is broken?

2012-12-04 Thread Shai Erera
ok done.

On Tue, Dec 4, 2012 at 3:50 PM, Steve Rowe sar...@gmail.com wrote:

 Hi Shai - it's safe to remove.
 On Dec 4, 2012 5:09 AM, Shai Erera ser...@gmail.com wrote:

 Hi

 It references lucene/analysis/icu/src/test-files but I don't see such
 directory, and eclipse complains.

 Is it safe to remove it from dot.classpath or has it been moved elsewhere?

 Shai




[jira] [Commented] (SOLR-1028) Automatic core loading unloading for multicore

2012-12-04 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1028?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509759#comment-13509759
 ] 

Yonik Seeley commented on SOLR-1028:


TestLazyCores seems to have failed a number of times since it was added here.
http://svn.apache.org/viewvc?rev=1403710view=rev
Anyone have any clues?

 Automatic core loading unloading for multicore
 --

 Key: SOLR-1028
 URL: https://issues.apache.org/jira/browse/SOLR-1028
 Project: Solr
  Issue Type: New Feature
  Components: multicore
Affects Versions: 4.0, 5.0
Reporter: Noble Paul
Assignee: Erick Erickson
 Fix For: 4.1, 5.0

 Attachments: SOLR-1028.patch, SOLR-1028.patch


 usecase: I have many small cores (say one per user) on a single Solr box . 
 All the cores are not be always needed . But when I need it I should be able 
 to directly issue a search request and the core must be STARTED automatically 
 and the request must be served.
 This also requires that I must have an upper limit on the no:of cores that 
 should be loaded at any given point in time. If the limit is crossed the 
 CoreContainer must unload a core (preferably the least recently used core)  
 There must be a choice of specifying some cores as fixed. These cores must 
 never be unloaded 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-3425) CloudSolrServer can't create cores when using the zkHost based constructor

2012-12-04 Thread Per Steffensen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3425?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509769#comment-13509769
 ] 

Per Steffensen edited comment on SOLR-3425 at 12/4/12 2:42 PM:
---

You talk about using CloudSolrServer and/or LBHttpSolrServer for access to Core 
Admin API. This a little strange because:
* A Core Admin API request should be sent to a specific Solr node in order for 
you to control where the shard is created (or removed from)
* CloudSolrServer and LBHttpSolrServer are all about sending requests that can 
end up going to a (randmon) node among several nodes
With the quick-fix code above (Tommaso Teofili 01/May/12) you will end up 
having your shard created on a random node among all live nodes - it is very 
rarely want you want.

So as long as we are talking about accessing Core Admin API you probably always 
want to use HttpSolrServer, which is aimed a sending the request to a specific 
node.

But when talking about creating an entire collection consisting of many shards, 
it is certainly meaningful to use CloudSolrServer. To create entire collections 
(without having to create each shard yourself using the Core Admin API) we now 
have the Collection API in 4.0.0. The Collection API can be used through 
CloudSolrServer, except...
* You cant create your first collection through CloudSolrServer
* You cant create a collection through a CloudSolrServer, where 
default-collection is the collection you want to create
* etc

...basically because CloudSolrServer wants an existing collection (pointed to 
by its default-collection or a collection-name provided in the actual request) 
before it can do anything.
This will be fixed with SOLR-4140, but is not in 4.0.0.

Other things not in Collection API of 4.0.0
* You cant have more than one shard per collection on a single node - fixed in 
TLT-4114
* You cant specify which Solr nodes the shards for a new collection is allowed 
to be spread across - they are just spread across all live Solr nodes - fixed 
in TLT-4120

Please state if you still believe something is missing or unclear. Or else you 
might want to state that your problems are solved with the 4.0.0 Collection 
API (maybe plus one or more of SOLR-4140, SOLR-4114 and SOLR-4120 which will 
hopefully be in the next release) by closing this issue SOLR-3425

  was (Author: steff1193):
You talk about using CloudSolrServer and/or LBHttpSolrServer for access to 
Core Admin API. This a little strange because:
* A Core Admin API request should be sent to a specific Solr node in order for 
you to control where the shard is created (or removed from)
* CloudSolrServer and LBHttpSolrServer are all about sending requests that can 
end up going to a (randmon) node among several nodes
With the quick-fix code above (Tommaso Teofili 01/May/12) you will end up 
having your shard created on a random node among all live nodes - it is very 
rarely want you want.

So as long as we are talking about accessing Core Admin API you probably always 
want to use HttpSolrServer, which is aimed a sending the request to a specific 
node.

But when talking about creating an entire collection consisting of many shards, 
it is certainly meaningful to use CloudSolrServer. To create entire collections 
(without having to create each shard yourself using the Core Admin API) we now 
have the Collection API in 4.0.0. The Collection API can be used through 
CloudSolrServer, except...
* You cant create your first collection through CloudSolrServer
* You cant create a collection through a CloudSolrServer, where 
default-collection is the collection you want to create
* etc
...basically because CloudSolrServer wants an existing collection (pointed to 
by its default-collection or a collection-name provided in the actual request) 
before it can do anything.
This will be fixed with SOLR-4140, but is not in 4.0.0.

Other things not in Collection API of 4.0.0
* You cant have more than one shard per collection on a single node - fixed in 
TLT-4114
* You cant specify which Solr nodes the shards for a new collection is allowed 
to be spread across - they are just spread across all live Solr nodes - fixed 
in TLT-4120

Please state if you still believe something is missing or unclear. Or else you 
might want to state that your problems are solved with the 4.0.0 Collection 
API (maybe plus one or more of SOLR-4140, SOLR-4114 and SOLR-4120 which will 
hopefully be in the next release) by closing this issue SOLR-3425
  
 CloudSolrServer can't create cores when using the zkHost based constructor
 --

 Key: SOLR-3425
 URL: https://issues.apache.org/jira/browse/SOLR-3425
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Reporter: Tommaso 

[jira] [Commented] (SOLR-3425) CloudSolrServer can't create cores when using the zkHost based constructor

2012-12-04 Thread Per Steffensen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3425?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509769#comment-13509769
 ] 

Per Steffensen commented on SOLR-3425:
--

You talk about using CloudSolrServer and/or LBHttpSolrServer for access to Core 
Admin API. This a little strange because:
* A Core Admin API request should be sent to a specific Solr node in order for 
you to control where the shard is created (or removed from)
* CloudSolrServer and LBHttpSolrServer are all about sending requests that can 
end up going to a (randmon) node among several nodes
With the quick-fix code above (Tommaso Teofili 01/May/12) you will end up 
having your shard created on a random node among all live nodes - it is very 
rarely want you want.

So as long as we are talking about accessing Core Admin API you probably always 
want to use HttpSolrServer, which is aimed a sending the request to a specific 
node.

But when talking about creating an entire collection consisting of many shards, 
it is certainly meaningful to use CloudSolrServer. To create entire collections 
(without having to create each shard yourself using the Core Admin API) we now 
have the Collection API in 4.0.0. The Collection API can be used through 
CloudSolrServer, except...
* You cant create your first collection through CloudSolrServer
* You cant create a collection through a CloudSolrServer, where 
default-collection is the collection you want to create
* etc
...basically because CloudSolrServer wants an existing collection (pointed to 
by its default-collection or a collection-name provided in the actual request) 
before it can do anything.
This will be fixed with SOLR-4140, but is not in 4.0.0.

Other things not in Collection API of 4.0.0
* You cant have more than one shard per collection on a single node - fixed in 
TLT-4114
* You cant specify which Solr nodes the shards for a new collection is allowed 
to be spread across - they are just spread across all live Solr nodes - fixed 
in TLT-4120

Please state if you still believe something is missing or unclear. Or else you 
might want to state that your problems are solved with the 4.0.0 Collection 
API (maybe plus one or more of SOLR-4140, SOLR-4114 and SOLR-4120 which will 
hopefully be in the next release) by closing this issue SOLR-3425

 CloudSolrServer can't create cores when using the zkHost based constructor
 --

 Key: SOLR-3425
 URL: https://issues.apache.org/jira/browse/SOLR-3425
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Reporter: Tommaso Teofili
Assignee: Mark Miller
Priority: Minor
 Fix For: 4.1

 Attachments: SOLR-3425-test.patch


 When programmatically creating cores with a running SolrCloud instance the 
 CloudSolrServer uses the slices nodes information to feed the underlying 
 LBHttpSolrServer so it fails to create cores as there aren't any slices for 
 any new collection (as it's still to be created).
 This happens when using the CloudSolrServer constructor which takes the ZK 
 host as only parameter while it can be avoided by using the constructor which 
 also takes the list of Solr URLs and the underlying LBHttpSolrServer is 
 actually used for making the core creation request.
 However it'd be good to use the ZK host live nodes information to 
 automatically issue a core creation command on one of the underlying Solr 
 hosts without having to specify the full list of URLs beforehand.
 The scenario is when one wants to create a collection with N shards so the 
 client sends N core creation requests for the same collection thus the 
 SolrCloud stuff should just take care of choosing the host where to issue the 
 core creation request and update the cluster state.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4140) Collection API: CloudSolrServer cannot be used to create collection with referencing an existing collection

2012-12-04 Thread Per Steffensen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4140?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509774#comment-13509774
 ] 

Per Steffensen commented on SOLR-4140:
--

SOLR-3425 is certainly related, but if there are parts in it that are not 
solved with SOLR-4140, I am not sure. I posted a comment to SOLR-3425, 
describing this and that, and basically asking for a new evaluation of whether 
SOLR-3425 is still relevant (or can be closed) after SOLR-4140

 Collection API: CloudSolrServer cannot be used to create collection with 
 referencing an existing collection
 ---

 Key: SOLR-4140
 URL: https://issues.apache.org/jira/browse/SOLR-4140
 Project: Solr
  Issue Type: Bug
  Components: multicore, SolrCloud
Affects Versions: 4.0
Reporter: Per Steffensen
Assignee: Per Steffensen
Priority: Minor
  Labels: collection-api, multicore
 Attachments: SOLR-4140_trunk.patch


 CloudSolrServer needs an existing collection to be able to forward requests. 
 For some requests it is not logical that you ned to have or reference an 
 existing collection. E.g. if you want to use CloudSolrServer to create a 
 collection - especially:
 * If you have no collections already and want to create your first collection
 * If you like to create your CloudSolrServer once and for all, and want to 
 set default-collection on it to point to your favorite collection, but also 
 use this CloudSolrServer to create this favorite collection
 Yes, HttpSolrServer can be used, but
 * Some of us like to use CloudSolrServer for everything
 * Using HttpSolrServer you need to check youself which nodes are running, in 
 order to send the create-request to one of them

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4137) FastVectorHighlighter: StringIndexOutOfBoundsException in BaseFragmentsBuilder

2012-12-04 Thread Markus Jelsma (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4137?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509777#comment-13509777
 ] 

Markus Jelsma commented on SOLR-4137:
-

I have to add that we're also seeing this rather frequently with analyzers that 
use the WDF.

 FastVectorHighlighter: StringIndexOutOfBoundsException in BaseFragmentsBuilder
 --

 Key: SOLR-4137
 URL: https://issues.apache.org/jira/browse/SOLR-4137
 Project: Solr
  Issue Type: Bug
  Components: highlighter
Affects Versions: 3.6.1
Reporter: Marcel

 under some circumstances the BaseFragmentsBuilder genereates a 
 StringIndexOutOfBoundsException inside the makeFragment method.
 The starting offset is higher than the end offset.
 I did a small patch checking the offsets and posted it over there at 
 Stackoverflow: 
 http://stackoverflow.com/questions/12456448/solr-highlight-bug-with-usefastvectorhighlighter
 The code in 4.0 seems to be the same as in 3.6.1
 Example how to reproduce the behaviour:
 There is a word called www.DAKgesundAktivBonus.de inside the index. If you 
 search for dak bonus some offset calculations went wrong.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-3425) CloudSolrServer can't create cores when using the zkHost based constructor

2012-12-04 Thread Per Steffensen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3425?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509769#comment-13509769
 ] 

Per Steffensen edited comment on SOLR-3425 at 12/4/12 2:58 PM:
---

You talk about using CloudSolrServer and/or LBHttpSolrServer for access to Core 
Admin API. This a little strange because:
* A Core Admin API request should be sent to a specific Solr node in order for 
you to control where the shard is created (or removed from)
* CloudSolrServer and LBHttpSolrServer are all about sending requests that can 
end up going to a (randmon) node among several nodes
With the quick-fix code above (Tommaso Teofili 01/May/12) you will end up 
having your shard created on a random node among all live nodes - it is very 
rarely want you want.

So as long as we are talking about accessing Core Admin API you probably always 
want to use HttpSolrServer, which is aimed a sending the request to a specific 
node.

But when talking about creating an entire collection consisting of many shards, 
it is certainly meaningful to use CloudSolrServer. To create entire collections 
(without having to create each shard yourself using the Core Admin API) we now 
have the Collection API in 4.0.0. The Collection API can be used through 
CloudSolrServer, except...
* You cant create your first collection through CloudSolrServer
* You cant create a collection through a CloudSolrServer, where 
default-collection is the collection you want to create
* etc

...basically because CloudSolrServer wants an existing collection (pointed to 
by its default-collection or a collection-name provided in the actual request) 
before it can do anything.
This will be fixed with SOLR-4140, but is not in 4.0.0.

Other things not in Collection API of 4.0.0
* You cant have more than one shard per collection on a single node - fixed in 
SOLR-4114
* You cant specify which Solr nodes the shards for a new collection is allowed 
to be spread across - they are just spread across all live Solr nodes - fixed 
in SOLR-4120

Please state if you still believe something is missing or unclear. Or else you 
might want to state that your problems are solved with the 4.0.0 Collection 
API (maybe plus one or more of SOLR-4140, SOLR-4114 and SOLR-4120 which will 
hopefully be in the next release) by closing this issue SOLR-3425

  was (Author: steff1193):
You talk about using CloudSolrServer and/or LBHttpSolrServer for access to 
Core Admin API. This a little strange because:
* A Core Admin API request should be sent to a specific Solr node in order for 
you to control where the shard is created (or removed from)
* CloudSolrServer and LBHttpSolrServer are all about sending requests that can 
end up going to a (randmon) node among several nodes
With the quick-fix code above (Tommaso Teofili 01/May/12) you will end up 
having your shard created on a random node among all live nodes - it is very 
rarely want you want.

So as long as we are talking about accessing Core Admin API you probably always 
want to use HttpSolrServer, which is aimed a sending the request to a specific 
node.

But when talking about creating an entire collection consisting of many shards, 
it is certainly meaningful to use CloudSolrServer. To create entire collections 
(without having to create each shard yourself using the Core Admin API) we now 
have the Collection API in 4.0.0. The Collection API can be used through 
CloudSolrServer, except...
* You cant create your first collection through CloudSolrServer
* You cant create a collection through a CloudSolrServer, where 
default-collection is the collection you want to create
* etc

...basically because CloudSolrServer wants an existing collection (pointed to 
by its default-collection or a collection-name provided in the actual request) 
before it can do anything.
This will be fixed with SOLR-4140, but is not in 4.0.0.

Other things not in Collection API of 4.0.0
* You cant have more than one shard per collection on a single node - fixed in 
TLT-4114
* You cant specify which Solr nodes the shards for a new collection is allowed 
to be spread across - they are just spread across all live Solr nodes - fixed 
in TLT-4120

Please state if you still believe something is missing or unclear. Or else you 
might want to state that your problems are solved with the 4.0.0 Collection 
API (maybe plus one or more of SOLR-4140, SOLR-4114 and SOLR-4120 which will 
hopefully be in the next release) by closing this issue SOLR-3425
  
 CloudSolrServer can't create cores when using the zkHost based constructor
 --

 Key: SOLR-3425
 URL: https://issues.apache.org/jira/browse/SOLR-3425
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Reporter: 

[jira] [Updated] (SOLR-4140) Collection API: CloudSolrServer cannot be used to create collection without referencing an existing collection

2012-12-04 Thread Per Steffensen (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4140?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Per Steffensen updated SOLR-4140:
-

Summary: Collection API: CloudSolrServer cannot be used to create 
collection without referencing an existing collection  (was: Collection API: 
CloudSolrServer cannot be used to create collection with referencing an 
existing collection)

 Collection API: CloudSolrServer cannot be used to create collection without 
 referencing an existing collection
 --

 Key: SOLR-4140
 URL: https://issues.apache.org/jira/browse/SOLR-4140
 Project: Solr
  Issue Type: Bug
  Components: multicore, SolrCloud
Affects Versions: 4.0
Reporter: Per Steffensen
Assignee: Per Steffensen
Priority: Minor
  Labels: collection-api, multicore
 Attachments: SOLR-4140_trunk.patch


 CloudSolrServer needs an existing collection to be able to forward requests. 
 For some requests it is not logical that you ned to have or reference an 
 existing collection. E.g. if you want to use CloudSolrServer to create a 
 collection - especially:
 * If you have no collections already and want to create your first collection
 * If you like to create your CloudSolrServer once and for all, and want to 
 set default-collection on it to point to your favorite collection, but also 
 use this CloudSolrServer to create this favorite collection
 Yes, HttpSolrServer can be used, but
 * Some of us like to use CloudSolrServer for everything
 * Using HttpSolrServer you need to check youself which nodes are running, in 
 order to send the create-request to one of them

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-4144) SolrCloud replication high heap consumption

2012-12-04 Thread Markus Jelsma (JIRA)
Markus Jelsma created SOLR-4144:
---

 Summary: SolrCloud replication high heap consumption
 Key: SOLR-4144
 URL: https://issues.apache.org/jira/browse/SOLR-4144
 Project: Solr
  Issue Type: Bug
  Components: replication (java), SolrCloud
Affects Versions: 5.0
 Environment: 5.0-SNAPSHOT 1366361:1416494M - markus - 2012-12-03 
14:09:13
Reporter: Markus Jelsma
Priority: Critical
 Fix For: 5.0


Recent versions of SolrCloud require a very high heap size vs. older versions. 
Another cluster of 5.0.0.2012.10.09.19.29.59 (~4GB per core) can restore an 
empty node without taking a lot of heap (xmx=256m). Recent versions and current 
trunk fail miserably even with a higher heap (750m). Both clusters have 10 
nodes, 10 shards and 2 cores per node. One note to add is that the cluster on 
which this fails has only about 1.5GB per core due to changing in the Lucene 
codec such as compression.

After start up everything goes fine...

{code}
2012-12-04 15:05:35,013 INFO [solr.cloud.RecoveryStrategy] - [RecoveryThread] - 
: Begin buffering updates. core=shard_c
2012-12-04 15:05:35,013 INFO [solr.cloud.RecoveryStrategy] - [RecoveryThread] - 
: Begin buffering updates. core=shard_b
2012-12-04 15:05:35,013 INFO [solr.update.UpdateLog] - [RecoveryThread] - : 
Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
2012-12-04 15:05:35,013 INFO [solr.update.UpdateLog] - [RecoveryThread] - : 
Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
2012-12-04 15:05:35,013 INFO [solr.cloud.RecoveryStrategy] - [RecoveryThread] - 
: Attempting to replicate from http://178.21.118.190:8080/solr/shard_b/. 
core=shard_b
2012-12-04 15:05:35,013 INFO [solr.cloud.RecoveryStrategy] - [RecoveryThread] - 
: Attempting to replicate from http://178.21.118.192:8080/solr/shard_c/. 
core=shard_c
2012-12-04 15:05:35,014 INFO [solrj.impl.HttpClientUtil] - [RecoveryThread] - : 
Creating new http client, 
config:maxConnections=128maxConnectionsPerHost=32followRedirects=false
2012-12-04 15:05:35,014 INFO [solrj.impl.HttpClientUtil] - [RecoveryThread] - : 
Creating new http client, 
config:maxConnections=128maxConnectionsPerHost=32followRedirects=false
2012-12-04 15:05:35,052 INFO [solr.handler.ReplicationHandler] - 
[RecoveryThread] - : Commits will be reserved for  1
2012-12-04 15:05:35,052 INFO [solr.handler.ReplicationHandler] - 
[RecoveryThread] - : Commits will be reserved for  1
2012-12-04 15:05:35,053 INFO [solrj.impl.HttpClientUtil] - [RecoveryThread] - : 
Creating new http client, 
config:connTimeout=5000socketTimeout=2allowCompression=falsemaxConnections=1maxConnectionsPerHost=1
2012-12-04 15:05:35,060 INFO [solr.handler.SnapPuller] - [RecoveryThread] - :  
No value set for 'pollInterval'. Timer Task not started.
2012-12-04 15:05:35,060 INFO [solr.handler.SnapPuller] - [RecoveryThread] - :  
No value set for 'pollInterval'. Timer Task not started.
2012-12-04 15:05:35,069 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
Master's generation: 48
2012-12-04 15:05:35,069 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
Slave's generation: 1
2012-12-04 15:05:35,069 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
Starting replication process
2012-12-04 15:05:35,069 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
Master's generation: 47
2012-12-04 15:05:35,069 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
Slave's generation: 1
2012-12-04 15:05:35,070 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
Starting replication process
2012-12-04 15:05:35,078 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
Number of files in latest index in master: 235
2012-12-04 15:05:35,079 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
Number of files in latest index in master: 287
2012-12-04 15:05:35,084 WARN [solr.core.CachingDirectoryFactory] - 
[RecoveryThread] - : No lockType configured for 
NRTCachingDirectory(org.apache.lucene.store.MMapDirectory@/opt/solr/cores/shard_c/data/index.20121204150535080
 lockFactory=org.apache.lucene.store.NativeFSLockFactory@57530551; 
maxCacheMB=48.0 maxMergeSizeMB=4.0) assuming 'simple'
2012-12-04 15:05:35,085 INFO [solr.core.CachingDirectoryFactory] - 
[RecoveryThread] - : return new directory for 
/opt/solr/cores/shard_c/data/index.20121204150535080 forceNew:false
2012-12-04 15:05:35,085 INFO [solr.core.CachingDirectoryFactory] - 
[RecoveryThread] - : Releasing directory:/opt/solr/cores/shard_c/data
2012-12-04 15:05:35,085 WARN [solr.core.CachingDirectoryFactory] - 
[RecoveryThread] - : No lockType configured for 
NRTCachingDirectory(org.apache.lucene.store.MMapDirectory@/opt/solr/cores/shard_b/data/index.20121204150535079
 lockFactory=org.apache.lucene.store.NativeFSLockFactory@512fb063; 
maxCacheMB=48.0 maxMergeSizeMB=4.0) assuming 'simple'
2012-12-04 15:05:35,085 INFO 

[jira] [Commented] (SOLR-3425) CloudSolrServer can't create cores when using the zkHost based constructor

2012-12-04 Thread Tommaso Teofili (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3425?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509793#comment-13509793
 ] 

Tommaso Teofili commented on SOLR-3425:
---

this issue was opened before the Collections API was designed, so basically 
things have changed along time but apart from that as long as there's SOLR-4140 
the use case I opened this for is safe, therefore I'll close it.
Thanks [~steff1193] for helping out.

 CloudSolrServer can't create cores when using the zkHost based constructor
 --

 Key: SOLR-3425
 URL: https://issues.apache.org/jira/browse/SOLR-3425
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Reporter: Tommaso Teofili
Assignee: Mark Miller
Priority: Minor
 Fix For: 4.1

 Attachments: SOLR-3425-test.patch


 When programmatically creating cores with a running SolrCloud instance the 
 CloudSolrServer uses the slices nodes information to feed the underlying 
 LBHttpSolrServer so it fails to create cores as there aren't any slices for 
 any new collection (as it's still to be created).
 This happens when using the CloudSolrServer constructor which takes the ZK 
 host as only parameter while it can be avoided by using the constructor which 
 also takes the list of Solr URLs and the underlying LBHttpSolrServer is 
 actually used for making the core creation request.
 However it'd be good to use the ZK host live nodes information to 
 automatically issue a core creation command on one of the underlying Solr 
 hosts without having to specify the full list of URLs beforehand.
 The scenario is when one wants to create a collection with N shards so the 
 client sends N core creation requests for the same collection thus the 
 SolrCloud stuff should just take care of choosing the host where to issue the 
 core creation request and update the cluster state.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-3425) CloudSolrServer can't create cores when using the zkHost based constructor

2012-12-04 Thread Tommaso Teofili (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3425?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tommaso Teofili resolved SOLR-3425.
---

Resolution: Fixed

 CloudSolrServer can't create cores when using the zkHost based constructor
 --

 Key: SOLR-3425
 URL: https://issues.apache.org/jira/browse/SOLR-3425
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Reporter: Tommaso Teofili
Assignee: Mark Miller
Priority: Minor
 Fix For: 4.1

 Attachments: SOLR-3425-test.patch


 When programmatically creating cores with a running SolrCloud instance the 
 CloudSolrServer uses the slices nodes information to feed the underlying 
 LBHttpSolrServer so it fails to create cores as there aren't any slices for 
 any new collection (as it's still to be created).
 This happens when using the CloudSolrServer constructor which takes the ZK 
 host as only parameter while it can be avoided by using the constructor which 
 also takes the list of Solr URLs and the underlying LBHttpSolrServer is 
 actually used for making the core creation request.
 However it'd be good to use the ZK host live nodes information to 
 automatically issue a core creation command on one of the underlying Solr 
 hosts without having to specify the full list of URLs beforehand.
 The scenario is when one wants to create a collection with N shards so the 
 client sends N core creation requests for the same collection thus the 
 SolrCloud stuff should just take care of choosing the host where to issue the 
 core creation request and update the cluster state.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Solrj question

2012-12-04 Thread Per Steffensen




Believe there are no setParser method on CloudSolrServer? 

Have you tried this:
XMLWriter.writeResponse(a java.io.Writer for your XML output,
your SolrQueryRequest, your SolrQueryResponse)
It will require you to include solr-core, since that is where XMLWriter
lives.

But it is a little inefficient, if I understand you correctly. If you
have a CloudSolrServer doing a request and asks for a response in e.g.
XML format, then this CloudSolrServer will receive a XML response over
the wire, parse it into a SolrQueryResponse object and then you will
take that SolrQueryResponse and serialize it back to XML. It would be
nicer if you just used whatever was received on the wire directly.

Regards, Per Steffensen

Erick Erickson skrev:
haven't tried this personally, but try this:
  
  server.setParser(new XMLResponseParser());
  
  
  Best
  Erick
  
  
  
  On Mon, Dec 3, 2012 at 10:24 AM, Steve
Molloy smol...@opentext.com
wrote:
  


I’ve been digging around a bit and could not
find any straight-forward way of getting back the XML response after
calling Solr using Solrj. The query method returns a QueryResponse
object, which is great in most cases. But our use case implies pushing
the response back to whatever format was requested, but still use the
CloudSolrServer client. This is very easy with the javabin format,
using the JavaBinCodec, but is there equivalent utilities to marshal
into other formats from a NamedList? Otherwise, could the functionality
be isolated for better reuse?
 
Thanks,
 

  

  
  Steve
Molloy
  Content
Analytics Lead  |  RD
  


  
  

  

Phone:


(514) 908-5406 406

  
  

Website:


www.opentext.com

  

  
  

  



Nstein is
now OpenText.
Please update your address book with my new email address.


This
email message is confidential, may be privileged, and is intended for
the exclusive use of the addressee. Any other person is strictly
prohibited from disclosing or reproducing it. If the addressee cannot
be reached or is unknown to you, please inform the sender by return
email and delete this email message and all copies immediately.
 


  
  
  
  






[jira] [Updated] (SOLR-3583) Percentiles for facets, pivot facets, and distributed pivot facets

2012-12-04 Thread Chris Russell (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3583?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Russell updated SOLR-3583:


Attachment: SOLR-3583.patch

Now works with facet.missing=true

 Percentiles for facets, pivot facets, and distributed pivot facets
 --

 Key: SOLR-3583
 URL: https://issues.apache.org/jira/browse/SOLR-3583
 Project: Solr
  Issue Type: Improvement
Reporter: Chris Russell
Priority: Minor
  Labels: newbie, patch
 Fix For: 4.1

 Attachments: SOLR-3583.patch, SOLR-3583.patch, SOLR-3583.patch


 Built on top of SOLR-2894, this patch adds percentiles and averages to 
 facets, pivot facets, and distributed pivot facets by making use of range 
 facet internals.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4586) Change default ResultMode of FacetRequest to PER_NODE_IN_TREE

2012-12-04 Thread Shai Erera (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4586?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shai Erera updated LUCENE-4586:
---

Attachment: LUCENE-4586.patch

Patch changes the default ResultMode to PER_NODE_IN_TREE. That uncovered a 
potential bug in TopKInEachNodeHandler where it didn't resolve ties on facet 
values by ordinal -- this lead to inconsistent ordering of facet results 
(discovered by SamplingWrapperTest).

I had to fix some tests that relied on GLOBAL_FLAT mode for results 
verification (it's easier to verify results in that mode).

Added a CHANGES entry under back-compat changes.

All tests pass (tried many iterations).

 Change default ResultMode of FacetRequest to PER_NODE_IN_TREE
 -

 Key: LUCENE-4586
 URL: https://issues.apache.org/jira/browse/LUCENE-4586
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/facet
Reporter: Shai Erera
Assignee: Shai Erera
Priority: Minor
 Fix For: 4.1, 5.0

 Attachments: LUCENE-4586.patch


 Today the default ResultMode is GLOBAL_FLAT, but it should be 
 PER_NODE_IN_TREE. ResultMode is being used whenever you set the depth of 
 FacetRequest to greater than 1. The difference between the two is:
 * PER_NODE_IN_TREE would then compute the top-K categories recursively, for 
 every top category at every level (up to depth). The results are returned in 
 a tree structure as well. For instance:
 {noformat}
 Date
   2010
 March
 February
   2011
 April
 May
 {noformat}
 * GLOBAL_FLAT computes the top categories among all the nodes up to depth, 
 and returns a flat list of categories.
 GLOBAL_FLAT is faster to compute than PER_NODE_IN_TREE (it just computes 
 top-K among N total categories), however I think that it's less intuitive, 
 and therefore should not be used as a default. In fact, I think this is kind 
 of an expert usage.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4586) Change default ResultMode of FacetRequest to PER_NODE_IN_TREE

2012-12-04 Thread Shai Erera (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4586?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shai Erera updated LUCENE-4586:
---

Attachment: LUCENE-4586.patch

Same patch, but moved the CHANGES entry under a Changes in Runtime Behavior 
section

 Change default ResultMode of FacetRequest to PER_NODE_IN_TREE
 -

 Key: LUCENE-4586
 URL: https://issues.apache.org/jira/browse/LUCENE-4586
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/facet
Reporter: Shai Erera
Assignee: Shai Erera
Priority: Minor
 Fix For: 4.1, 5.0

 Attachments: LUCENE-4586.patch, LUCENE-4586.patch


 Today the default ResultMode is GLOBAL_FLAT, but it should be 
 PER_NODE_IN_TREE. ResultMode is being used whenever you set the depth of 
 FacetRequest to greater than 1. The difference between the two is:
 * PER_NODE_IN_TREE would then compute the top-K categories recursively, for 
 every top category at every level (up to depth). The results are returned in 
 a tree structure as well. For instance:
 {noformat}
 Date
   2010
 March
 February
   2011
 April
 May
 {noformat}
 * GLOBAL_FLAT computes the top categories among all the nodes up to depth, 
 and returns a flat list of categories.
 GLOBAL_FLAT is faster to compute than PER_NODE_IN_TREE (it just computes 
 top-K among N total categories), however I think that it's less intuitive, 
 and therefore should not be used as a default. In fact, I think this is kind 
 of an expert usage.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3030) StopTermTypesFilter and Factory allows filtering based on the TermTypeAttribute

2012-12-04 Thread John Wooden (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3030?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

John Wooden updated SOLR-3030:
--

Attachment: SOLR-3030 4.0.patch

Updated this patch for 4.0.0

 StopTermTypesFilter and Factory allows filtering based on the 
 TermTypeAttribute
 ---

 Key: SOLR-3030
 URL: https://issues.apache.org/jira/browse/SOLR-3030
 Project: Solr
  Issue Type: New Feature
  Components: Schema and Analysis
Affects Versions: 4.0-ALPHA
Reporter: Monica Skidmore
Priority: Trivial
  Labels: features, newbie, patch
 Attachments: SOLR-3030 4.0.patch, SOLR-3030.patch

   Original Estimate: 168h
  Remaining Estimate: 168h

 This filter will remove terms based on the TermTypeAttribute, using a list of 
 StopTermTypes from the StopTermTypes file specified by the user as an 
 attribute of the filter factory in their schema.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3030) StopTermTypesFilter and Factory allows filtering based on the TermTypeAttribute

2012-12-04 Thread Steven Rowe (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3030?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509828#comment-13509828
 ] 

Steven Rowe commented on SOLR-3030:
---

@John Wooden: as Tommaso Teofili mentioned in the previous post, the 
functionality provided by this issue already exists in Lucene/Solr: 
TypeTokenFilter(Factory).  In fact, TypeTokenFilter also can handle whitelisted 
types in addition to the blacklisting functionality provided here.  Is there 
something here that TypeTokenFilter can't do?

AFAICT, this issue should be closed as Won't Fix.

 StopTermTypesFilter and Factory allows filtering based on the 
 TermTypeAttribute
 ---

 Key: SOLR-3030
 URL: https://issues.apache.org/jira/browse/SOLR-3030
 Project: Solr
  Issue Type: New Feature
  Components: Schema and Analysis
Affects Versions: 4.0-ALPHA
Reporter: Monica Skidmore
Priority: Trivial
  Labels: features, newbie, patch
 Attachments: SOLR-3030 4.0.patch, SOLR-3030.patch

   Original Estimate: 168h
  Remaining Estimate: 168h

 This filter will remove terms based on the TermTypeAttribute, using a list of 
 StopTermTypes from the StopTermTypes file specified by the user as an 
 attribute of the filter factory in their schema.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4586) Change default ResultMode of FacetRequest to PER_NODE_IN_TREE

2012-12-04 Thread Gilad Barkai (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4586?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509847#comment-13509847
 ] 

Gilad Barkai commented on LUCENE-4586:
--

Patch looks good, ready for a commit.
+1

 Change default ResultMode of FacetRequest to PER_NODE_IN_TREE
 -

 Key: LUCENE-4586
 URL: https://issues.apache.org/jira/browse/LUCENE-4586
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/facet
Reporter: Shai Erera
Assignee: Shai Erera
Priority: Minor
 Fix For: 4.1, 5.0

 Attachments: LUCENE-4586.patch, LUCENE-4586.patch


 Today the default ResultMode is GLOBAL_FLAT, but it should be 
 PER_NODE_IN_TREE. ResultMode is being used whenever you set the depth of 
 FacetRequest to greater than 1. The difference between the two is:
 * PER_NODE_IN_TREE would then compute the top-K categories recursively, for 
 every top category at every level (up to depth). The results are returned in 
 a tree structure as well. For instance:
 {noformat}
 Date
   2010
 March
 February
   2011
 April
 May
 {noformat}
 * GLOBAL_FLAT computes the top categories among all the nodes up to depth, 
 and returns a flat list of categories.
 GLOBAL_FLAT is faster to compute than PER_NODE_IN_TREE (it just computes 
 top-K among N total categories), however I think that it's less intuitive, 
 and therefore should not be used as a default. In fact, I think this is kind 
 of an expert usage.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4586) Change default ResultMode of FacetRequest to PER_NODE_IN_TREE

2012-12-04 Thread Shai Erera (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4586?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509852#comment-13509852
 ] 

Shai Erera commented on LUCENE-4586:


Thanks. I'll give tests a couple more runs on my box before I commit.

 Change default ResultMode of FacetRequest to PER_NODE_IN_TREE
 -

 Key: LUCENE-4586
 URL: https://issues.apache.org/jira/browse/LUCENE-4586
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/facet
Reporter: Shai Erera
Assignee: Shai Erera
Priority: Minor
 Fix For: 4.1, 5.0

 Attachments: LUCENE-4586.patch, LUCENE-4586.patch


 Today the default ResultMode is GLOBAL_FLAT, but it should be 
 PER_NODE_IN_TREE. ResultMode is being used whenever you set the depth of 
 FacetRequest to greater than 1. The difference between the two is:
 * PER_NODE_IN_TREE would then compute the top-K categories recursively, for 
 every top category at every level (up to depth). The results are returned in 
 a tree structure as well. For instance:
 {noformat}
 Date
   2010
 March
 February
   2011
 April
 May
 {noformat}
 * GLOBAL_FLAT computes the top categories among all the nodes up to depth, 
 and returns a flat list of categories.
 GLOBAL_FLAT is faster to compute than PER_NODE_IN_TREE (it just computes 
 top-K among N total categories), however I think that it's less intuitive, 
 and therefore should not be used as a default. In fact, I think this is kind 
 of an expert usage.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4585) Spatial RecursivePrefixTreeFilter has some bugs with indexing non-point shapes

2012-12-04 Thread David Smiley (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4585?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Smiley updated LUCENE-4585:
-

Attachment: LUCENE-4585_PrefixTree_bugs.patch

The attached patch resolves the aforementioned problems.

However note that when a point is indexed, the final full-length token is not 
indexed with a trailing leaf '+' byte variant; it was before.  The up-side is 
that we save one token per indexed point (~1/12th savings if 11 maxLevels).  
The semantics of that '+' are intended to be that the entire grid cell 
represents an indexed shape for matching, so compare the rectangle for it with 
the query shape.  But for points, it should be the center point of the cell, so 
no '+'.

If a user doesn't re-index, then an indexed point is 1/2 the smallest grid cell 
closer to a query shape and as such might match when it didn't before.  Quite 
minor I think but worth mentioning.

 Spatial RecursivePrefixTreeFilter has some bugs with indexing non-point shapes
 --

 Key: LUCENE-4585
 URL: https://issues.apache.org/jira/browse/LUCENE-4585
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/spatial
Affects Versions: 4.0
Reporter: David Smiley
 Attachments: LUCENE-4585_PrefixTree_bugs.patch


 RecursivePrefixTreeFilter has some bugs that can occur when searching indexed 
 shapes.  One bug is an unpositioned termsEnum.  It through an exception in 
 testing; I'm not sure what its effects would be in production.  The other 
 couple bugs are hard to describe here but were rare to occur in extensive 
 testing. The effects were probably a slim chance of matching an indexed shape 
 near the query shape. And SpatialPrefixTree does not support an indexed shape 
 that covers the entire globe.
 These bugs were discovered during development of tests for RPTF LUCENE-4419 
 which I will submit shortly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (LUCENE-4585) Spatial RecursivePrefixTreeFilter has some bugs with indexing non-point shapes

2012-12-04 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4585?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509854#comment-13509854
 ] 

David Smiley edited comment on LUCENE-4585 at 12/4/12 4:58 PM:
---

The attached patch resolves the aforementioned problems.

However note that when a point is indexed, the final full-length token is not 
indexed with a trailing leaf '\+' byte variant; it was before.  The up-side is 
that we save one token per indexed point (~1/12th savings if 11 maxLevels).  
The semantics of that '\+' are intended to be that the entire grid cell 
represents an indexed shape for matching, so compare the rectangle for it with 
the query shape.  But for points, it should be the center point of the cell, so 
no '+'.

If a user doesn't re-index, then an indexed point is 1/2 the smallest grid cell 
closer to a query shape and as such might match when it didn't before.  Quite 
minor I think but worth mentioning.

  was (Author: dsmiley):
The attached patch resolves the aforementioned problems.

However note that when a point is indexed, the final full-length token is not 
indexed with a trailing leaf '+' byte variant; it was before.  The up-side is 
that we save one token per indexed point (~1/12th savings if 11 maxLevels).  
The semantics of that '+' are intended to be that the entire grid cell 
represents an indexed shape for matching, so compare the rectangle for it with 
the query shape.  But for points, it should be the center point of the cell, so 
no '+'.

If a user doesn't re-index, then an indexed point is 1/2 the smallest grid cell 
closer to a query shape and as such might match when it didn't before.  Quite 
minor I think but worth mentioning.
  
 Spatial RecursivePrefixTreeFilter has some bugs with indexing non-point shapes
 --

 Key: LUCENE-4585
 URL: https://issues.apache.org/jira/browse/LUCENE-4585
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/spatial
Affects Versions: 4.0
Reporter: David Smiley
 Attachments: LUCENE-4585_PrefixTree_bugs.patch


 RecursivePrefixTreeFilter has some bugs that can occur when searching indexed 
 shapes.  One bug is an unpositioned termsEnum.  It through an exception in 
 testing; I'm not sure what its effects would be in production.  The other 
 couple bugs are hard to describe here but were rare to occur in extensive 
 testing. The effects were probably a slim chance of matching an indexed shape 
 near the query shape. And SpatialPrefixTree does not support an indexed shape 
 that covers the entire globe.
 These bugs were discovered during development of tests for RPTF LUCENE-4419 
 which I will submit shortly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4583) StraightBytesDocValuesField fails if bytes 32k

2012-12-04 Thread Barakat Barakat (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4583?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Barakat Barakat updated LUCENE-4583:


Attachment: LUCENE-4583.patch

I updated the code to work when start isn't zero. The code can still crash if 
you ask for a length that goes beyond the total size of the paged bytes, but 
I'm not sure how you guys like to prevent things like that. The code seems to 
be working fine with our Solr core so far. I am new to posting patches and 
writing test units in Java so please let me know if there is anything wrong 
with the code.

 StraightBytesDocValuesField fails if bytes  32k
 

 Key: LUCENE-4583
 URL: https://issues.apache.org/jira/browse/LUCENE-4583
 Project: Lucene - Core
  Issue Type: Bug
  Components: core/index
Affects Versions: 4.0, 4.1, 5.0
Reporter: David Smiley
Priority: Critical
 Fix For: 4.1

 Attachments: LUCENE-4583.patch


 I didn't observe any limitations on the size of a bytes based DocValues field 
 value in the docs.  It appears that the limit is 32k, although I didn't get 
 any friendly error telling me that was the limit.  32k is kind of small IMO; 
 I suspect this limit is unintended and as such is a bug.The following 
 test fails:
 {code:java}
   public void testBigDocValue() throws IOException {
 Directory dir = newDirectory();
 IndexWriter writer = new IndexWriter(dir, writerConfig(false));
 Document doc = new Document();
 BytesRef bytes = new BytesRef((4+4)*4097);//4096 works
 bytes.length = bytes.bytes.length;//byte data doesn't matter
 doc.add(new StraightBytesDocValuesField(dvField, bytes));
 writer.addDocument(doc);
 writer.commit();
 writer.close();
 DirectoryReader reader = DirectoryReader.open(dir);
 DocValues docValues = MultiDocValues.getDocValues(reader, dvField);
 //FAILS IF BYTES IS BIG!
 docValues.getSource().getBytes(0, bytes);
 reader.close();
 dir.close();
   }
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-1575) loading solr.xml takes a lot of time when there are lot of cores

2012-12-04 Thread olivier soyez (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1575?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509857#comment-13509857
 ] 

olivier soyez commented on SOLR-1575:
-

We have around 50K cores with loadOnStart=false and same issue to start the 
server.
It takes a very long time to start, and it is just the time taken for loading 
the config file listing the available solr cores : solr.xml.

To solve this, in the load function, I just replace the line 532 of 
CoreContainer.java (Apache Solr 4.0 release) from 
p.setCoreProperties(readProperties(cfg, node)); to 
p.setCoreProperties(readProperties(cfg, 
node.getParentNode().removeChild(node)));

Now, it takes a few seconds to start the server


 loading solr.xml takes a lot of time when there are lot of cores
 

 Key: SOLR-1575
 URL: https://issues.apache.org/jira/browse/SOLR-1575
 Project: Solr
  Issue Type: Improvement
  Components: multicore
Reporter: Noble Paul
Priority: Minor
 Fix For: 1.5


 We have around 38K cores which takes up around 25 mins to start the server. 
 Keep in mind that no cores are started and this is the time taken for loading 
 the solr.xml.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4114) Collection API: Allow multiple shards from one collection on the same Solr server

2012-12-04 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509861#comment-13509861
 ] 

Mark Miller commented on SOLR-4114:
---

Yeah, I minimized the patch down to just dealing with this issue. I'm away from 
home and just looking to get this issue completed with minimum fuss. 
'nodeUrlWithoutProtocolPart' is just so long and I didn't see it helping much 
in terms of code readability - if you have a better fitting name that is a 
little less verbose, I think I'd be more into it.

bq. I see that you removed the auto-reduce replica-per-shard t

Yeah, I don't think I agree with changing the users params on him - I'd rather 
warn and let the user do what he wants to do rather than trying to outthink him 
ahead of time. If he decides he wants more than one repica on an instance for 
some reason, that's his deal - we can warn him though.

bq. You removed the following lines (because you just want default-values for 
instance-dir and data-dir)

Right - it should match collection1 - eg newcollection/data should be the data 
dir just like collection1/data and rather than something like 
newcollection_data. In my experience and testing, data ends up in the cores own 
instance dir - not some common dir.

bq. making shardToUrls-map (renamed) concurrency protected 

Yup - it seemed unrelated and I'm busy so I didn't want to think about it. My 
goal is to get the essence of this thing committed - it's a lot easier to then 
fight over smaller pieces on top of that. Progress not perfection. 

 Collection API: Allow multiple shards from one collection on the same Solr 
 server
 -

 Key: SOLR-4114
 URL: https://issues.apache.org/jira/browse/SOLR-4114
 Project: Solr
  Issue Type: New Feature
  Components: multicore, SolrCloud
Affects Versions: 4.0
 Environment: Solr 4.0.0 release
Reporter: Per Steffensen
Assignee: Per Steffensen
  Labels: collection-api, multicore, shard, shard-allocation
 Attachments: SOLR-4114.patch, SOLR-4114.patch, SOLR-4114.patch, 
 SOLR-4114.patch, SOLR-4114_trunk.patch


 We should support running multiple shards from one collection on the same 
 Solr server - the run a collection with 8 shards on a 4 Solr server cluster 
 (each Solr server running 2 shards).
 Performance tests at our side has shown that this is a good idea, and it is 
 also a good idea for easy elasticity later on - it is much easier to move an 
 entire existing shards from one Solr server to another one that just joined 
 the cluter than it is to split an exsiting shard among the Solr that used to 
 run it and the new Solr.
 See dev mailing list discussion Multiple shards for one collection on the 
 same Solr server

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4114) Collection API: Allow multiple shards from one collection on the same Solr server

2012-12-04 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509864#comment-13509864
 ] 

Mark Miller commented on SOLR-4114:
---

bq. In checkForCollection we are willing to wait up to 120 secs (used to be 60 
secs) for the collection to be created (correctly). So it is kinda obvious to 
also wait at least 60 secs to verify that nothing related to a collection was 
actually created in cases where you want to be sure that nothing is created. 

We poll so that we wait up to 120 seconds for a slow comp, but a fast comp 
won't need to wait nearly that long. The 60 wait hits everyone no matter what. 
We generally try avoid any hard waits like that. I understand you can't really 
poll in this case, so I'm not sure the best way to test that - I made it a TODO 
for now.

 Collection API: Allow multiple shards from one collection on the same Solr 
 server
 -

 Key: SOLR-4114
 URL: https://issues.apache.org/jira/browse/SOLR-4114
 Project: Solr
  Issue Type: New Feature
  Components: multicore, SolrCloud
Affects Versions: 4.0
 Environment: Solr 4.0.0 release
Reporter: Per Steffensen
Assignee: Per Steffensen
  Labels: collection-api, multicore, shard, shard-allocation
 Attachments: SOLR-4114.patch, SOLR-4114.patch, SOLR-4114.patch, 
 SOLR-4114.patch, SOLR-4114_trunk.patch


 We should support running multiple shards from one collection on the same 
 Solr server - the run a collection with 8 shards on a 4 Solr server cluster 
 (each Solr server running 2 shards).
 Performance tests at our side has shown that this is a good idea, and it is 
 also a good idea for easy elasticity later on - it is much easier to move an 
 entire existing shards from one Solr server to another one that just joined 
 the cluter than it is to split an exsiting shard among the Solr that used to 
 run it and the new Solr.
 See dev mailing list discussion Multiple shards for one collection on the 
 same Solr server

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4114) Collection API: Allow multiple shards from one collection on the same Solr server

2012-12-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509882#comment-13509882
 ] 

Commit Tag Bot commented on SOLR-4114:
--

[trunk commit] Mark Robert Miller
http://svn.apache.org/viewvc?view=revisionrevision=1417045

SOLR-4114: Allow creating more than one shard per instance with the Collection 
API.



 Collection API: Allow multiple shards from one collection on the same Solr 
 server
 -

 Key: SOLR-4114
 URL: https://issues.apache.org/jira/browse/SOLR-4114
 Project: Solr
  Issue Type: New Feature
  Components: multicore, SolrCloud
Affects Versions: 4.0
 Environment: Solr 4.0.0 release
Reporter: Per Steffensen
Assignee: Per Steffensen
  Labels: collection-api, multicore, shard, shard-allocation
 Attachments: SOLR-4114.patch, SOLR-4114.patch, SOLR-4114.patch, 
 SOLR-4114.patch, SOLR-4114_trunk.patch


 We should support running multiple shards from one collection on the same 
 Solr server - the run a collection with 8 shards on a 4 Solr server cluster 
 (each Solr server running 2 shards).
 Performance tests at our side has shown that this is a good idea, and it is 
 also a good idea for easy elasticity later on - it is much easier to move an 
 entire existing shards from one Solr server to another one that just joined 
 the cluter than it is to split an exsiting shard among the Solr that used to 
 run it and the new Solr.
 See dev mailing list discussion Multiple shards for one collection on the 
 same Solr server

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Odd pattern of segment creation - branch_4x bug?

2012-12-04 Thread Shawn Heisey

On 12/3/2012 9:46 PM, Yonik Seeley wrote:

On Mon, Dec 3, 2012 at 11:36 PM, Shawn Heisey s...@elyograg.org wrote:

updateHandler class=solr.DirectUpdateHandler2
   autoCommit
 maxDocs65536/maxDocs
 maxTime30/maxTime
   /autoCommit
   updateLog /
/updateHandler

Yeah, seems like that just should generate hard commits.
Are you using delete by querys at all?  The current implementation can
do an NRT reopen as part of it's implementation.


When I do deletes, it's done exclusively using deleteByQuery in SolrJ, 
with something like did:(1000 OR 1001 OR 1002 OR 1003) as the query.  
Before doing the delete, I do a search using the same query and only 
proceed to do the delete on that shard if numFound from the query is 
more than zero.


I don't know if this is necessary information, but no deletes are 
happening while I do my full-import.  Nothing else happens to the index 
at all while DIH is doing its thing.


This is new experimentation with autocommit.  By using autocommit, the 
commit at the end of an import is much faster. When autocommit is turned 
off, it is several minutes between the last document add and the handler 
returning to idle status.  I like this new faster commit, but if I 
can't get an autocommit to only produce one segment, I may go back to 
the old way, to reduce the I/O impact from the more frequent merges.


Thanks,
Shawn


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4114) Collection API: Allow multiple shards from one collection on the same Solr server

2012-12-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509910#comment-13509910
 ] 

Commit Tag Bot commented on SOLR-4114:
--

[branch_4x commit] Mark Robert Miller
http://svn.apache.org/viewvc?view=revisionrevision=1417070

SOLR-4114: Allow creating more than one shard per instance with the Collection 
API.



 Collection API: Allow multiple shards from one collection on the same Solr 
 server
 -

 Key: SOLR-4114
 URL: https://issues.apache.org/jira/browse/SOLR-4114
 Project: Solr
  Issue Type: New Feature
  Components: multicore, SolrCloud
Affects Versions: 4.0
 Environment: Solr 4.0.0 release
Reporter: Per Steffensen
Assignee: Per Steffensen
  Labels: collection-api, multicore, shard, shard-allocation
 Attachments: SOLR-4114.patch, SOLR-4114.patch, SOLR-4114.patch, 
 SOLR-4114.patch, SOLR-4114_trunk.patch


 We should support running multiple shards from one collection on the same 
 Solr server - the run a collection with 8 shards on a 4 Solr server cluster 
 (each Solr server running 2 shards).
 Performance tests at our side has shown that this is a good idea, and it is 
 also a good idea for easy elasticity later on - it is much easier to move an 
 entire existing shards from one Solr server to another one that just joined 
 the cluter than it is to split an exsiting shard among the Solr that used to 
 run it and the new Solr.
 See dev mailing list discussion Multiple shards for one collection on the 
 same Solr server

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4585) Spatial RecursivePrefixTreeFilter has some bugs with indexing non-point shapes

2012-12-04 Thread David Smiley (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4585?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Smiley updated LUCENE-4585:
-

Attachment: LUCENE-4585_PrefixTree_bugs.patch

Updated patch with the correct set of files.

 Spatial RecursivePrefixTreeFilter has some bugs with indexing non-point shapes
 --

 Key: LUCENE-4585
 URL: https://issues.apache.org/jira/browse/LUCENE-4585
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/spatial
Affects Versions: 4.0
Reporter: David Smiley
 Attachments: LUCENE-4585_PrefixTree_bugs.patch, 
 LUCENE-4585_PrefixTree_bugs.patch


 RecursivePrefixTreeFilter has some bugs that can occur when searching indexed 
 shapes.  One bug is an unpositioned termsEnum.  It through an exception in 
 testing; I'm not sure what its effects would be in production.  The other 
 couple bugs are hard to describe here but were rare to occur in extensive 
 testing. The effects were probably a slim chance of matching an indexed shape 
 near the query shape. And SpatialPrefixTree does not support an indexed shape 
 that covers the entire globe.
 These bugs were discovered during development of tests for RPTF LUCENE-4419 
 which I will submit shortly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org




[JENKINS] Lucene-Solr-trunk-Windows (32bit/jdk1.7.0_09) - Build # 2052 - Failure!

2012-12-04 Thread Policeman Jenkins Server
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-trunk-Windows/2052/
Java: 32bit/jdk1.7.0_09 -server -XX:+UseG1GC

All tests passed

Build Log:
[...truncated 13648 lines...]
BUILD FAILED
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\build.xml:88: The 
following files contain @author tags, tabs or nocommits:
* solr/core/src/test/org/apache/solr/cloud/BasicDistributedZkTest.java

Total time: 52 minutes 34 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 32bit/jdk1.7.0_09 -server -XX:+UseG1GC
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[JENKINS] Lucene-Solr-trunk-Linux (64bit/jdk1.6.0_37) - Build # 3061 - Failure!

2012-12-04 Thread Policeman Jenkins Server
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-trunk-Linux/3061/
Java: 64bit/jdk1.6.0_37 -XX:+UseParallelGC

All tests passed

Build Log:
[...truncated 12949 lines...]
BUILD FAILED
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:88: The following 
files contain @author tags, tabs or nocommits:
* solr/core/src/test/org/apache/solr/cloud/BasicDistributedZkTest.java

Total time: 27 minutes 56 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 64bit/jdk1.6.0_37 -XX:+UseParallelGC
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

RE: Solrj question

2012-12-04 Thread Steve Molloy
Thanks, for some reason I got stopped with the fact that XMLWriter wanted 
SolrQueryRequest/Response instead of the NamedList available in solrj's 
QueryResponse and never looked at building those from the info I already had. 
It is working now.

For the comment on efficiency, I do have to agree that parsing to Java object 
and then serializing back is less than optimal to say the least. The reason it 
has to be done in our case is because the component acting as proxy also 
supports extra parameters for some of our business logic that will modify the 
incoming query in some cases, decide where it goes in others, etc. So to avoid 
rewriting the logic inside of CloudSolrServer, I want to reuse it (I really 
hate duplicating code). But the output is a Java object (QueryResponse), so I 
have to serialize it back. Unless there's a way to get the original message 
before it's unmarshaled into the QueryResponse, in which case I'd be more than 
happy to use that as-is. :)

Anyways, as I mentioned, it seems to be working for now. And I do use javabin 
between that component and Solr no matter what was asked in the request since I 
can then marshal it to anything.

Steve Molloy
Content Analytics Lead  |  RD

Phone:

(514) 908-5406 406

Website:

www.opentext.comhttp://www.opentext.com/


[http://www.opentext.com/2/emailsupport-logo-opentext-2010.gif]http://www.opentext.com/2/email-signature-logo




http://www.opentext.com/2/email-signature-event

We can't solve problems by using the same kind of thinking we used when we 
created them.
Attributed to Albert Einstein

From: Per Steffensen [mailto:st...@designware.dk]
Sent: Tuesday, December 04, 2012 10:24 AM
To: dev@lucene.apache.org
Subject: Re: Solrj question

Believe there are no setParser method on CloudSolrServer?

Have you tried this:
XMLWriter.writeResponse(a java.io.Writer for your XML output, your 
SolrQueryRequest, your SolrQueryResponse)
It will require you to include solr-core, since that is where XMLWriter lives.

But it is a little inefficient, if I understand you correctly. If you have a 
CloudSolrServer doing a request and asks for a response in e.g. XML format, 
then this CloudSolrServer will receive a XML response over the wire, parse it 
into a SolrQueryResponse object and then you will take that SolrQueryResponse 
and serialize it back to XML. It would be nicer if you just used whatever was 
received on the wire directly.

Regards, Per Steffensen

Erick Erickson skrev:
haven't tried this personally, but try this:

server.setParser(new XMLResponseParser());





Best

Erick

On Mon, Dec 3, 2012 at 10:24 AM, Steve Molloy 
smol...@opentext.commailto:smol...@opentext.com wrote:
I've been digging around a bit and could not find any straight-forward way of 
getting back the XML response after calling Solr using Solrj. The query method 
returns a QueryResponse object, which is great in most cases. But our use case 
implies pushing the response back to whatever format was requested, but still 
use the CloudSolrServer client. This is very easy with the javabin format, 
using the JavaBinCodec, but is there equivalent utilities to marshal into other 
formats from a NamedList? Otherwise, could the functionality be isolated for 
better reuse?

Thanks,

Steve Molloy
Content Analytics Lead  |  RD

Phone:

(514) 908-5406 406tel:908-5406%20406

Website:

www.opentext.comhttp://www.opentext.com/


[http://www.opentext.com/2/emailsupport-logo-opentext-2010.gif]http://www.opentext.com/2/email-signature-logo

Nstein is now OpenText.
Please update your address book with my new email address.

This email message is confidential, may be privileged, and is intended for the 
exclusive use of the addressee. Any other person is strictly prohibited from 
disclosing or reproducing it. If the addressee cannot be reached or is unknown 
to you, please inform the sender by return email and delete this email message 
and all copies immediately.



inline: image001.gifinline: image002.gif

Re: JSON Format Mbeans Stat Error

2012-12-04 Thread Chris Hostetter

: The result that we have is :
: 
: solr-mbeans:[
: CORE,{...},
: QUERYHANDLER,{...},
: CACHE,{...},...
: 
: In other terms, a table with a succession of
: key,value,key,value,key,value,... and not a map.

Correct.  see the json.nl param which controls how NamedList data is 
included in the response...

http://wiki.apache.org/solr/SolJSON#JSON_specific_parameters


-Hoss

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Mark Miller
You should take a look at the precommit target you pointed me at!

- Mark

On Dec 2, 2012, at 3:00 PM, Michael McCandless luc...@mikemccandless.com 
wrote:

 Woops, I committed a fix ...
 
 Mike McCandless
 
 http://blog.mikemccandless.com
 
 
 On Sun, Dec 2, 2012 at 5:53 PM, Policeman Jenkins Server
 jenk...@sd-datasolutions.de wrote:
 Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-trunk-Linux/3026/
 Java: 32bit/jdk1.7.0_09 -server -XX:+UseParallelGC
 
 All tests passed
 
 Build Log:
 [...truncated 25027 lines...]
 BUILD FAILED
 /mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:60: The 
 following error occurred while executing this line:
 /mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/build.xml:285: The 
 following error occurred while executing this line:
 /mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/common-build.xml:1526:
  The following error occurred while executing this line:
 /mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/common-build.xml:1560:
  Compile failed; see the compiler error output for details.
 
 Total time: 28 minutes 52 seconds
 Build step 'Invoke Ant' marked build as failure
 Archiving artifacts
 Recording test results
 Description set: Java: 32bit/jdk1.7.0_09 -server -XX:+UseParallelGC
 Email was triggered for: Failure
 Sending email for trigger: Failure
 
 
 
 
 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org
 
 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org
 


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Michael McCandless
Yes, my bad, sorry :(

I share your view that it's annoyingly slow to run the top-level ant
precommit, especially when it's a tiny change.

So this time I took a shortcut and ran ant javadocs in lucene/core
as a proxy for a top-level ant precommit, yet that was not good
enough because the ecj compiler is more picky.

I'd really love to have a per-module ant precommit.

Mike McCandless

http://blog.mikemccandless.com

On Tue, Dec 4, 2012 at 2:25 PM, Mark Miller markrmil...@gmail.com wrote:
 You should take a look at the precommit target you pointed me at!

 - Mark

 On Dec 2, 2012, at 3:00 PM, Michael McCandless luc...@mikemccandless.com 
 wrote:

 Woops, I committed a fix ...

 Mike McCandless

 http://blog.mikemccandless.com


 On Sun, Dec 2, 2012 at 5:53 PM, Policeman Jenkins Server
 jenk...@sd-datasolutions.de wrote:
 Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-trunk-Linux/3026/
 Java: 32bit/jdk1.7.0_09 -server -XX:+UseParallelGC

 All tests passed

 Build Log:
 [...truncated 25027 lines...]
 BUILD FAILED
 /mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:60: The 
 following error occurred while executing this line:
 /mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/build.xml:285: 
 The following error occurred while executing this line:
 /mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/common-build.xml:1526:
  The following error occurred while executing this line:
 /mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/common-build.xml:1560:
  Compile failed; see the compiler error output for details.

 Total time: 28 minutes 52 seconds
 Build step 'Invoke Ant' marked build as failure
 Archiving artifacts
 Recording test results
 Description set: Java: 32bit/jdk1.7.0_09 -server -XX:+UseParallelGC
 Email was triggered for: Failure
 Sending email for trigger: Failure




 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org

 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org



 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4114) Collection API: Allow multiple shards from one collection on the same Solr server

2012-12-04 Thread Per Steffensen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509960#comment-13509960
 ] 

Per Steffensen commented on SOLR-4114:
--

bq. We poll so that we wait up to 120 seconds for a slow comp, but a fast comp 
won't need to wait nearly that long. The 60 wait hits everyone no matter what. 
We generally try avoid any hard waits like that. I understand you can't really 
poll in this case, so I'm not sure the best way to test that - I made it a TODO 
for now.

Ok with a TODO, but it should be about making a more clever solution than the 
60 sec wait. You commented out the assert-method checkCollectionIsNotCreated, 
which means that we now have an unprotected feature in the code-base. Anyone 
can go ruin this feature tomorrow and no one will notice. Yes, I believe the 
main value of tests is their ability to protect other people from accidently 
ruining existing fuctionality. All the comments below are very unimportant 
compared to this - I am really serious about this. Get 
checkCollectionIsNotCreated running now so that the code is protected, then 
think about a more clever solution later (if you think such a thing exists). 
TODO's have a tendency to be forgotten.

bq. Yeah, I minimized the patch down to just dealing with this issue. I'm away 
from home and just looking to get this issue completed with minimum fuss.

Then the easiest thing would probably have been, to take everything in, except 
for things you thought would actually ruin existing stuff. Instead of using 
time to find every little detail that could be left out. Do not misunderstand 
me, I am glad you used your time to get it committed, but I also want to 
influence the way you committers work, whenever I have the chance. Only 
thinking about our common interrest - a good Solr. I have a bad gut feeling 
that the code-base is so messy because no one ever refactors. Refactoring is 
something you usually do while solving some specific (potentially) unrelated 
task. No one goes refactoring just to do refactoring, but it is extremely 
important that refactoring has an easy way into the code-base.

bq. 'nodeUrlWithoutProtocolPart' is just so long and I didn't see it helping 
much in terms of code readability - if you have a better fitting name that is a 
little less verbose, I think I'd be more into it.

Well, first of all a long saying name is much better than a short misleading 
name, and second of all that name really isnt very long :-)

bq. Yeah, I don't think I agree with changing the users params on him - I'd 
rather warn and let the user do what he wants to do rather than trying to 
outthink him ahead of time. If he decides he wants more than one repica on an 
instance for some reason, that's his deal - we can warn him though.

Ok, cool

bq. Right - it should match collection1 - eg newcollection/data should be the 
data dir just like collection1/data and rather than something like 
newcollection_data. In my experience and testing, data ends up in the cores own 
instance dir - not some common dir.

Didnt learn much from this explanation, but I will have to do a little studying 
on instance-dir and data-dir to understand how your solution will ever work. I 
will get back with an official complaint (just a comment here or a mail on the 
mailing list :-) ) if I still do not think it will work after I have done my 
studying.

bq. Yup - it seemed unrelated and I'm busy so I didn't want to think about it...

Still easier to just take it in, unless you saw it harming more than it helped. 
I am worried about refactoring in this project! Trust your test-suite :-)


 Collection API: Allow multiple shards from one collection on the same Solr 
 server
 -

 Key: SOLR-4114
 URL: https://issues.apache.org/jira/browse/SOLR-4114
 Project: Solr
  Issue Type: New Feature
  Components: multicore, SolrCloud
Affects Versions: 4.0
 Environment: Solr 4.0.0 release
Reporter: Per Steffensen
Assignee: Per Steffensen
  Labels: collection-api, multicore, shard, shard-allocation
 Attachments: SOLR-4114.patch, SOLR-4114.patch, SOLR-4114.patch, 
 SOLR-4114.patch, SOLR-4114_trunk.patch


 We should support running multiple shards from one collection on the same 
 Solr server - the run a collection with 8 shards on a 4 Solr server cluster 
 (each Solr server running 2 shards).
 Performance tests at our side has shown that this is a good idea, and it is 
 also a good idea for easy elasticity later on - it is much easier to move an 
 entire existing shards from one Solr server to another one that just joined 
 the cluter than it is to split an exsiting shard among the Solr that used to 
 run it and the new Solr.
 See dev mailing list discussion 

Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Steve Rowe
On Dec 4, 2012, at 2:45 PM, Michael McCandless luc...@mikemccandless.com 
wrote:
 I'd really love to have a per-module ant precommit.

+1
-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Chris Hostetter

: I'd really love to have a per-module ant precommit.

That's not really viable is it?

small tweaks in one class might leave everything fine and dandy in that 
module, but another module that depends on it could break as a result (eg: 
removing a method/variable that is {@link}ed to, etc...)


-Hoss

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4114) Collection API: Allow multiple shards from one collection on the same Solr server

2012-12-04 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13509967#comment-13509967
 ] 

Mark Miller commented on SOLR-4114:
---

bq. Still easier to just take it in

For me it's easier to not take it in :) I have to vet what I take in. I think 
you will find it easier to get stuff in if your refactoring is related to the 
issue. Otherwise make a refactoring issue.

bq. Trust your test-suite 

It's not my test-suite, it's a huge shard test-suite and I don't blindly trust 
it. It certainly doesn't test everything.

 Collection API: Allow multiple shards from one collection on the same Solr 
 server
 -

 Key: SOLR-4114
 URL: https://issues.apache.org/jira/browse/SOLR-4114
 Project: Solr
  Issue Type: New Feature
  Components: multicore, SolrCloud
Affects Versions: 4.0
 Environment: Solr 4.0.0 release
Reporter: Per Steffensen
Assignee: Per Steffensen
  Labels: collection-api, multicore, shard, shard-allocation
 Attachments: SOLR-4114.patch, SOLR-4114.patch, SOLR-4114.patch, 
 SOLR-4114.patch, SOLR-4114_trunk.patch


 We should support running multiple shards from one collection on the same 
 Solr server - the run a collection with 8 shards on a 4 Solr server cluster 
 (each Solr server running 2 shards).
 Performance tests at our side has shown that this is a good idea, and it is 
 also a good idea for easy elasticity later on - it is much easier to move an 
 entire existing shards from one Solr server to another one that just joined 
 the cluter than it is to split an exsiting shard among the Solr that used to 
 run it and the new Solr.
 See dev mailing list discussion Multiple shards for one collection on the 
 same Solr server

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Mark Miller
Well perhaps there is a middle ground - 'quasi-precommit' that just checks fast 
stuff per module… (eg doesn't build all the javadocs)

Then we would catch more, but perhaps not everything.

Precommit is scary slow and if it's so slow no one will use it, it's almost not 
so helpful.

- Mark

On Dec 4, 2012, at 11:53 AM, Chris Hostetter hossman_luc...@fucit.org wrote:

 
 : I'd really love to have a per-module ant precommit.
 
 That's not really viable is it?
 
 small tweaks in one class might leave everything fine and dandy in that 
 module, but another module that depends on it could break as a result (eg: 
 removing a method/variable that is {@link}ed to, etc...)
 
 
 -Hoss
 
 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org
 


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Michael McCandless
On Tue, Dec 4, 2012 at 3:00 PM, Mark Miller markrmil...@gmail.com wrote:
 Well perhaps there is a middle ground - 'quasi-precommit' that just checks 
 fast stuff per module… (eg doesn't build all the javadocs)

Right, is seems like a good number off the checks (no tabs, nocommits,
author tags, ecj is happy) can be done only on the current module ...

 Precommit is scary slow and if it's so slow no one will use it, it's almost 
 not so helpful.

Especially when the original change is trivial (all I did was rename a
boolean param, and add some javadocs about it, here).

Hmm I wonder if we could parallelize it ...

Mike McCandless

http://blog.mikemccandless.com

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Steve Rowe
I think one useful middle point might be to only run checks on modules that 
have pending changes.  Yes, as Hoss points out, an unchanged module's javadocs 
might no longer work (etc.), but I don't think that's the common case.

On Dec 4, 2012, at 3:00 PM, Mark Miller markrmil...@gmail.com wrote:

 Well perhaps there is a middle ground - 'quasi-precommit' that just checks 
 fast stuff per module… (eg doesn't build all the javadocs)
 
 Then we would catch more, but perhaps not everything.
 
 Precommit is scary slow and if it's so slow no one will use it, it's almost 
 not so helpful.
 
 - Mark
 
 On Dec 4, 2012, at 11:53 AM, Chris Hostetter hossman_luc...@fucit.org wrote:
 
 
 : I'd really love to have a per-module ant precommit.
 
 That's not really viable is it?
 
 small tweaks in one class might leave everything fine and dandy in that 
 module, but another module that depends on it could break as a result (eg: 
 removing a method/variable that is {@link}ed to, etc...)
 
 
 -Hoss
 
 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org
 
 
 
 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org
 


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Robert Muir
I think the slowness problem can be fixed way easier than that.

Actually in my opinion the main bug is the speed of building javadocs
themselves. Today this requires lots of building of useless jars and some
other things that are unnecessary.
I also think javadocs for a module get rebuilt across the whole thing
because they are depended on by some other module's javadocs, and so on.

I really don't think the checks should be weakened in any way. These checks
are necessary (people will complain about them at release vote time, so we
have to keep the codebase in good order and not just all make a mess for
months, waiting for some RM to clean it up).


On Tue, Dec 4, 2012 at 3:16 PM, Steve Rowe sar...@gmail.com wrote:

 I think one useful middle point might be to only run checks on modules
 that have pending changes.  Yes, as Hoss points out, an unchanged module's
 javadocs might no longer work (etc.), but I don't think that's the common
 case.

 On Dec 4, 2012, at 3:00 PM, Mark Miller markrmil...@gmail.com wrote:

  Well perhaps there is a middle ground - 'quasi-precommit' that just
 checks fast stuff per module… (eg doesn't build all the javadocs)
 
  Then we would catch more, but perhaps not everything.
 
  Precommit is scary slow and if it's so slow no one will use it, it's
 almost not so helpful.
 
  - Mark
 
  On Dec 4, 2012, at 11:53 AM, Chris Hostetter hossman_luc...@fucit.org
 wrote:
 
 
  : I'd really love to have a per-module ant precommit.
 
  That's not really viable is it?
 
  small tweaks in one class might leave everything fine and dandy in that
  module, but another module that depends on it could break as a result
 (eg:
  removing a method/variable that is {@link}ed to, etc...)
 
 
  -Hoss
 
  -
  To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
  For additional commands, e-mail: dev-h...@lucene.apache.org
 
 
 
  -
  To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
  For additional commands, e-mail: dev-h...@lucene.apache.org
 


 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org




[jira] [Created] (LUCENE-4587) WordBreakSpellChecker treats bytes as chars

2012-12-04 Thread James Dyer (JIRA)
James Dyer created LUCENE-4587:
--

 Summary: WordBreakSpellChecker treats bytes as chars
 Key: LUCENE-4587
 URL: https://issues.apache.org/jira/browse/LUCENE-4587
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/spellchecker
Affects Versions: 4.0
Reporter: Andreas Hubold
Assignee: James Dyer
 Fix For: 4.1, 5.0


Originally opened as SOLR-4115.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-4115) WordBreakSpellChecker throws ArrayIndexOutOfBoundsException for random query string

2012-12-04 Thread James Dyer (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4115?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Dyer resolved SOLR-4115.
--

Resolution: Duplicate
  Assignee: James Dyer

I opened LUCENE-4587 and will resolve with a fix there.

 WordBreakSpellChecker throws ArrayIndexOutOfBoundsException for random query 
 string
 ---

 Key: SOLR-4115
 URL: https://issues.apache.org/jira/browse/SOLR-4115
 Project: Solr
  Issue Type: Bug
  Components: spellchecker
Affects Versions: 4.0
 Environment: java version 1.6.0_37
 Java(TM) SE Runtime Environment (build 1.6.0_37-b06)
 Java HotSpot(TM) 64-Bit Server VM (build 20.12-b01, mixed mode)
Reporter: Andreas Hubold
Assignee: James Dyer

 The following SolrJ test code causes an ArrayIndexOutOfBoundsException in the 
 WordBreakSpellChecker. I tested this with the Solr 4.0.0 example webapp 
 started with {{java -jar start.jar}}.
 {code:java}
   @Test
   public void testWordbreakSpellchecker() throws Exception {
 SolrQuery q = new SolrQuery(\uD864\uDC79);
 q.setRequestHandler(/browse);
 q.setParam(spellcheck.dictionary, wordbreak);
 HttpSolrServer server = new HttpSolrServer(http://localhost:8983/solr;);
 server.query(q, SolrRequest.METHOD.POST);
   }
 {code}
 {noformat}
 INFO: [collection1] webapp=/solr path=/browse 
 params={spellcheck.dictionary=wordbreakqt=/browsewt=javabinq=?version=2} 
 hits=0 status=500 QTime=11 
 Nov 28, 2012 11:23:01 AM org.apache.solr.common.SolrException log
 SEVERE: null:java.lang.ArrayIndexOutOfBoundsException: 1
   at org.apache.lucene.util.UnicodeUtil.UTF8toUTF16(UnicodeUtil.java:599)
   at org.apache.lucene.util.BytesRef.utf8ToString(BytesRef.java:165)
   at org.apache.lucene.index.Term.text(Term.java:72)
   at 
 org.apache.lucene.search.spell.WordBreakSpellChecker.generateSuggestWord(WordBreakSpellChecker.java:350)
   at 
 org.apache.lucene.search.spell.WordBreakSpellChecker.generateBreakUpSuggestions(WordBreakSpellChecker.java:283)
   at 
 org.apache.lucene.search.spell.WordBreakSpellChecker.suggestWordBreaks(WordBreakSpellChecker.java:122)
   at 
 org.apache.solr.spelling.WordBreakSolrSpellChecker.getSuggestions(WordBreakSolrSpellChecker.java:229)
   at 
 org.apache.solr.handler.component.SpellCheckComponent.process(SpellCheckComponent.java:172)
   at 
 org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:206)
   at 
 org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129)
   at org.apache.solr.core.SolrCore.execute(SolrCore.java:1699)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:455)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:276)
   at 
 org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1337)
   at 
 org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:484)
   at 
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:119)
   at 
 org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:524)
   at 
 org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:233)
   at 
 org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1065)
   at 
 org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:413)
   at 
 org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:192)
   at 
 org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:999)
   at 
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117)
   at 
 org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:250)
   at 
 org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:149)
   at 
 org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:111)
   at org.eclipse.jetty.server.Server.handle(Server.java:351)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:454)
   at 
 org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:47)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection.content(AbstractHttpConnection.java:900)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.content(AbstractHttpConnection.java:954)
   at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:857)
   at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235)
   at 
 org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:66)
   at 
 

Re: Solrj question

2012-12-04 Thread Per Steffensen

Steve Molloy skrev:

I really hate duplicating code

I love you :-)
Unless there's a way to get the original message before it's 
unmarshaled into the QueryResponse, in which case I'd be more than 
happy to use that as-is. J
Ha ha. Just had a short look at the code, and tried to hack it, but got 
tired, so I didnt complete or test it. But maybe this will help you 
(completely untested - dont even know if it will compile):


   CloudSolrServer client = new 
CloudSolrServer(zk-connect-string, new LBHttpSolrServer(null) {

   @Override
   protected HttpSolrServer makeServer(String 
baseUrl) throws MalformedURLException {
   return new HttpSolrServer(baseUrl, 
getHttpClient(), null) {

   @Override
   public NamedListObject request(final 
SolrRequest request) throws SolrServerException, IOException{

   return request(request, null);
   }
   };
   }
   });
   InputStream responseDirectlyFromWire = 
(InputStream)client.query(query).getResponse().get(stream);




[jira] [Updated] (LUCENE-4499) Multi-word synonym filter (synonym expansion)

2012-12-04 Thread Roman Chyla (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Chyla updated LUCENE-4499:


Attachment: LUCENE-4499.patch

A new patch, as the old version was extending wrong class (which cause web 
tests to fail)

 Multi-word synonym filter (synonym expansion)
 -

 Key: LUCENE-4499
 URL: https://issues.apache.org/jira/browse/LUCENE-4499
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/other
Affects Versions: 4.1, 5.0
Reporter: Roman Chyla
Priority: Minor
  Labels: analysis, multi-word, synonyms
 Fix For: 5.0

 Attachments: LUCENE-4499.patch, LUCENE-4499.patch


 I apologize for bringing the multi-token synonym expansion up again. There is 
 an old, unresolved issue at LUCENE-1622 [1]
 While solving the problem for our needs [2], I discovered that the current 
 SolrSynonym parser (and the wonderful FTS) have almost everything to 
 satisfactorily handle both the query and index time synonym expansion. It 
 seems that people often need to use the synonym filter *slightly* differently 
 at indexing and query time.
 In our case, we must do different things during indexing and querying.
 Example sentence: Mirrors of the Hubble space telescope pointed at XA5
 This is what we need (comma marks position bump):
 indexing: mirrors,hubble|hubble space 
 telescope|hst,space,telescope,pointed,xa5|astroobject#5
 querying: +mirrors +(hubble space telescope | hst) +pointed 
 +(xa5|astroboject#5)
 This translated to following needs:
   indexing time: 
 single-token synonyms = return only synonyms
 multi-token synonyms = return original tokens *AND* the synonyms
   query time:
 single-token: return only synonyms (but preserve case)
 multi-token: return only synonyms
  
 We need the original tokens for the proximity queries, if we indexed 'hubble 
 space telescope'
 as one token, we cannot search for 'hubble NEAR telescope'
 You may (not) be surprised, but Lucene already supports ALL of these 
 requirements. The patch is an attempt to state the problem differently. I am 
 not sure if it is the best option, however it works perfectly for our needs 
 and it seems it could work for general public too. Especially if the 
 SynonymFilterFactory had a preconfigured sets of SynonymMapBuilders - and 
 people would just choose what situation they use. Please look at the unittest.
 links:
 [1] https://issues.apache.org/jira/browse/LUCENE-1622
 [2] http://labs.adsabs.harvard.edu/trac/ads-invenio/ticket/158
 [3] seems to have similar request: 
 http://lucene.472066.n3.nabble.com/Proposal-Full-support-for-multi-word-synonyms-at-query-time-td4000522.html

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Mark Miller
We are not talking about weakening them on Jenkins though. It won't wait till 
release time, it will be caught 10 min later by Jenkins. 

But if we can make it faster and keep all the checks, then by all means!

Mark

Sent from my iPhone

On Dec 4, 2012, at 12:23 PM, Robert Muir rcm...@gmail.com wrote:

 I think the slowness problem can be fixed way easier than that.
 
 Actually in my opinion the main bug is the speed of building javadocs 
 themselves. Today this requires lots of building of useless jars and some 
 other things that are unnecessary.
 I also think javadocs for a module get rebuilt across the whole thing 
 because they are depended on by some other module's javadocs, and so on.
 
 I really don't think the checks should be weakened in any way. These checks 
 are necessary (people will complain about them at release vote time, so we 
 have to keep the codebase in good order and not just all make a mess for 
 months, waiting for some RM to clean it up).
 
 
 On Tue, Dec 4, 2012 at 3:16 PM, Steve Rowe sar...@gmail.com wrote:
 I think one useful middle point might be to only run checks on modules that 
 have pending changes.  Yes, as Hoss points out, an unchanged module's 
 javadocs might no longer work (etc.), but I don't think that's the common 
 case.
 
 On Dec 4, 2012, at 3:00 PM, Mark Miller markrmil...@gmail.com wrote:
 
  Well perhaps there is a middle ground - 'quasi-precommit' that just checks 
  fast stuff per module… (eg doesn't build all the javadocs)
 
  Then we would catch more, but perhaps not everything.
 
  Precommit is scary slow and if it's so slow no one will use it, it's 
  almost not so helpful.
 
  - Mark
 
  On Dec 4, 2012, at 11:53 AM, Chris Hostetter hossman_luc...@fucit.org 
  wrote:
 
 
  : I'd really love to have a per-module ant precommit.
 
  That's not really viable is it?
 
  small tweaks in one class might leave everything fine and dandy in that
  module, but another module that depends on it could break as a result (eg:
  removing a method/variable that is {@link}ed to, etc...)
 
 
  -Hoss
 
  -
  To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
  For additional commands, e-mail: dev-h...@lucene.apache.org
 
 
 
  -
  To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
  For additional commands, e-mail: dev-h...@lucene.apache.org
 
 
 
 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org
 


Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Robert Muir
On Tue, Dec 4, 2012 at 3:33 PM, Mark Miller markrmil...@gmail.com wrote:

 We are not talking about weakening them on Jenkins though. It won't wait
 till release time, it will be caught 10 min later by Jenkins.


Right but that then doesnt really improve the situation (its less than 10
minutes to check locally),
because if someone does commit-and-run we are still left with a broke
build.

I'm not concerned about this stuff tripping at all if people are looking
out and fixing stuff.

I do have some concerns about not running the javadocs checks, because part
of the point is check everything we can automatically since it makes
someone go and look at the file and perhaps then they also see some text is
out of date too (not just syntax). I'm not trying to say the system is
perfect or even works, only time will tell. But I do feel like it helps
keep the documentation from falling out of date, especially when the stuff
is fresh in the committers mind.


[jira] [Commented] (SOLR-4114) Collection API: Allow multiple shards from one collection on the same Solr server

2012-12-04 Thread Per Steffensen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13510020#comment-13510020
 ] 

Per Steffensen commented on SOLR-4114:
--

Ok, will you please consider enabling the 60 sec test (maybe reduce it to 10 or 
30 sec) so that the feature is protected until someone comes up with a 
better/faster test. Please!!!

 Collection API: Allow multiple shards from one collection on the same Solr 
 server
 -

 Key: SOLR-4114
 URL: https://issues.apache.org/jira/browse/SOLR-4114
 Project: Solr
  Issue Type: New Feature
  Components: multicore, SolrCloud
Affects Versions: 4.0
 Environment: Solr 4.0.0 release
Reporter: Per Steffensen
Assignee: Per Steffensen
  Labels: collection-api, multicore, shard, shard-allocation
 Attachments: SOLR-4114.patch, SOLR-4114.patch, SOLR-4114.patch, 
 SOLR-4114.patch, SOLR-4114_trunk.patch


 We should support running multiple shards from one collection on the same 
 Solr server - the run a collection with 8 shards on a 4 Solr server cluster 
 (each Solr server running 2 shards).
 Performance tests at our side has shown that this is a good idea, and it is 
 also a good idea for easy elasticity later on - it is much easier to move an 
 entire existing shards from one Solr server to another one that just joined 
 the cluter than it is to split an exsiting shard among the Solr that used to 
 run it and the new Solr.
 See dev mailing list discussion Multiple shards for one collection on the 
 same Solr server

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4114) Collection API: Allow multiple shards from one collection on the same Solr server

2012-12-04 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13510038#comment-13510038
 ] 

Mark Miller commented on SOLR-4114:
---

I'm happy to commit a test for this, but lets come up with something that 
doesn't have a long sleep  like this. 
I suppose a 10 second sleep is more agreeable, but these things add up and  and 
I'd rather come up with a better test. 

 Collection API: Allow multiple shards from one collection on the same Solr 
 server
 -

 Key: SOLR-4114
 URL: https://issues.apache.org/jira/browse/SOLR-4114
 Project: Solr
  Issue Type: New Feature
  Components: multicore, SolrCloud
Affects Versions: 4.0
 Environment: Solr 4.0.0 release
Reporter: Per Steffensen
Assignee: Per Steffensen
  Labels: collection-api, multicore, shard, shard-allocation
 Attachments: SOLR-4114.patch, SOLR-4114.patch, SOLR-4114.patch, 
 SOLR-4114.patch, SOLR-4114_trunk.patch


 We should support running multiple shards from one collection on the same 
 Solr server - the run a collection with 8 shards on a 4 Solr server cluster 
 (each Solr server running 2 shards).
 Performance tests at our side has shown that this is a good idea, and it is 
 also a good idea for easy elasticity later on - it is much easier to move an 
 entire existing shards from one Solr server to another one that just joined 
 the cluter than it is to split an exsiting shard among the Solr that used to 
 run it and the new Solr.
 See dev mailing list discussion Multiple shards for one collection on the 
 same Solr server

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4419) Test RecursivePrefixTree indexing non-point data

2012-12-04 Thread David Smiley (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4419?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Smiley updated LUCENE-4419:
-

Attachment: LUCENE-4419_Test_indexing_non-point_shapes.patch

The attached test SpatialOpRecursivePrefixTreeTest tests the 
RecursivePrefixTree extensively.  It yielded the bugs in LUCENE-4585.  It uses 
the QuadPrefixTree with random levels and random scan threshold for 
RecursivePrefixTreeFilter.  Some randomly generated rectangles are asserted as 
a MUST match, and some are optional depending on the grid approximation.

QuadPrefixTree was chosen because it supports non-geo, and this test has some 
simplistic logic in going from shape - grid-snapped shape that would be more 
complicated in geospatial that ultimately RecursivePrefixTree doesn't care 
about any way -- it's Spatial4j + GeohashPrefixTree that deal with that.

This test added an evaluate() method to SpatialOperation which I found quite 
handy.  With this in place and this test as a template, it should be easy to 
test for operations other than intersect once they are supported.

 Test RecursivePrefixTree indexing non-point data
 

 Key: LUCENE-4419
 URL: https://issues.apache.org/jira/browse/LUCENE-4419
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/spatial
Reporter: David Smiley
Assignee: David Smiley
 Attachments: LUCENE-4419_Test_indexing_non-point_shapes.patch


 RecursivePrefixTreeFilter was modified in ~July 2011 to support spatial 
 filtering of non-point indexed shapes.  It seems to work when playing with 
 the capability but it isn't tested.  It really needs to be as this is a major 
 feature.
 I imagine an approach in which some randomly generated rectangles are indexed 
 and then a randomly generated rectangle is queried.  The right answer can be 
 calculated brute-force and then compared with the filter.  In order to deal 
 with shape imprecision, the randomly generated shapes could be generated to 
 fit a course grid (e.g. round everything to a 1 degree interval).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Mark Miller
I guess my point is that since no one runs it (except perhaps you), it's not 
really helping anything.

And the broken build is really more minor than a true broken build if tests 
at least still pass. Personally, I'm pretty okay with a nocommit failing the 
build for 30 minutes if the tests still pass.

Most of these issues are basic enough that anyone can jump in and fix it if 
it's causing them issues as well. Just remove the tabs, fix the javadoc, etc.

- Mark

On Dec 4, 2012, at 12:56 PM, Robert Muir rcm...@gmail.com wrote:

 
 
 On Tue, Dec 4, 2012 at 3:33 PM, Mark Miller markrmil...@gmail.com wrote:
 We are not talking about weakening them on Jenkins though. It won't wait till 
 release time, it will be caught 10 min later by Jenkins. 
 
 
 Right but that then doesnt really improve the situation (its less than 10 
 minutes to check locally),
 because if someone does commit-and-run we are still left with a broke build.
 
 I'm not concerned about this stuff tripping at all if people are looking out 
 and fixing stuff.
 
 I do have some concerns about not running the javadocs checks, because part 
 of the point is check everything we can automatically since it makes 
 someone go and look at the file and perhaps then they also see some text is 
 out of date too (not just syntax). I'm not trying to say the system is 
 perfect or even works, only time will tell. But I do feel like it helps keep 
 the documentation from falling out of date, especially when the stuff is 
 fresh in the committers mind.


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4419) Test RecursivePrefixTree indexing non-point data

2012-12-04 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4419?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13510077#comment-13510077
 ] 

David Smiley commented on LUCENE-4419:
--

Whoops; there's a stupid if (i != 5) continue; on line ~68 which I should 
have labelled with a NOCOMMIT that's in the patch.

 Test RecursivePrefixTree indexing non-point data
 

 Key: LUCENE-4419
 URL: https://issues.apache.org/jira/browse/LUCENE-4419
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/spatial
Reporter: David Smiley
Assignee: David Smiley
 Attachments: LUCENE-4419_Test_indexing_non-point_shapes.patch


 RecursivePrefixTreeFilter was modified in ~July 2011 to support spatial 
 filtering of non-point indexed shapes.  It seems to work when playing with 
 the capability but it isn't tested.  It really needs to be as this is a major 
 feature.
 I imagine an approach in which some randomly generated rectangles are indexed 
 and then a randomly generated rectangle is queried.  The right answer can be 
 calculated brute-force and then compared with the filter.  In order to deal 
 with shape imprecision, the randomly generated shapes could be generated to 
 fit a course grid (e.g. round everything to a 1 degree interval).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-4145) Getting a NPE when the result set is null

2012-12-04 Thread Amandeep Saini (JIRA)
Amandeep Saini created SOLR-4145:


 Summary: Getting a NPE when the result set is null
 Key: SOLR-4145
 URL: https://issues.apache.org/jira/browse/SOLR-4145
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.0
Reporter: Amandeep Saini


Re: org.apache.solr.handler.dataimport.JdbcDataSource

In line 250, resultSet.getMetaData() is called. However, if the resultSet is 
null, a NPE is thrown.

In line 254, the null check is done (if (resultSet == null))
The contents of line 250 should occur after the null check.

I am getting the NPE after trying to delete some data in my DB using the DIH. 
The DIH is probably not meant for that but it would benefit me if it worked and 
I don't see a big change in the code in order to do it.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4144) SolrCloud replication high heap consumption

2012-12-04 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13510105#comment-13510105
 ] 

Mark Miller commented on SOLR-4144:
---

Thanks for the report! I will try and investigate this soon - I'm on an 11 inch 
low powered macbook air for the week, so I may not get to it till next week, 
but we will see...

 SolrCloud replication high heap consumption
 ---

 Key: SOLR-4144
 URL: https://issues.apache.org/jira/browse/SOLR-4144
 Project: Solr
  Issue Type: Bug
  Components: replication (java), SolrCloud
Affects Versions: 5.0
 Environment: 5.0-SNAPSHOT 1366361:1416494M - markus - 2012-12-03 
 14:09:13
Reporter: Markus Jelsma
Priority: Critical
 Fix For: 5.0


 Recent versions of SolrCloud require a very high heap size vs. older 
 versions. Another cluster of 5.0.0.2012.10.09.19.29.59 (~4GB per core) can 
 restore an empty node without taking a lot of heap (xmx=256m). Recent 
 versions and current trunk fail miserably even with a higher heap (750m). 
 Both clusters have 10 nodes, 10 shards and 2 cores per node. One note to add 
 is that the cluster on which this fails has only about 1.5GB per core due to 
 changing in the Lucene codec such as compression.
 After start up everything goes fine...
 {code}
 2012-12-04 15:05:35,013 INFO [solr.cloud.RecoveryStrategy] - [RecoveryThread] 
 - : Begin buffering updates. core=shard_c
 2012-12-04 15:05:35,013 INFO [solr.cloud.RecoveryStrategy] - [RecoveryThread] 
 - : Begin buffering updates. core=shard_b
 2012-12-04 15:05:35,013 INFO [solr.update.UpdateLog] - [RecoveryThread] - : 
 Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
 2012-12-04 15:05:35,013 INFO [solr.update.UpdateLog] - [RecoveryThread] - : 
 Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
 2012-12-04 15:05:35,013 INFO [solr.cloud.RecoveryStrategy] - [RecoveryThread] 
 - : Attempting to replicate from http://178.21.118.190:8080/solr/shard_b/. 
 core=shard_b
 2012-12-04 15:05:35,013 INFO [solr.cloud.RecoveryStrategy] - [RecoveryThread] 
 - : Attempting to replicate from http://178.21.118.192:8080/solr/shard_c/. 
 core=shard_c
 2012-12-04 15:05:35,014 INFO [solrj.impl.HttpClientUtil] - [RecoveryThread] - 
 : Creating new http client, 
 config:maxConnections=128maxConnectionsPerHost=32followRedirects=false
 2012-12-04 15:05:35,014 INFO [solrj.impl.HttpClientUtil] - [RecoveryThread] - 
 : Creating new http client, 
 config:maxConnections=128maxConnectionsPerHost=32followRedirects=false
 2012-12-04 15:05:35,052 INFO [solr.handler.ReplicationHandler] - 
 [RecoveryThread] - : Commits will be reserved for  1
 2012-12-04 15:05:35,052 INFO [solr.handler.ReplicationHandler] - 
 [RecoveryThread] - : Commits will be reserved for  1
 2012-12-04 15:05:35,053 INFO [solrj.impl.HttpClientUtil] - [RecoveryThread] - 
 : Creating new http client, 
 config:connTimeout=5000socketTimeout=2allowCompression=falsemaxConnections=1maxConnectionsPerHost=1
 2012-12-04 15:05:35,060 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
  No value set for 'pollInterval'. Timer Task not started.
 2012-12-04 15:05:35,060 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
  No value set for 'pollInterval'. Timer Task not started.
 2012-12-04 15:05:35,069 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
 Master's generation: 48
 2012-12-04 15:05:35,069 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
 Slave's generation: 1
 2012-12-04 15:05:35,069 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
 Starting replication process
 2012-12-04 15:05:35,069 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
 Master's generation: 47
 2012-12-04 15:05:35,069 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
 Slave's generation: 1
 2012-12-04 15:05:35,070 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
 Starting replication process
 2012-12-04 15:05:35,078 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
 Number of files in latest index in master: 235
 2012-12-04 15:05:35,079 INFO [solr.handler.SnapPuller] - [RecoveryThread] - : 
 Number of files in latest index in master: 287
 2012-12-04 15:05:35,084 WARN [solr.core.CachingDirectoryFactory] - 
 [RecoveryThread] - : No lockType configured for 
 NRTCachingDirectory(org.apache.lucene.store.MMapDirectory@/opt/solr/cores/shard_c/data/index.20121204150535080
  lockFactory=org.apache.lucene.store.NativeFSLockFactory@57530551; 
 maxCacheMB=48.0 maxMergeSizeMB=4.0) assuming 'simple'
 2012-12-04 15:05:35,085 INFO [solr.core.CachingDirectoryFactory] - 
 [RecoveryThread] - : return new directory for 
 /opt/solr/cores/shard_c/data/index.20121204150535080 forceNew:false
 2012-12-04 15:05:35,085 INFO [solr.core.CachingDirectoryFactory] - 
 

[jira] [Resolved] (SOLR-3987) Provide Collection API request results beyond manual inspection.

2012-12-04 Thread Mark Miller (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3987?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller resolved SOLR-3987.
---

Resolution: Duplicate
  Assignee: Mark Miller

I'm closing this as a dupe of SOLR-4043 since this one has a patch.

 Provide Collection API request results beyond manual inspection.
 

 Key: SOLR-3987
 URL: https://issues.apache.org/jira/browse/SOLR-3987
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Reporter: Mark Miller
Assignee: Mark Miller
 Fix For: 4.1, 5.0




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4043) Add ability to get success/failure responses from Collections API.

2012-12-04 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4043?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13510114#comment-13510114
 ] 

Mark Miller commented on SOLR-4043:
---

Thanks a lot for this patch Raintung! I'm going to try and get it up to date 
with trunk shortly.

 Add ability to get success/failure responses from Collections API.
 --

 Key: SOLR-4043
 URL: https://issues.apache.org/jira/browse/SOLR-4043
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Affects Versions: 4.0-ALPHA, 4.0-BETA, 4.0
 Environment: Solr cloud cluster
Reporter: Raintung Li
Assignee: Mark Miller
 Fix For: 4.1, 5.0

 Attachments: patch-4043.txt


 The create/delete/reload collections are asynchronous process, the client 
 can't get the right response, only make sure the information have been saved 
 into the OverseerCollectionQueue. The client will get the response directly 
 that don't wait the result of behavior(create/delete/reload collection) 
 whatever successful. 
 The easy solution is client wait until the asynchronous process success, the 
 create/delete/reload collection thread will save the response into 
 OverseerCollectionQueue, then notify client to get the response. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4138) currency field isn't compatible with frange query

2012-12-04 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4138?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-4138:
---

Attachment: SOLR-4138.patch

Thanks for reporting this Grzegorz,

The general problem seems to be that while CurrencyField has a 
CurrencyValueSource that it uses for range queries and sorting, it doesn't 
override FieldType.getValueSource() to return one of these in cases when people 
use currency fields in functions. 

Even with this simple addition however, your test as written still doesn't 
pass, because CurrencyValueSource operates using the internal long values of 
the currency, based on Currency.getDefaultFractionDigits() for the default 
currency .. so if USD is the configured default, you get a value of 500.0 
(cents) from the value source instead of 5.0 when a document is indexed with 
a value of 5,USD.

I've attached a patch with the trivial getVaueSource() and tests showing the 
_current_ behavior, but i'm not entirely certain if this is really the behavior 
we want ... changing it would probably require some meaty rewrites to 
CurrencyValueSource to take in options about whether the 
getDefaultFractionDigits() logic should be undone to produce the more human 
friendly values ... or perhaps just wrapping it in a special ValueSource that 
applies that logic in the getValueSource case so existing usages in 
getRangeQuery would be unaffected.

 currency field isn't compatible with frange query
 -

 Key: SOLR-4138
 URL: https://issues.apache.org/jira/browse/SOLR-4138
 Project: Solr
  Issue Type: Bug
  Components: search
Affects Versions: 4.0
Reporter: Grzegorz Sobczyk
 Attachments: SOLR-4135-test.patch, SOLR-4138.patch


 Filtering using {!frange} syntax isn't work properly. (rather at all)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4138) currency field doesn't work with functions (ie: isn't compatible with frange query)

2012-12-04 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4138?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-4138:
---

Description: 
In general, using CurrencyField with FunctionQueries doesn't work

In particular, as originally reported...

Filtering using {!frange} syntax isn't work properly. (rather at all)


  was:Filtering using {!frange} syntax isn't work properly. (rather at all)

Summary: currency field doesn't work with functions (ie: isn't 
compatible with frange query)  (was: currency field isn't compatible with 
frange query)

enhancing summary  description to reflect broader scope of issue

 currency field doesn't work with functions (ie: isn't compatible with frange 
 query)
 ---

 Key: SOLR-4138
 URL: https://issues.apache.org/jira/browse/SOLR-4138
 Project: Solr
  Issue Type: Bug
  Components: search
Affects Versions: 4.0
Reporter: Grzegorz Sobczyk
 Attachments: SOLR-4135-test.patch, SOLR-4138.patch


 In general, using CurrencyField with FunctionQueries doesn't work
 In particular, as originally reported...
 Filtering using {!frange} syntax isn't work properly. (rather at all)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3428) SolrCmdDistributor flushAdds/flushDeletes problems

2012-12-04 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3428?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13510132#comment-13510132
 ] 

Mark Miller commented on SOLR-3428:
---

This can likely be closed - at some point I fixed 1 and 2 (adding Per to the 
CHANGES entries) and added targeted tests for them and I'm not yet convinced 
that 3 is a problem (which doesn't guarantee it's not - but I have not found a 
situation where it is a problem).

 SolrCmdDistributor flushAdds/flushDeletes problems
 --

 Key: SOLR-3428
 URL: https://issues.apache.org/jira/browse/SOLR-3428
 Project: Solr
  Issue Type: Bug
  Components: replication (java), SolrCloud, update
Affects Versions: 4.0-ALPHA
Reporter: Per Steffensen
Assignee: Per Steffensen
  Labels: add, delete, replica, solrcloud, update
 Fix For: 4.1

   Original Estimate: 24h
  Remaining Estimate: 24h

 A few problems with SolrCmdDistributor.flushAdds/flushDeletes
 * If number of AddRequests/DeleteRequests in alist/dlist is below limit for a 
 specific node the method returns immediately and doesnt flush for subsequent 
 nodes
 * When returning immediately because there is below limit requests for a 
 given node, then previous nodes that have already been flushed/submitted are 
 not removed from adds/deletes maps (causing them to be flushed/submitted 
 again the next time flushAdds/flushDeletes is executed)
 * The idea about just combining params does not work for SEEN_LEADER params 
 (and probably others as well). Since SEEN_LEADER cannot be expressed (unlike 
 commitWithin and overwrite) for individual operations in the request, you 
 need to sent two separate submits. One containing requests with 
 SEEN_LEADER=true and one with SEEN_LEADER=false.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4043) Add ability to get success/failure responses from Collections API.

2012-12-04 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4043?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13510136#comment-13510136
 ] 

Mark Miller commented on SOLR-4043:
---

Raintung, is this patch against 4.0 rather than 4x or 5x on Nov 7?

 Add ability to get success/failure responses from Collections API.
 --

 Key: SOLR-4043
 URL: https://issues.apache.org/jira/browse/SOLR-4043
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Affects Versions: 4.0-ALPHA, 4.0-BETA, 4.0
 Environment: Solr cloud cluster
Reporter: Raintung Li
Assignee: Mark Miller
 Fix For: 4.1, 5.0

 Attachments: patch-4043.txt


 The create/delete/reload collections are asynchronous process, the client 
 can't get the right response, only make sure the information have been saved 
 into the OverseerCollectionQueue. The client will get the response directly 
 that don't wait the result of behavior(create/delete/reload collection) 
 whatever successful. 
 The easy solution is client wait until the asynchronous process success, the 
 create/delete/reload collection thread will save the response into 
 OverseerCollectionQueue, then notify client to get the response. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3239) Function for returning a CurrencyField as a virtual field

2012-12-04 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3239?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13510157#comment-13510157
 ] 

Hoss Man commented on SOLR-3239:


bq. A special case would be that you want only formatting, not conversion.

I'm not sure if that's a good idea ... it's tempting to look at your examples 
and think it should be the original currency, but what about when it's wrapped 
in another function?

This comes up because of some of the questions in SOLR-4138 -- I think it would 
make a lot of sense if currency(price_c) returned the human value of the 
price_c field in the default currency, (ie: 5.5 if the indexed value is 
equivalent to 5.5,USD) regardless of which currency was used when the doc was 
indexed.  That way you could do things like 
sum(currency(price_field),prod($ship_cost,weight_field)) and get comparable 
values across all documents regardless of the currency used in each document.

it might make more sense to have a currency(...) function that explicitly 
deals with the conversion of different currencies, and then have an independent 
\[currencyft\] DocTransformer for dealing with the locale based formatting of 
currency.  We'd get the currency conversion as input to other functions 
benefit i described above independently from the need to format; and you could 
have the format currencies in locale benefit independently of whether someone 
actually used CurrencyField

not to mention then you cold wrap the formating around some complex function 
that involved the currency values...

{code}
?q=foo
// uses default currency for consistency of value comparisons
total_price=sum(currency(price_field),prod($ship_cost,weight_field))
sort=$total_price+asc
// fmt overrides default currency to be users currency
fl=id,[currencyfmt v=$total_price currency=GBP locale=en_us]
{code}




 Function for returning a CurrencyField as a virtual field
 -

 Key: SOLR-3239
 URL: https://issues.apache.org/jira/browse/SOLR-3239
 Project: Solr
  Issue Type: New Feature
Reporter: Jan Høydahl
 Fix For: 4.1


 With the new CurrencyField from SOLR-2202, we can now do range queries, sort 
 and soon faceting on CurrencyField.
 One thing we cannot do is return a converted value. We can only return the 
 stored value which must be in the same currency as it was entered, e.g. 
 10.0,USD. People now have to do this conversion in application layer, 
 reading the same currency.xml file as in Solr. This is a mess to keep in 
 sync, and is hard to support with other ExchangeRateProviders or in other 
 client languages than Java.
 This patch should implement a Function which takes a field name and requested 
 currency as input and returns the converted value. Optionally it could also 
 take care of localized display, e.g. $10.0 instead of 10.0,USD. Proposed 
 syntax:
 {code}
 fl=title,price:currency(price_c,EUR) # Returns price_c in EUR as 
 price
 fl=title,price:currency(price_c,EUR,en_US)   # Returns price_c in EUR, 
 formatted according to en_US locale
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Steve Rowe
I'm running it.  It takes just over 4 minutes for me.  Pretty minor IMHO.

On Dec 4, 2012, at 5:26 PM, Mark Miller markrmil...@gmail.com wrote:

 I guess my point is that since no one runs it (except perhaps you), it's not 
 really helping anything.
 
 And the broken build is really more minor than a true broken build if tests 
 at least still pass. Personally, I'm pretty okay with a nocommit failing the 
 build for 30 minutes if the tests still pass.
 
 Most of these issues are basic enough that anyone can jump in and fix it if 
 it's causing them issues as well. Just remove the tabs, fix the javadoc, etc.
 
 - Mark
 
 On Dec 4, 2012, at 12:56 PM, Robert Muir rcm...@gmail.com wrote:
 
 
 
 On Tue, Dec 4, 2012 at 3:33 PM, Mark Miller markrmil...@gmail.com wrote:
 We are not talking about weakening them on Jenkins though. It won't wait 
 till release time, it will be caught 10 min later by Jenkins. 
 
 
 Right but that then doesnt really improve the situation (its less than 10 
 minutes to check locally),
 because if someone does commit-and-run we are still left with a broke 
 build.
 
 I'm not concerned about this stuff tripping at all if people are looking out 
 and fixing stuff.
 
 I do have some concerns about not running the javadocs checks, because part 
 of the point is check everything we can automatically since it makes 
 someone go and look at the file and perhaps then they also see some text is 
 out of date too (not just syntax). I'm not trying to say the system is 
 perfect or even works, only time will tell. But I do feel like it helps keep 
 the documentation from falling out of date, especially when the stuff is 
 fresh in the committers mind.
 
 
 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org
 


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.7.0_09) - Build # 3026 - Failure!

2012-12-04 Thread Mark Miller

On Dec 4, 2012, at 4:18 PM, Steve Rowe sar...@gmail.com wrote:

 I'm running it.  It takes just over 4 minutes for me.  Pretty minor IMHO.

I suppose if you have long tests anyway - typically my solr tests run takes 
3:30 to 4 minutes, so doubling my time is not really something I'm ready to 
swallow.

- Mark


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3575) solr.xml should default to persist=true

2012-12-04 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3575?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13510183#comment-13510183
 ] 

Mark Miller commented on SOLR-3575:
---

We would still support loading and unloading.

bq. Also, what would happen if a core was unloaded?

The simple properties file or whatever in each core folder would indicate the 
core was not 'loaded'. Or one of many other solutions.

 solr.xml should default to persist=true
 ---

 Key: SOLR-3575
 URL: https://issues.apache.org/jira/browse/SOLR-3575
 Project: Solr
  Issue Type: Improvement
Reporter: Mark Miller
Assignee: Mark Miller
 Fix For: 4.0-BETA, 5.0


 The default of false is kind of silly IMO.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Solrj question

2012-12-04 Thread Chris Hostetter

: Believe there are no setParser method on CloudSolrServer?

I'm not sure why CloudSolrServer doesn't let you specify the parser ... i 
would suggest filing a feature request to make that possible.

Even assuming you have that, Erick's suggestion isn't quite what you 
want...

:  haven't tried this personally, but try this:
:  server.setParser(new XMLResponseParser());

...because that tells the SolrServer instance to use the XMLResponseParser 
to parse the data into a QueryResponse.

What you want is essentially to write your own ResponesParser that doesn't 
do any parsing, and just returns a pointer to the the InputStream it's 
given.

Assuming you have that, it may not matter that CloudSolrServer doesn't let 
you specify a default RequestParser, you *should* be able to just call 
SolrRequest.setResponseParser() to force your specific parser to be used 
(but i haven't sanity checked that works with CloudSolrServer)


-Hoss

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4043) Add ability to get success/failure responses from Collections API.

2012-12-04 Thread Raintung Li (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4043?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13510206#comment-13510206
 ] 

Raintung Li commented on SOLR-4043:
---

Mark, this patch is base on 4.0

 Add ability to get success/failure responses from Collections API.
 --

 Key: SOLR-4043
 URL: https://issues.apache.org/jira/browse/SOLR-4043
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Affects Versions: 4.0-ALPHA, 4.0-BETA, 4.0
 Environment: Solr cloud cluster
Reporter: Raintung Li
Assignee: Mark Miller
 Fix For: 4.1, 5.0

 Attachments: patch-4043.txt


 The create/delete/reload collections are asynchronous process, the client 
 can't get the right response, only make sure the information have been saved 
 into the OverseerCollectionQueue. The client will get the response directly 
 that don't wait the result of behavior(create/delete/reload collection) 
 whatever successful. 
 The easy solution is client wait until the asynchronous process success, the 
 create/delete/reload collection thread will save the response into 
 OverseerCollectionQueue, then notify client to get the response. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-1581) Facet by Function

2012-12-04 Thread Ted Strauss (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1581?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13510230#comment-13510230
 ] 

Ted Strauss commented on SOLR-1581:
---

I'd like to offer to co-sponsor work on this issue if anyone here knows where 
to go with it.
This relates to my issue https://issues.apache.org/jira/browse/SOLR-3742
Thanks

 Facet by Function
 -

 Key: SOLR-1581
 URL: https://issues.apache.org/jira/browse/SOLR-1581
 Project: Solr
  Issue Type: New Feature
Reporter: Grant Ingersoll
 Fix For: 4.1

 Attachments: SOLR-1581.patch


 It would be really great if we could execute a function and quantize it into 
 buckets that could then be returned as facets.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-4.x-Windows (32bit/jdk1.7.0_09) - Build # 2050 - Failure!

2012-12-04 Thread Policeman Jenkins Server
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows/2050/
Java: 32bit/jdk1.7.0_09 -server -XX:+UseG1GC

All tests passed

Build Log:
[...truncated 28804 lines...]
BUILD FAILED
C:\Users\JenkinsSlave\workspace\Lucene-Solr-4.x-Windows\build.xml:60: The 
following error occurred while executing this line:
C:\Users\JenkinsSlave\workspace\Lucene-Solr-4.x-Windows\solr\build.xml:499: The 
following error occurred while executing this line:
C:\Users\JenkinsSlave\workspace\Lucene-Solr-4.x-Windows\lucene\common-build.xml:1961:
 Can't get https://issues.apache.org/jira/rest/api/2/project/SOLR to 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-4.x-Windows\solr\build\docs\changes\jiraVersionList.json

Total time: 58 minutes 55 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 32bit/jdk1.7.0_09 -server -XX:+UseG1GC
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org