Re: Two hour recovery in non-replica setup?

2013-08-15 Thread Per Steffensen

Hi

Thanks, Mark. I am pretty sure SOLR-4899 is not it, but our problem 
might we solved after 4.0.0 anyway. We will consider if we will upgrade 
to latest 4.x and/or if we will dive into the problem to see if we can 
figure out what happens.


Regards, Per Steffensen

On 8/14/13 4:13 PM, Mark Miller wrote:

There have been hundreds of fixes since 4.0 - many around zookeeper 
integration. Really hard to say what it might be - I can say nothing should 
take that long by design.

A pretty important somewhat recent bug found was 
https://issues.apache.org/jira/browse/SOLR-4899 When reconnecting after 
ZooKeeper expiration, we need to be willing to wait forever, not for 30 seconds.

Don't know that it has anything to do with what you are seeing, or if perhaps 
that didn't exist in 4.0 - but worth looking at.

- Mark

On Aug 14, 2013, at 9:10 AM, Per Steffensen st...@designware.dk wrote:


Hi

We have a fairly large SolrCloud installation where we continuously index a lot 
of documents. Users are doing searches against the system from time to time.

 From time to time our Solrs lose their Zookeeper connection. Guess thats what happens. But it 
takes two hours before a Solr that loses its ZK connection has its shards active again. I cant 
imagine what it needs to do for two hours? We are not using replica, so synch of data among replica 
is not it. At the time we havnt dived into the problem, and we do not know there the time is spent, 
whether it is between the connection is lost and the Solr realizes it, or whether it is 
from when it realizes it and until the shards are declared active again.

We will dive into the problem, but before doing that I wanted to ask here if 
this is a known problem, and, if yes, whether or not it is solved? FYI we are 
using Solr 4.0.0.

Regards, Per Steffensen

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org





-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5056) Further clean up of ConfigSolr interface and CoreContainer construction

2013-08-15 Thread David Smiley (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5056?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Smiley updated SOLR-5056:
---

Fix Version/s: 4.5

 Further clean up of ConfigSolr interface and CoreContainer construction
 ---

 Key: SOLR-5056
 URL: https://issues.apache.org/jira/browse/SOLR-5056
 Project: Solr
  Issue Type: Improvement
Reporter: Alan Woodward
Assignee: Alan Woodward
Priority: Minor
 Fix For: 4.5

 Attachments: SOLR-5056.patch, SOLR-5056.patch


 Makes ConfigSolr a bit more typesafe, and pushes a bunch of cloud-specific 
 config into ZkController.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5176) remove _TestUtil.keepFullyDeletedSegments

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5176?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740772#comment-13740772
 ] 

ASF subversion and git services commented on LUCENE-5176:
-

Commit 1514184 from [~thetaphi] in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1514184 ]

Merged revision(s) 1514183 from lucene/dev/trunk:
LUCENE-5176: Reuse new AllDeletedFilterReader in TestAddIndexes

 remove _TestUtil.keepFullyDeletedSegments
 -

 Key: LUCENE-5176
 URL: https://issues.apache.org/jira/browse/LUCENE-5176
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5176.patch, LUCENE-5176.patch


 The few tests that need to test something here can use the index-writer 
 pkg-private method.
 Other tests (like QueryUtils) can use a FilterAtomicReader to do their crazy 
 tests.
 Otherwise, we risk lots of stupid false test failures.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5176) remove _TestUtil.keepFullyDeletedSegments

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5176?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740769#comment-13740769
 ] 

ASF subversion and git services commented on LUCENE-5176:
-

Commit 1514183 from [~thetaphi] in branch 'dev/trunk'
[ https://svn.apache.org/r1514183 ]

LUCENE-5176: Reuse new AllDeletedFilterReader in TestAddIndexes

 remove _TestUtil.keepFullyDeletedSegments
 -

 Key: LUCENE-5176
 URL: https://issues.apache.org/jira/browse/LUCENE-5176
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5176.patch, LUCENE-5176.patch


 The few tests that need to test something here can use the index-writer 
 pkg-private method.
 Other tests (like QueryUtils) can use a FilterAtomicReader to do their crazy 
 tests.
 Otherwise, we risk lots of stupid false test failures.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Problem threading JIRA emails in thunderbird?

2013-08-15 Thread Shawn Heisey
I'm wondering if anyone else is using Thunderbird and having problems
with incorrect threading of emails from Jira.  A few months ago I filed
a bug with mozilla that has a screenshot and header info, but nothing's
happened with it:

https://bugzilla.mozilla.org/show_bug.cgi?id=864545

If anyone is seeing similar behavior, could you please update the bug to
let Mozilla know?  If you happen to know of a solution, I'd love to hear it.

Thanks,
Shawn

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740829#comment-13740829
 ] 

Dawid Weiss edited comment on LUCENE-5168 at 8/15/13 9:21 AM:
--

Yeah, I agree. It isn't a trivial single-method issue though. nextSlice is 
indeed the problem, but it's inlined heavily in a number of places (and 
probably optimized wherever it's inlined). I've experimented with the following:
- -XX:-Inline - no errors
- excluding the parent method from compilation (which seems to be the root of 
the problem): org/apache/lucene/index/FreqProxTermsWriterPerField,flush - no 
errors
- -XX:-DoEscapeAnalysis - preventing escape analysis - no errors
- enforcing value flush in ByteSliceReader.nextSlice by modifying:
{code}
buffer  = pool.buffers[bufferUpto];
upto = nextIndex  ByteBlockPool.BYTE_BLOCK_MASK;
{code}
to
{code}
buffer  = pool.buffers[bufferUpto];
foo = upto = nextIndex  ByteBlockPool.BYTE_BLOCK_MASK;
{code}
where foo is defined as a static field with no other accesses:
{code}
static int foo;
{code}
This also results in no errors.

I also dumped the generated assembly for the flush() method (you have to dump 
the parent method because of the inlines) but it's huge and, sigh, it's 
sometimes different even for two runs with identical options. I'm guessing 
parallel compilation tasks hit different stats and hence the difference.


  was (Author: dweiss):
Yeah, I agree. It isn't a trivial single-method issue though. nextSlice is 
indeed the problem, but it's inlined heavily in a number of places (and 
probably optimized wherever it's inlined). I've experimented with the following:
- -XX:-Inline - no errors
- excluding the parent method from compilation (which seems to be the root of 
the problem): org/apache/lucene/index/FreqProxTermsWriterPerField,flush - no 
errors
- -XX:-DoEscapeAnalysis - preventing escape analysis - no errors
- enforcing value flush in ByteSliceReader.nextSlice by modifying:
{code}
buffer  = pool.buffers[bufferUpto];
upto = nextIndex  ByteBlockPool.BYTE_BLOCK_MASK;
{code}
to
{code}
buffer  = pool.buffers[bufferUpto];
foo = upto = nextIndex  ByteBlockPool.BYTE_BLOCK_MASK;
{code}
where foo is defined as a static field with no other accesses:
{code}
static int foo;
{code}

I also dumped the generated assembly for the flush() method (you have to dump 
the parent method because of the inlines) but it's huge and, sigh, it's 
sometimes different even for two runs with identical options. I'm guessing 
parallel compilation tasks hit different stats and hence the difference.

  
 ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC
 ---

 Key: LUCENE-5168
 URL: https://issues.apache.org/jira/browse/LUCENE-5168
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Attachments: java8-windows-4x-3075-console.txt


 This assertion trips (sometimes from different tests), if you run the 
 highlighting tests on branch_4x with r1512807.
 It reproduces about half the time, always only with 32bit + G1GC (other 
 combinations do not seem to trip it, i didnt try looping or anything really 
 though).
 {noformat}
 rmuir@beast:~/workspace/branch_4x$ svn up -r 1512807
 rmuir@beast:~/workspace/branch_4x$ ant clean
 rmuir@beast:~/workspace/branch_4x$ rm -rf .caches #this is important,
 otherwise master seed does not work!
 rmuir@beast:~/workspace/branch_4x/lucene/highlighter$ ant test
 -Dtests.jvms=2 -Dtests.seed=EBBFA6F4E80A7365 -Dargs=-server
 -XX:+UseG1GC
 {noformat}
 Originally showed up like this:
 {noformat}
 Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6874/
 Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC
 1 tests failed.
 REGRESSION:  
 org.apache.lucene.search.postingshighlight.TestPostingsHighlighter.testUserFailedToIndexOffsets
 Error Message:
 Stack Trace:
 java.lang.AssertionError
 at 
 __randomizedtesting.SeedInfo.seed([EBBFA6F4E80A7365:1FBF811885F2D611]:0)
 at 
 org.apache.lucene.index.ByteSliceReader.readByte(ByteSliceReader.java:73)
 at org.apache.lucene.store.DataInput.readVInt(DataInput.java:108)
 at 
 org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:453)
 at 
 org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
 at org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
 at org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
 at 
 org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
 at 
 org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, 

[jira] [Commented] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740829#comment-13740829
 ] 

Dawid Weiss commented on LUCENE-5168:
-

Yeah, I agree. It isn't a trivial single-method issue though. nextSlice is 
indeed the problem, but it's inlined heavily in a number of places (and 
probably optimized wherever it's inlined). I've experimented with the following:
- -XX:-Inline - no errors
- excluding the parent method from compilation (which seems to be the root of 
the problem): org/apache/lucene/index/FreqProxTermsWriterPerField,flush - no 
errors
- -XX:-DoEscapeAnalysis - preventing escape analysis - no errors
- enforcing value flush in ByteSliceReader.nextSlice by modifying:
{code}
buffer  = pool.buffers[bufferUpto];
upto = nextIndex  ByteBlockPool.BYTE_BLOCK_MASK;
{code}
to
{code}
buffer  = pool.buffers[bufferUpto];
foo = upto = nextIndex  ByteBlockPool.BYTE_BLOCK_MASK;
{code}
where foo is defined as a static field with no other accesses:
{code}
static int foo;
{code}

I also dumped the generated assembly for the flush() method (you have to dump 
the parent method because of the inlines) but it's huge and, sigh, it's 
sometimes different even for two runs with identical options. I'm guessing 
parallel compilation tasks hit different stats and hence the difference.


 ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC
 ---

 Key: LUCENE-5168
 URL: https://issues.apache.org/jira/browse/LUCENE-5168
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Attachments: java8-windows-4x-3075-console.txt


 This assertion trips (sometimes from different tests), if you run the 
 highlighting tests on branch_4x with r1512807.
 It reproduces about half the time, always only with 32bit + G1GC (other 
 combinations do not seem to trip it, i didnt try looping or anything really 
 though).
 {noformat}
 rmuir@beast:~/workspace/branch_4x$ svn up -r 1512807
 rmuir@beast:~/workspace/branch_4x$ ant clean
 rmuir@beast:~/workspace/branch_4x$ rm -rf .caches #this is important,
 otherwise master seed does not work!
 rmuir@beast:~/workspace/branch_4x/lucene/highlighter$ ant test
 -Dtests.jvms=2 -Dtests.seed=EBBFA6F4E80A7365 -Dargs=-server
 -XX:+UseG1GC
 {noformat}
 Originally showed up like this:
 {noformat}
 Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6874/
 Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC
 1 tests failed.
 REGRESSION:  
 org.apache.lucene.search.postingshighlight.TestPostingsHighlighter.testUserFailedToIndexOffsets
 Error Message:
 Stack Trace:
 java.lang.AssertionError
 at 
 __randomizedtesting.SeedInfo.seed([EBBFA6F4E80A7365:1FBF811885F2D611]:0)
 at 
 org.apache.lucene.index.ByteSliceReader.readByte(ByteSliceReader.java:73)
 at org.apache.lucene.store.DataInput.readVInt(DataInput.java:108)
 at 
 org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:453)
 at 
 org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
 at org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
 at org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
 at 
 org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
 at 
 org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740832#comment-13740832
 ] 

Uwe Schindler commented on LUCENE-5168:
---

I am glad it is not {{DataInput#readVInt()}} :-)

 ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC
 ---

 Key: LUCENE-5168
 URL: https://issues.apache.org/jira/browse/LUCENE-5168
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Attachments: java8-windows-4x-3075-console.txt


 This assertion trips (sometimes from different tests), if you run the 
 highlighting tests on branch_4x with r1512807.
 It reproduces about half the time, always only with 32bit + G1GC (other 
 combinations do not seem to trip it, i didnt try looping or anything really 
 though).
 {noformat}
 rmuir@beast:~/workspace/branch_4x$ svn up -r 1512807
 rmuir@beast:~/workspace/branch_4x$ ant clean
 rmuir@beast:~/workspace/branch_4x$ rm -rf .caches #this is important,
 otherwise master seed does not work!
 rmuir@beast:~/workspace/branch_4x/lucene/highlighter$ ant test
 -Dtests.jvms=2 -Dtests.seed=EBBFA6F4E80A7365 -Dargs=-server
 -XX:+UseG1GC
 {noformat}
 Originally showed up like this:
 {noformat}
 Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6874/
 Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC
 1 tests failed.
 REGRESSION:  
 org.apache.lucene.search.postingshighlight.TestPostingsHighlighter.testUserFailedToIndexOffsets
 Error Message:
 Stack Trace:
 java.lang.AssertionError
 at 
 __randomizedtesting.SeedInfo.seed([EBBFA6F4E80A7365:1FBF811885F2D611]:0)
 at 
 org.apache.lucene.index.ByteSliceReader.readByte(ByteSliceReader.java:73)
 at org.apache.lucene.store.DataInput.readVInt(DataInput.java:108)
 at 
 org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:453)
 at 
 org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
 at org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
 at org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
 at 
 org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
 at 
 org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740840#comment-13740840
 ] 

Dawid Weiss commented on LUCENE-5168:
-

Don't be so sure. :) FreqProxTermsWriterPerField::flush inlines half of the 
world (and is compiled multiple times, in fact). readVInt *is* in fact inlined 
there (but it doesn't seem to be the root of the problem to me).
{code}
   8206 1158 
org.apache.lucene.index.FreqProxTermsWriterPerField::flush (1207 bytes)
@ 4   org.apache.lucene.index.FieldInfo::isIndexed 
(5 bytes)   inline (hot)
@ 16   
org.apache.lucene.codecs.perfield.PerFieldPostingsFormat$FieldsWriter::addField 
(327 bytes)   too big
@ 16   
org.apache.lucene.codecs.BlockTreeTermsWriter::addField (53 bytes)   too big
@ 23   
org.apache.lucene.codecs.BlockTreeTermsWriter$TermsWriter::getComparator (4 
bytes)   inline (hot)
@ 23   
org.apache.lucene.codecs.blockterms.BlockTermsWriter$TermsWriter::getComparator 
(4 bytes)   inline (hot)
  @ 0   
org.apache.lucene.util.BytesRef::getUTF8SortedAsUnicodeComparator (4 bytes)   
inline (hot)
  @ 0   
org.apache.lucene.util.BytesRef::getUTF8SortedAsUnicodeComparator (4 bytes)   
inline (hot)
@ 32   
org.apache.lucene.index.FieldInfo::getIndexOptions (5 bytes)   inline (hot)
@ 61   java.lang.Enum::compareTo (44 bytes)   too 
big
@ 79   java.lang.Enum::compareTo (44 bytes)   too 
big
@ 97   java.lang.Enum::compareTo (44 bytes)   too 
big
@ 238   java.util.HashMap::size (5 bytes)   inline 
(hot)
@ 267   
org.apache.lucene.index.TermsHashPerField::sortPostings (9 bytes)   executed  
MinInliningThreshold times
@ 279   org.apache.lucene.util.BytesRefHash::size 
(5 bytes)   inline (hot)
@ 288   org.apache.lucene.util.BytesRef::init (8 
bytes)   inline (hot)
  @ 4   org.apache.lucene.util.BytesRef::init (9 
bytes)   inline (hot)
@ 5   org.apache.lucene.util.BytesRef::init 
(41 bytes)   inline (hot)
  @ 1   java.lang.Object::init (1 bytes)   
inline (hot)
  @ 26   
org.apache.lucene.util.BytesRef::isValid (329 bytes)   hot method too big
@ 309   
org.apache.lucene.index.ByteSliceReader::init (5 bytes)   inline (hot)
  @ 1   org.apache.lucene.store.DataInput::init 
(5 bytes)   inline (hot)
@ 1   java.lang.Object::init (1 bytes)   
inline (hot)
@ 318   
org.apache.lucene.index.ByteSliceReader::init (5 bytes)   inline (hot)
  @ 1   org.apache.lucene.store.DataInput::init 
(5 bytes)   inline (hot)
@ 1   java.lang.Object::init (1 bytes)   
inline (hot)
@ 331   
org.apache.lucene.index.SegmentInfo::getDocCount (23 bytes)   inline (hot)
@ 334   org.apache.lucene.util.FixedBitSet::init 
(29 bytes)   inline (hot)
  @ 1   org.apache.lucene.search.DocIdSet::init 
(5 bytes)   inline (hot)
@ 1   java.lang.Object::init (1 bytes)   
inline (hot)
  @ 11   
org.apache.lucene.util.FixedBitSet::bits2words (17 bytes)   inline (hot)
@ 350   org.apache.lucene.index.Term::init (13 
bytes)   inline (hot)
  @ 6   org.apache.lucene.util.BytesRef::init (8 
bytes)   inline (hot)
@ 4   org.apache.lucene.util.BytesRef::init 
(9 bytes)   inline (hot)
  @ 5   org.apache.lucene.util.BytesRef::init 
(41 bytes)   inline (hot)
@ 1   java.lang.Object::init (1 bytes)   
inline (hot)
@ 26   
org.apache.lucene.util.BytesRef::isValid (329 bytes)   hot method too big
  @ 9   org.apache.lucene.index.Term::init (15 
bytes)   inline (hot)
@ 1   java.lang.Object::init (1 bytes)   
inline (hot)
@ 393   
org.apache.lucene.util.ByteBlockPool::setBytesRef (107 bytes)   inline (hot)
@ 405   
org.apache.lucene.index.TermsHashPerField::initReader (87 bytes)   inline (hot)
  @ 83   
org.apache.lucene.index.ByteSliceReader::init (153 bytes)   inline (hot)
@ 427   

[jira] [Commented] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740847#comment-13740847
 ] 

Uwe Schindler commented on LUCENE-5168:
---

What happens if you make upto volatile? I assume the same happens like with 
the static dummy field.

Maybe for now we use the static field hack. I don't think this causes 
performance trouble. We can add a good comment and once the bug is fixed in JDK 
we can remove it in 5 years or so :-)

To me it looks like we hit this bug more often because of optimized tests that 
seem to run this method more often. This is a good sign, we are improving our 
tests! :-)

 ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC
 ---

 Key: LUCENE-5168
 URL: https://issues.apache.org/jira/browse/LUCENE-5168
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Attachments: java8-windows-4x-3075-console.txt


 This assertion trips (sometimes from different tests), if you run the 
 highlighting tests on branch_4x with r1512807.
 It reproduces about half the time, always only with 32bit + G1GC (other 
 combinations do not seem to trip it, i didnt try looping or anything really 
 though).
 {noformat}
 rmuir@beast:~/workspace/branch_4x$ svn up -r 1512807
 rmuir@beast:~/workspace/branch_4x$ ant clean
 rmuir@beast:~/workspace/branch_4x$ rm -rf .caches #this is important,
 otherwise master seed does not work!
 rmuir@beast:~/workspace/branch_4x/lucene/highlighter$ ant test
 -Dtests.jvms=2 -Dtests.seed=EBBFA6F4E80A7365 -Dargs=-server
 -XX:+UseG1GC
 {noformat}
 Originally showed up like this:
 {noformat}
 Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6874/
 Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC
 1 tests failed.
 REGRESSION:  
 org.apache.lucene.search.postingshighlight.TestPostingsHighlighter.testUserFailedToIndexOffsets
 Error Message:
 Stack Trace:
 java.lang.AssertionError
 at 
 __randomizedtesting.SeedInfo.seed([EBBFA6F4E80A7365:1FBF811885F2D611]:0)
 at 
 org.apache.lucene.index.ByteSliceReader.readByte(ByteSliceReader.java:73)
 at org.apache.lucene.store.DataInput.readVInt(DataInput.java:108)
 at 
 org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:453)
 at 
 org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
 at org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
 at org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
 at 
 org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
 at 
 org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5158) PostingsHighlighter - optionally return unhighlighted fields

2013-08-15 Thread Stewart Sims (JIRA)
Stewart Sims created SOLR-5158:
--

 Summary: PostingsHighlighter - optionally return unhighlighted 
fields
 Key: SOLR-5158
 URL: https://issues.apache.org/jira/browse/SOLR-5158
 Project: Solr
  Issue Type: Improvement
  Components: highlighter
Affects Versions: 4.2.1
Reporter: Stewart Sims
Priority: Minor


The PostingsHighlighter mechanism currently returns all fields specified in the 
hl.fl parameter for each document, even if they do not have hits within them. 
This can lead to a lot of empty / self-closing tags if highlighting on a large 
number of fields which may be undesirable.

It would be useful if a user could specify using a parameter such as 
'hl.ignoreUnhighlightedFields=true' that these fields are not returned in the 
response.

See 
http://lucene.472066.n3.nabble.com/PostingsHighlighter-returning-fields-which-don-t-match-td4084495.html
 for a suggestions on implementation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740851#comment-13740851
 ] 

Dawid Weiss commented on LUCENE-5168:
-

Yeah... we could temporarily commit the static field hack and at least see if 
this proves me wrong (perhaps it just happens to work for this particular seed 
but will fail elsewhere).

 ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC
 ---

 Key: LUCENE-5168
 URL: https://issues.apache.org/jira/browse/LUCENE-5168
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Attachments: java8-windows-4x-3075-console.txt


 This assertion trips (sometimes from different tests), if you run the 
 highlighting tests on branch_4x with r1512807.
 It reproduces about half the time, always only with 32bit + G1GC (other 
 combinations do not seem to trip it, i didnt try looping or anything really 
 though).
 {noformat}
 rmuir@beast:~/workspace/branch_4x$ svn up -r 1512807
 rmuir@beast:~/workspace/branch_4x$ ant clean
 rmuir@beast:~/workspace/branch_4x$ rm -rf .caches #this is important,
 otherwise master seed does not work!
 rmuir@beast:~/workspace/branch_4x/lucene/highlighter$ ant test
 -Dtests.jvms=2 -Dtests.seed=EBBFA6F4E80A7365 -Dargs=-server
 -XX:+UseG1GC
 {noformat}
 Originally showed up like this:
 {noformat}
 Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6874/
 Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC
 1 tests failed.
 REGRESSION:  
 org.apache.lucene.search.postingshighlight.TestPostingsHighlighter.testUserFailedToIndexOffsets
 Error Message:
 Stack Trace:
 java.lang.AssertionError
 at 
 __randomizedtesting.SeedInfo.seed([EBBFA6F4E80A7365:1FBF811885F2D611]:0)
 at 
 org.apache.lucene.index.ByteSliceReader.readByte(ByteSliceReader.java:73)
 at org.apache.lucene.store.DataInput.readVInt(DataInput.java:108)
 at 
 org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:453)
 at 
 org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
 at org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
 at org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
 at 
 org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
 at 
 org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5149) Query facet to respect mincount

2013-08-15 Thread Markus Jelsma (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5149?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740912#comment-13740912
 ] 

Markus Jelsma commented on SOLR-5149:
-

Yes, that makes sense. But we definately have some cases in which we would like 
to omit them from the result so maybe having a facet.querymincount would be a 
good idea?

Thanks

 Query facet to respect mincount
 ---

 Key: SOLR-5149
 URL: https://issues.apache.org/jira/browse/SOLR-5149
 Project: Solr
  Issue Type: Bug
  Components: SearchComponents - other
Affects Versions: 4.4
Reporter: Markus Jelsma
Priority: Minor
 Fix For: 4.5, 5.0

 Attachments: SOLR-5149-trunk.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Manuel Amoabeng (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manuel Amoabeng updated LUCENE-5166:


Attachment: LUCENE-5166-revisited.patch

 PostingsHighlighter fails with IndexOutOfBoundsException
 

 Key: LUCENE-5166
 URL: https://issues.apache.org/jira/browse/LUCENE-5166
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/highlighter
Affects Versions: 4.4
Reporter: Manuel Amoabeng
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166-revisited.patch


 Given a document with a match at a startIndex  PostingsHighlighter.maxlength 
 and an endIndex  PostingsHighlighter.maxLength, DefaultPassageFormatter will 
 throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
 invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Manuel Amoabeng (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740925#comment-13740925
 ] 

Manuel Amoabeng commented on LUCENE-5166:
-

I just found another problem here: If we have both, matches that do and matches 
that don't span the content boundary the formatter is asked to highlight the 
spanning match.
Please find attached additional tests and a possible fix for this. 

 PostingsHighlighter fails with IndexOutOfBoundsException
 

 Key: LUCENE-5166
 URL: https://issues.apache.org/jira/browse/LUCENE-5166
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/highlighter
Affects Versions: 4.4
Reporter: Manuel Amoabeng
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166-revisited.patch


 Given a document with a match at a startIndex  PostingsHighlighter.maxlength 
 and an endIndex  PostingsHighlighter.maxLength, DefaultPassageFormatter will 
 throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
 invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Reopened] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Manuel Amoabeng (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manuel Amoabeng reopened LUCENE-5166:
-


 PostingsHighlighter fails with IndexOutOfBoundsException
 

 Key: LUCENE-5166
 URL: https://issues.apache.org/jira/browse/LUCENE-5166
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/highlighter
Affects Versions: 4.4
Reporter: Manuel Amoabeng
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166-revisited.patch


 Given a document with a match at a startIndex  PostingsHighlighter.maxlength 
 and an endIndex  PostingsHighlighter.maxLength, DefaultPassageFormatter will 
 throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
 invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3069) Lucene should have an entirely memory resident term dictionary

2013-08-15 Thread Han Jiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3069?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Han Jiang updated LUCENE-3069:
--

Attachment: LUCENE-3069.patch

Patch, update BlockTerms dict so that it follows refactored API.

 Lucene should have an entirely memory resident term dictionary
 --

 Key: LUCENE-3069
 URL: https://issues.apache.org/jira/browse/LUCENE-3069
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index, core/search
Affects Versions: 4.0-ALPHA
Reporter: Simon Willnauer
Assignee: Han Jiang
  Labels: gsoc2013
 Fix For: 5.0, 4.5

 Attachments: df-ttf-estimate.txt, example.png, LUCENE-3069.patch, 
 LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, 
 LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, 
 LUCENE-3069.patch


 FST based TermDictionary has been a great improvement yet it still uses a 
 delta codec file for scanning to terms. Some environments have enough memory 
 available to keep the entire FST based term dict in memory. We should add a 
 TermDictionary implementation that encodes all needed information for each 
 term into the FST (custom fst.Output) and builds a FST from the entire term 
 not just the delta.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3069) Lucene should have an entirely memory resident term dictionary

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740948#comment-13740948
 ] 

ASF subversion and git services commented on LUCENE-3069:
-

Commit 1514253 from [~billy] in branch 'dev/branches/lucene3069'
[ https://svn.apache.org/r1514253 ]

LUCENE-3069: API refactoring on BlockTerms dict

 Lucene should have an entirely memory resident term dictionary
 --

 Key: LUCENE-3069
 URL: https://issues.apache.org/jira/browse/LUCENE-3069
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index, core/search
Affects Versions: 4.0-ALPHA
Reporter: Simon Willnauer
Assignee: Han Jiang
  Labels: gsoc2013
 Fix For: 5.0, 4.5

 Attachments: df-ttf-estimate.txt, example.png, LUCENE-3069.patch, 
 LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, 
 LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, LUCENE-3069.patch, 
 LUCENE-3069.patch


 FST based TermDictionary has been a great improvement yet it still uses a 
 delta codec file for scanning to terms. Some environments have enough memory 
 available to keep the entire FST based term dict in memory. We should add a 
 TermDictionary implementation that encodes all needed information for each 
 term into the FST (custom fst.Output) and builds a FST from the entire term 
 not just the delta.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740969#comment-13740969
 ] 

Dawid Weiss edited comment on LUCENE-5168 at 8/15/13 1:49 PM:
--

I've built a fastdebug binary of openjdk 1.8 (top of the master branch). 
Retrying my 1000-random seed sample right now (1.7's seed doesn't reproduce on 
my fresh-from-the-oven binary, so perhaps it's been fixed in between).

  was (Author: dweiss):
I've built a fastdebug binary of Java 1.8. Retrying my 1000-random seed 
sample right now (1.7's seed doesn't reproduce on my fresh-from-the-oven 
binary, so perhaps it's been fixed in between).
  
 ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC
 ---

 Key: LUCENE-5168
 URL: https://issues.apache.org/jira/browse/LUCENE-5168
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Attachments: java8-windows-4x-3075-console.txt


 This assertion trips (sometimes from different tests), if you run the 
 highlighting tests on branch_4x with r1512807.
 It reproduces about half the time, always only with 32bit + G1GC (other 
 combinations do not seem to trip it, i didnt try looping or anything really 
 though).
 {noformat}
 rmuir@beast:~/workspace/branch_4x$ svn up -r 1512807
 rmuir@beast:~/workspace/branch_4x$ ant clean
 rmuir@beast:~/workspace/branch_4x$ rm -rf .caches #this is important,
 otherwise master seed does not work!
 rmuir@beast:~/workspace/branch_4x/lucene/highlighter$ ant test
 -Dtests.jvms=2 -Dtests.seed=EBBFA6F4E80A7365 -Dargs=-server
 -XX:+UseG1GC
 {noformat}
 Originally showed up like this:
 {noformat}
 Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6874/
 Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC
 1 tests failed.
 REGRESSION:  
 org.apache.lucene.search.postingshighlight.TestPostingsHighlighter.testUserFailedToIndexOffsets
 Error Message:
 Stack Trace:
 java.lang.AssertionError
 at 
 __randomizedtesting.SeedInfo.seed([EBBFA6F4E80A7365:1FBF811885F2D611]:0)
 at 
 org.apache.lucene.index.ByteSliceReader.readByte(ByteSliceReader.java:73)
 at org.apache.lucene.store.DataInput.readVInt(DataInput.java:108)
 at 
 org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:453)
 at 
 org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
 at org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
 at org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
 at 
 org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
 at 
 org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5168) ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC

2013-08-15 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740969#comment-13740969
 ] 

Dawid Weiss commented on LUCENE-5168:
-

I've built a fastdebug binary of Java 1.8. Retrying my 1000-random seed sample 
right now (1.7's seed doesn't reproduce on my fresh-from-the-oven binary, so 
perhaps it's been fixed in between).

 ByteSliceReader assert trips with 32-bit oracle 1.7.0_25 + G1GC
 ---

 Key: LUCENE-5168
 URL: https://issues.apache.org/jira/browse/LUCENE-5168
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Attachments: java8-windows-4x-3075-console.txt


 This assertion trips (sometimes from different tests), if you run the 
 highlighting tests on branch_4x with r1512807.
 It reproduces about half the time, always only with 32bit + G1GC (other 
 combinations do not seem to trip it, i didnt try looping or anything really 
 though).
 {noformat}
 rmuir@beast:~/workspace/branch_4x$ svn up -r 1512807
 rmuir@beast:~/workspace/branch_4x$ ant clean
 rmuir@beast:~/workspace/branch_4x$ rm -rf .caches #this is important,
 otherwise master seed does not work!
 rmuir@beast:~/workspace/branch_4x/lucene/highlighter$ ant test
 -Dtests.jvms=2 -Dtests.seed=EBBFA6F4E80A7365 -Dargs=-server
 -XX:+UseG1GC
 {noformat}
 Originally showed up like this:
 {noformat}
 Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6874/
 Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC
 1 tests failed.
 REGRESSION:  
 org.apache.lucene.search.postingshighlight.TestPostingsHighlighter.testUserFailedToIndexOffsets
 Error Message:
 Stack Trace:
 java.lang.AssertionError
 at 
 __randomizedtesting.SeedInfo.seed([EBBFA6F4E80A7365:1FBF811885F2D611]:0)
 at 
 org.apache.lucene.index.ByteSliceReader.readByte(ByteSliceReader.java:73)
 at org.apache.lucene.store.DataInput.readVInt(DataInput.java:108)
 at 
 org.apache.lucene.index.FreqProxTermsWriterPerField.flush(FreqProxTermsWriterPerField.java:453)
 at 
 org.apache.lucene.index.FreqProxTermsWriter.flush(FreqProxTermsWriter.java:85)
 at org.apache.lucene.index.TermsHash.flush(TermsHash.java:116)
 at org.apache.lucene.index.DocInverter.flush(DocInverter.java:53)
 at 
 org.apache.lucene.index.DocFieldProcessor.flush(DocFieldProcessor.java:81)
 at 
 org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:501)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740975#comment-13740975
 ] 

Robert Muir commented on LUCENE-5166:
-

Hi Manuel: thank you! Another bug, or a bug in my fix to the other bug :)

I'll investigate deeper in a bit.

 PostingsHighlighter fails with IndexOutOfBoundsException
 

 Key: LUCENE-5166
 URL: https://issues.apache.org/jira/browse/LUCENE-5166
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/highlighter
Affects Versions: 4.4
Reporter: Manuel Amoabeng
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166-revisited.patch


 Given a document with a match at a startIndex  PostingsHighlighter.maxlength 
 and an endIndex  PostingsHighlighter.maxLength, DefaultPassageFormatter will 
 throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
 invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to ArithmeticException: / by zero

2013-08-15 Thread James Dyer (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740999#comment-13740999
 ] 

James Dyer commented on SOLR-5122:
--

Hoss,

This looks reasonable to me.  The test is more forgiving to variations caused 
by random merges, no more integer division, etc.  I appreciate your working on 
this as I wouldn't have much more time until next week.  I think your method of 
estimating the hits would be at least as good of what I attempted to do before.

 spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
 to ArithmeticException: / by zero
 

 Key: SOLR-5122
 URL: https://issues.apache.org/jira/browse/SOLR-5122
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.4
Reporter: Hoss Man
Assignee: James Dyer
 Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch


 As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
 and this (aparently) led to a failure in testEstimatedHitCounts.
 As far as i can tell: the test assumes that specific values would be returned 
 as the _estimated_ hits for a colleation, and it appears that the change in 
 MergePolicy however resulted in different segments with different term stats, 
 causing the estimation code to produce different values then what is expected.
 I made a quick attempt to improve the test to:
  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
 set such that the estimate' should actually be exact (ie: 
 collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
 docs in the index
  * randomize the values used for collateMaxCollectDocs and confirm that the 
 estimates are never more then the num docs in the index
 This lead to an odd ArithmeticException: / by zero error in the test, which 
 seems to suggest that there is a genuine bug in the code for estimating the 
 hits that only gets tickled in certain 
 mergepolicy/segment/collateMaxCollectDocs combinations.
 *Update:* This appears to be a general problem with collecting docs out of 
 order and the estimation of hits -- i believe even if there is no divide by 
 zero error, the estimates are largely meaningless since the docs are 
 collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Artem Karpenko (JIRA)
Artem Karpenko created SOLR-5159:


 Summary: Manifest includes non-parsed maven variables
 Key: SOLR-5159
 URL: https://issues.apache.org/jira/browse/SOLR-5159
 Project: Solr
  Issue Type: Bug
  Components: Build
Affects Versions: 4.4, 4.5, 5.0
 Environment: Apache Maven 3.0.5
Reporter: Artem Karpenko
Priority: Minor


When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
into JAR artifacts contain non-parsed POM variables: namely, there are entries 
like

Specification-Version: 5.0.0.${now.version}

In the end, Solr displays these values on admin page in Versions section.

This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Artem Karpenko (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Karpenko updated SOLR-5159:
-

Attachment: SOLR-5159.patch

Patch moves declaration of Manifest headers from archive configuration of 
maven-jar-plugin into instructions section of maven-bundle-plugin, which 
latter plugin does parse for variables correctly.

 Manifest includes non-parsed maven variables
 

 Key: SOLR-5159
 URL: https://issues.apache.org/jira/browse/SOLR-5159
 Project: Solr
  Issue Type: Bug
  Components: Build
Affects Versions: 4.4, 4.5, 5.0
 Environment: Apache Maven 3.0.5
Reporter: Artem Karpenko
Priority: Minor
  Labels: maven-bundle-plugin, maven3,
 Attachments: SOLR-5159.patch


 When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
 into JAR artifacts contain non-parsed POM variables: namely, there are 
 entries like
 Specification-Version: 5.0.0.${now.version}
 In the end, Solr displays these values on admin page in Versions section.
 This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Artem Karpenko (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Artem Karpenko updated SOLR-5159:
-

Description: 
When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
into JAR artifacts contain non-parsed POM variables: namely, there are entries 
like

Specification-Version: 5.0.0.$\{now.version}

In the end, Solr displays these values on admin page in Versions section.

This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

  was:
When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
into JAR artifacts contain non-parsed POM variables: namely, there are entries 
like

Specification-Version: 5.0.0.${now.version}

In the end, Solr displays these values on admin page in Versions section.

This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 


 Manifest includes non-parsed maven variables
 

 Key: SOLR-5159
 URL: https://issues.apache.org/jira/browse/SOLR-5159
 Project: Solr
  Issue Type: Bug
  Components: Build
Affects Versions: 4.4, 4.5, 5.0
 Environment: Apache Maven 3.0.5
Reporter: Artem Karpenko
Priority: Minor
  Labels: maven-bundle-plugin, maven3,
 Attachments: SOLR-5159.patch


 When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
 into JAR artifacts contain non-parsed POM variables: namely, there are 
 entries like
 Specification-Version: 5.0.0.$\{now.version}
 In the end, Solr displays these values on admin page in Versions section.
 This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Artem Karpenko (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741061#comment-13741061
 ] 

Artem Karpenko commented on SOLR-5159:
--

I've checked maven 2 builds as well - after applying changes manifest files 
stay the same (correct).

 Manifest includes non-parsed maven variables
 

 Key: SOLR-5159
 URL: https://issues.apache.org/jira/browse/SOLR-5159
 Project: Solr
  Issue Type: Bug
  Components: Build
Affects Versions: 4.4, 4.5, 5.0
 Environment: Apache Maven 3.0.5
Reporter: Artem Karpenko
Priority: Minor
  Labels: maven-bundle-plugin, maven3,
 Attachments: SOLR-5159.patch


 When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
 into JAR artifacts contain non-parsed POM variables: namely, there are 
 entries like
 Specification-Version: 5.0.0.$\{now.version}
 In the end, Solr displays these values on admin page in Versions section.
 This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5149) Query facet to respect mincount

2013-08-15 Thread Erik Hatcher (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5149?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741112#comment-13741112
 ] 

Erik Hatcher commented on SOLR-5149:


I'm curious about the use cases - why not filter out zero counts in the client? 
  

I'm ok with another parameter, thought that feels odd too, but at least one 
would have the option to filter zeros from fields without filtering out 
facet.query's too.   facet.query.mincount?  (like facet.range.*)

 Query facet to respect mincount
 ---

 Key: SOLR-5149
 URL: https://issues.apache.org/jira/browse/SOLR-5149
 Project: Solr
  Issue Type: Bug
  Components: SearchComponents - other
Affects Versions: 4.4
Reporter: Markus Jelsma
Priority: Minor
 Fix For: 4.5, 5.0

 Attachments: SOLR-5149-trunk.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: MoreLikeThis (MLT) - AND operator between the fields

2013-08-15 Thread Kranti Parisa
I was looking at the code and found that it is hard coded to Occur.SHOULD
in MoreLikeThisQuery.

I customized the code to pass a new parameter *mlt.operator*=AND/OR  based
on that it computes the MLT documents. Default operator is set to OR.
And I also want to have *mlt.sort* option, So I will be trying for that as
well.

Do you guys think, we should make this part of the MLT feature?
Please share your ideas. I can submit this change.


Thanks  Regards,
Kranti K Parisa
http://www.linkedin.com/in/krantiparisa



On Thu, Aug 15, 2013 at 12:05 AM, Kranti Parisa kranti.par...@gmail.comwrote:

 Hi,

 It seems that when we pass multiple field names with mlt.fl parameter, it
 is ORing them to find the MLT documents.

 Is there a way to specify AND operator? Means if mlt.fl=language,year,
 then we should return back the MLT documents which has language AND year
 field values as same as the main query result document.


 http://localhost:8180/solr/mltCore/mlt?q=id:1wt=jsonmlt=truemlt.fl=language,yearfl=*,scoremlt.mindf=0mlt.mintf=0mlt.match.include=false

 The above query should return those documents whose field values
 (language, year) are exactly matching with the document id:1.

 Is this possible thru any config or param? If not, I think it's worth
 having as a feature because we don't know the values of those fields to
 apply as FQ.


 Thanks  Regards,
 Kranti K Parisa
 http://www.linkedin.com/in/krantiparisa




[jira] [Commented] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741140#comment-13741140
 ] 

Robert Muir commented on LUCENE-5166:
-

Manuel: your fix is correct, thank you.

To explain: I had totally forgotten about this little loop on tf within the 
passage (i had removed this optimization in LUCENE-4909, which didnt turn out 
to work that great, so wasn't committed).

We might at some point want to still just remove the optimization just based on 
the reason that it makes this thing more complicated, it was just intended to 
speed up the worst case (where someone has very common stopwords and stuff like 
that).

But for now to complete the bugfix, we should commit your patch 
(LUCENE-5166-revisited.patch).


 PostingsHighlighter fails with IndexOutOfBoundsException
 

 Key: LUCENE-5166
 URL: https://issues.apache.org/jira/browse/LUCENE-5166
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/highlighter
Affects Versions: 4.4
Reporter: Manuel Amoabeng
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166-revisited.patch


 Given a document with a match at a startIndex  PostingsHighlighter.maxlength 
 and an endIndex  PostingsHighlighter.maxLength, DefaultPassageFormatter will 
 throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
 invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5160) MoreLikeThis (MLT) - New Optional Params - mlt.operator mlt.sort

2013-08-15 Thread Kranti Parisa (JIRA)
Kranti Parisa created SOLR-5160:
---

 Summary: MoreLikeThis (MLT) - New Optional Params -  mlt.operator 
 mlt.sort
 Key: SOLR-5160
 URL: https://issues.apache.org/jira/browse/SOLR-5160
 Project: Solr
  Issue Type: Improvement
  Components: MoreLikeThis
Reporter: Kranti Parisa
Priority: Minor


It seems that when we pass multiple field names with mlt.fl parameter, it is 
ORing them to find the MLT documents.

Is there a way to specify AND operator? Means if mlt.fl=language,year, then we 
should return back the MLT documents which has language AND year field values 
as same as the main query result document.

http://localhost:8180/solr/mltCore/mlt?q=id:1wt=jsonmlt=truemlt.fl=language,yearfl=*,scoremlt.mindf=0mlt.mintf=0mlt.match.include=false

The above query should return those documents whose field values (language, 
year) are exactly matching with the document id:1.

Is this possible thru any config or param? If not, I think it's worth having as 
a feature because we don't know the values of those fields to apply as FQ. 


 was looking at the code and found that it is hard coded to Occur.SHOULD in 
MoreLikeThisQuery.

I customized the code to pass a new parameter mlt.operator=AND/OR  based on 
that it computes the MLT documents. Default operator is set to OR.
And I also want to have mlt.sort option, So I will be trying for that as well.

Do you guys think, we should make this part of the MLT feature? 
Please share your ideas. I can submit this change.



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741174#comment-13741174
 ] 

ASF subversion and git services commented on LUCENE-5166:
-

Commit 1514367 from [~rcmuir] in branch 'dev/trunk'
[ https://svn.apache.org/r1514367 ]

LUCENE-5166: also fix and test this case where tf  1 within the passage for a 
term

 PostingsHighlighter fails with IndexOutOfBoundsException
 

 Key: LUCENE-5166
 URL: https://issues.apache.org/jira/browse/LUCENE-5166
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/highlighter
Affects Versions: 4.4
Reporter: Manuel Amoabeng
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166-revisited.patch


 Given a document with a match at a startIndex  PostingsHighlighter.maxlength 
 and an endIndex  PostingsHighlighter.maxLength, DefaultPassageFormatter will 
 throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
 invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5160) MoreLikeThis (MLT) - New Optional Params - mlt.operator mlt.sort

2013-08-15 Thread Jack Krupansky (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5160?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741172#comment-13741172
 ] 

Jack Krupansky commented on SOLR-5160:
--

In truth, a more-like-this feature should have two modes: 1) Discovery mode, to 
broaden the search or increase recall (OR), and 2) Refinement mode, to narrow 
the search or increase precision (AND).


 MoreLikeThis (MLT) - New Optional Params -  mlt.operator  mlt.sort
 ---

 Key: SOLR-5160
 URL: https://issues.apache.org/jira/browse/SOLR-5160
 Project: Solr
  Issue Type: Improvement
  Components: MoreLikeThis
Reporter: Kranti Parisa
Priority: Minor
  Labels: features
   Original Estimate: 24h
  Remaining Estimate: 24h

 It seems that when we pass multiple field names with mlt.fl parameter, it is 
 ORing them to find the MLT documents.
 Is there a way to specify AND operator? Means if mlt.fl=language,year, then 
 we should return back the MLT documents which has language AND year field 
 values as same as the main query result document.
 http://localhost:8180/solr/mltCore/mlt?q=id:1wt=jsonmlt=truemlt.fl=language,yearfl=*,scoremlt.mindf=0mlt.mintf=0mlt.match.include=false
 The above query should return those documents whose field values (language, 
 year) are exactly matching with the document id:1.
 Is this possible thru any config or param? If not, I think it's worth having 
 as a feature because we don't know the values of those fields to apply as FQ. 
  was looking at the code and found that it is hard coded to Occur.SHOULD in 
 MoreLikeThisQuery.
 I customized the code to pass a new parameter mlt.operator=AND/OR  based on 
 that it computes the MLT documents. Default operator is set to OR.
 And I also want to have mlt.sort option, So I will be trying for that as well.
 Do you guys think, we should make this part of the MLT feature? 
 Please share your ideas. I can submit this change.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Steve Rowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Rowe reassigned SOLR-5159:


Assignee: Steve Rowe

 Manifest includes non-parsed maven variables
 

 Key: SOLR-5159
 URL: https://issues.apache.org/jira/browse/SOLR-5159
 Project: Solr
  Issue Type: Bug
  Components: Build
Affects Versions: 4.4, 4.5, 5.0
 Environment: Apache Maven 3.0.5
Reporter: Artem Karpenko
Assignee: Steve Rowe
Priority: Minor
  Labels: maven-bundle-plugin, maven3,
 Attachments: SOLR-5159.patch


 When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
 into JAR artifacts contain non-parsed POM variables: namely, there are 
 entries like
 Specification-Version: 5.0.0.$\{now.version}
 In the end, Solr displays these values on admin page in Versions section.
 This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5159) Manifest includes non-parsed maven variables

2013-08-15 Thread Steve Rowe (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741176#comment-13741176
 ] 

Steve Rowe commented on SOLR-5159:
--

Hi Artem, I've run and compared with and without your patch on trunk with 
Maven3, looks good so far: ${svn.revision} and ${now.timestamp} and one other 
timestazmp that weren't being interpolated now are.  I want to verify Maven2 
locally, and I also want to compare all manifest entries with the Ant-produced 
ones - the solr entries were changed recently, and I want to keep them in sync. 
 Once these issues are settled, I'll commit.

 Manifest includes non-parsed maven variables
 

 Key: SOLR-5159
 URL: https://issues.apache.org/jira/browse/SOLR-5159
 Project: Solr
  Issue Type: Bug
  Components: Build
Affects Versions: 4.4, 4.5, 5.0
 Environment: Apache Maven 3.0.5
Reporter: Artem Karpenko
Assignee: Steve Rowe
Priority: Minor
  Labels: maven-bundle-plugin, maven3,
 Attachments: SOLR-5159.patch


 When building Lucene/Solr with Apache Maven 3, all MANIFEST.MF files included 
 into JAR artifacts contain non-parsed POM variables: namely, there are 
 entries like
 Specification-Version: 5.0.0.$\{now.version}
 In the end, Solr displays these values on admin page in Versions section.
 This is caused by unresolved bug in maven-bundle-plugin (FELIX-3392). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-trunk-Windows (32bit/jdk1.7.0_25) - Build # 3148 - Failure!

2013-08-15 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Windows/3148/
Java: 32bit/jdk1.7.0_25 -server -XX:+UseG1GC

No tests ran.

Build Log:
[...truncated 1326 lines...]
FATAL: hudson.remoting.RequestAbortedException: 
hudson.remoting.Channel$OrderlyShutdown
hudson.remoting.RequestAbortedException: 
hudson.remoting.RequestAbortedException: hudson.remoting.Channel$OrderlyShutdown
at 
hudson.remoting.RequestAbortedException.wrapForRethrow(RequestAbortedException.java:41)
at 
hudson.remoting.RequestAbortedException.wrapForRethrow(RequestAbortedException.java:34)
at hudson.remoting.Request.call(Request.java:174)
at hudson.remoting.Channel.call(Channel.java:714)
at 
hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:167)
at com.sun.proxy.$Proxy84.join(Unknown Source)
at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:925)
at hudson.Launcher$ProcStarter.join(Launcher.java:360)
at hudson.tasks.Ant.perform(Ant.java:217)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:19)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:804)
at hudson.model.Build$BuildExecution.build(Build.java:199)
at hudson.model.Build$BuildExecution.doRun(Build.java:160)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:586)
at hudson.model.Run.execute(Run.java:1597)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
at hudson.model.ResourceController.execute(ResourceController.java:88)
at hudson.model.Executor.run(Executor.java:247)
Caused by: hudson.remoting.RequestAbortedException: 
hudson.remoting.Channel$OrderlyShutdown
at hudson.remoting.Request.abort(Request.java:299)
at hudson.remoting.Channel.terminate(Channel.java:774)
at hudson.remoting.Channel$CloseCommand.execute(Channel.java:919)
at hudson.remoting.Channel$2.handle(Channel.java:461)
at 
hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:60)
Caused by: hudson.remoting.Channel$OrderlyShutdown
... 3 more
Caused by: Command close created at
at hudson.remoting.Command.init(Command.java:56)
at hudson.remoting.Channel$CloseCommand.init(Channel.java:913)
at hudson.remoting.Channel$CloseCommand.init(Channel.java:911)
at hudson.remoting.Channel.close(Channel.java:994)
at hudson.remoting.Channel.close(Channel.java:977)
at hudson.remoting.Channel$CloseCommand.execute(Channel.java:918)
... 2 more



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (SOLR-5160) MoreLikeThis (MLT) - New Optional Params - mlt.operator mlt.sort

2013-08-15 Thread Kranti Parisa (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5160?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741184#comment-13741184
 ] 

Kranti Parisa commented on SOLR-5160:
-

Yes, AND will help to get MLT documents based on the criteria. the criteria 
could be I want to get similar documents whose values are same as the original 
document. for example: all products manufactured in 2013 + mobiles category 
(these are attributes of the original document for which we are trying to find 
exact matches). And if the use case is to get either manufactured in 2013 OR 
mobiles category, then we can ask for OR.

Hence, I think allowing the user to specify the operator (mlt.operator or 
something) will help them get MLT documents based on their criteria and the 
results can be sorted based on mlt.sort

 MoreLikeThis (MLT) - New Optional Params -  mlt.operator  mlt.sort
 ---

 Key: SOLR-5160
 URL: https://issues.apache.org/jira/browse/SOLR-5160
 Project: Solr
  Issue Type: Improvement
  Components: MoreLikeThis
Reporter: Kranti Parisa
Priority: Minor
  Labels: features
   Original Estimate: 24h
  Remaining Estimate: 24h

 It seems that when we pass multiple field names with mlt.fl parameter, it is 
 ORing them to find the MLT documents.
 Is there a way to specify AND operator? Means if mlt.fl=language,year, then 
 we should return back the MLT documents which has language AND year field 
 values as same as the main query result document.
 http://localhost:8180/solr/mltCore/mlt?q=id:1wt=jsonmlt=truemlt.fl=language,yearfl=*,scoremlt.mindf=0mlt.mintf=0mlt.match.include=false
 The above query should return those documents whose field values (language, 
 year) are exactly matching with the document id:1.
 Is this possible thru any config or param? If not, I think it's worth having 
 as a feature because we don't know the values of those fields to apply as FQ. 
  was looking at the code and found that it is hard coded to Occur.SHOULD in 
 MoreLikeThisQuery.
 I customized the code to pass a new parameter mlt.operator=AND/OR  based on 
 that it computes the MLT documents. Default operator is set to OR.
 And I also want to have mlt.sort option, So I will be trying for that as well.
 Do you guys think, we should make this part of the MLT feature? 
 Please share your ideas. I can submit this change.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741185#comment-13741185
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1514377 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1514377 ]

SOLR-4952: SolrCmdDistributorTest needs to force MergePolicy for consistent 
optimize() behavior

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man
 Fix For: 4.5, 5.0


 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741188#comment-13741188
 ] 

ASF subversion and git services commented on LUCENE-5166:
-

Commit 1514379 from [~rcmuir] in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1514379 ]

LUCENE-5166: also fix and test this case where tf  1 within the passage for a 
term

 PostingsHighlighter fails with IndexOutOfBoundsException
 

 Key: LUCENE-5166
 URL: https://issues.apache.org/jira/browse/LUCENE-5166
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/highlighter
Affects Versions: 4.4
Reporter: Manuel Amoabeng
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166-revisited.patch


 Given a document with a match at a startIndex  PostingsHighlighter.maxlength 
 and an endIndex  PostingsHighlighter.maxLength, DefaultPassageFormatter will 
 throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
 invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-5166) PostingsHighlighter fails with IndexOutOfBoundsException

2013-08-15 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-5166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir resolved LUCENE-5166.
-

Resolution: Fixed

Thank you again!

 PostingsHighlighter fails with IndexOutOfBoundsException
 

 Key: LUCENE-5166
 URL: https://issues.apache.org/jira/browse/LUCENE-5166
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/highlighter
Affects Versions: 4.4
Reporter: Manuel Amoabeng
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5166-2.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, LUCENE-5166.patch, 
 LUCENE-5166.patch, LUCENE-5166-revisited.patch


 Given a document with a match at a startIndex  PostingsHighlighter.maxlength 
 and an endIndex  PostingsHighlighter.maxLength, DefaultPassageFormatter will 
 throw an IndexOutOfBoundsException when DefaultPassageFormatter.append() is 
 invoked. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5161) Distributed MoreLikeThis has problems when q.op=AND

2013-08-15 Thread Shawn Heisey (JIRA)
Shawn Heisey created SOLR-5161:
--

 Summary: Distributed MoreLikeThis has problems when q.op=AND
 Key: SOLR-5161
 URL: https://issues.apache.org/jira/browse/SOLR-5161
 Project: Solr
  Issue Type: Bug
  Components: MoreLikeThis
Affects Versions: 4.2.1
Reporter: Shawn Heisey
 Fix For: 4.5, 5.0


Distributed MoreLikeThis seems to result in regular distributed queries using 
the termvectors found in the example document.

If q.op=AND in the initial request, then the distributed queries with the 
termvectors will ALSO have q.op=AND, which for my tests, has resulted in no 
matches.  If I send the MLT request to a handler using edismax, or include 
q.op=OR, then it works as expected.  It's very slow, which I'll bring up in 
another issue.

Should it remove that parameter or assign it to OR?  Should it ultimately pay 
attention to the mlt.operator mentioned in SOLR-5160, which is not implemented 
at this time?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to ArithmeticException: / by zero

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741262#comment-13741262
 ] 

ASF subversion and git services commented on SOLR-5122:
---

Commit 1514402 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1514402 ]

SOLR-5122: Fixed bug in spellcheck.collateMaxCollectDocs.  Eliminates risk of 
divide by zero, and makes estimated hit counts meaningful in non-optimized 
indexes.

 spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
 to ArithmeticException: / by zero
 

 Key: SOLR-5122
 URL: https://issues.apache.org/jira/browse/SOLR-5122
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.4
Reporter: Hoss Man
Assignee: James Dyer
 Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch


 As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
 and this (aparently) led to a failure in testEstimatedHitCounts.
 As far as i can tell: the test assumes that specific values would be returned 
 as the _estimated_ hits for a colleation, and it appears that the change in 
 MergePolicy however resulted in different segments with different term stats, 
 causing the estimation code to produce different values then what is expected.
 I made a quick attempt to improve the test to:
  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
 set such that the estimate' should actually be exact (ie: 
 collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
 docs in the index
  * randomize the values used for collateMaxCollectDocs and confirm that the 
 estimates are never more then the num docs in the index
 This lead to an odd ArithmeticException: / by zero error in the test, which 
 seems to suggest that there is a genuine bug in the code for estimating the 
 hits that only gets tickled in certain 
 mergepolicy/segment/collateMaxCollectDocs combinations.
 *Update:* This appears to be a general problem with collecting docs out of 
 order and the estimation of hits -- i believe even if there is no divide by 
 zero error, the estimates are largely meaningless since the docs are 
 collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to ArithmeticException: / by zero

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741270#comment-13741270
 ] 

ASF subversion and git services commented on SOLR-5122:
---

Commit 1514408 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1514408 ]

SOLR-5122: Fixed bug in spellcheck.collateMaxCollectDocs.  Eliminates risk of 
divide by zero, and makes estimated hit counts meaningful in non-optimized 
indexes. (merge r1514402)

 spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
 to ArithmeticException: / by zero
 

 Key: SOLR-5122
 URL: https://issues.apache.org/jira/browse/SOLR-5122
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.4
Reporter: Hoss Man
Assignee: James Dyer
 Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch


 As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
 and this (aparently) led to a failure in testEstimatedHitCounts.
 As far as i can tell: the test assumes that specific values would be returned 
 as the _estimated_ hits for a colleation, and it appears that the change in 
 MergePolicy however resulted in different segments with different term stats, 
 causing the estimation code to produce different values then what is expected.
 I made a quick attempt to improve the test to:
  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
 set such that the estimate' should actually be exact (ie: 
 collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
 docs in the index
  * randomize the values used for collateMaxCollectDocs and confirm that the 
 estimates are never more then the num docs in the index
 This lead to an odd ArithmeticException: / by zero error in the test, which 
 seems to suggest that there is a genuine bug in the code for estimating the 
 hits that only gets tickled in certain 
 mergepolicy/segment/collateMaxCollectDocs combinations.
 *Update:* This appears to be a general problem with collecting docs out of 
 order and the estimation of hits -- i believe even if there is no divide by 
 zero error, the estimates are largely meaningless since the docs are 
 collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741263#comment-13741263
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1514404 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1514404 ]

SOLR-4952: SolrCmdDistributorTest needs to force MergePolicy for consistent 
optimize() behavior (merge r1514377)

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man
 Fix For: 4.5, 5.0


 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to ArithmeticException: / by zero

2013-08-15 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man resolved SOLR-5122.


   Resolution: Fixed
Fix Version/s: 5.0
   4.5
 Assignee: Hoss Man  (was: James Dyer)

Committed revision 1514402.
Committed revision 1514408.


 spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
 to ArithmeticException: / by zero
 

 Key: SOLR-5122
 URL: https://issues.apache.org/jira/browse/SOLR-5122
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.4
Reporter: Hoss Man
Assignee: Hoss Man
 Fix For: 4.5, 5.0

 Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch


 As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
 and this (aparently) led to a failure in testEstimatedHitCounts.
 As far as i can tell: the test assumes that specific values would be returned 
 as the _estimated_ hits for a colleation, and it appears that the change in 
 MergePolicy however resulted in different segments with different term stats, 
 causing the estimation code to produce different values then what is expected.
 I made a quick attempt to improve the test to:
  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
 set such that the estimate' should actually be exact (ie: 
 collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
 docs in the index
  * randomize the values used for collateMaxCollectDocs and confirm that the 
 estimates are never more then the num docs in the index
 This lead to an odd ArithmeticException: / by zero error in the test, which 
 seems to suggest that there is a genuine bug in the code for estimating the 
 hits that only gets tickled in certain 
 mergepolicy/segment/collateMaxCollectDocs combinations.
 *Update:* This appears to be a general problem with collecting docs out of 
 order and the estimation of hits -- i believe even if there is no divide by 
 zero error, the estimates are largely meaningless since the docs are 
 collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741379#comment-13741379
 ] 

Michael McCandless commented on LUCENE-5177:


+1 to cutover to non-weak HashMap.

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-Tests-4.x-Java7 - Build # 1506 - Failure

2013-08-15 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-4.x-Java7/1506/

All tests passed

Build Log:
[...truncated 28316 lines...]
  [javadoc] Generating Javadoc
  [javadoc] Javadoc execution
  [javadoc] warning: [options] bootstrap class path not set in conjunction with 
-source 1.6
  [javadoc] Loading source files for package org.apache.lucene...
  [javadoc] Loading source files for package org.apache.lucene.analysis...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.tokenattributes...
  [javadoc] Loading source files for package org.apache.lucene.codecs...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.compressing...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene3x...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene40...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene41...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene42...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.perfield...
  [javadoc] Loading source files for package org.apache.lucene.document...
  [javadoc] Loading source files for package org.apache.lucene.index...
  [javadoc] Loading source files for package org.apache.lucene.search...
  [javadoc] Loading source files for package 
org.apache.lucene.search.payloads...
  [javadoc] Loading source files for package 
org.apache.lucene.search.similarities...
  [javadoc] Loading source files for package org.apache.lucene.search.spans...
  [javadoc] Loading source files for package org.apache.lucene.store...
  [javadoc] Loading source files for package org.apache.lucene.util...
  [javadoc] Loading source files for package org.apache.lucene.util.automaton...
  [javadoc] Loading source files for package org.apache.lucene.util.fst...
  [javadoc] Loading source files for package org.apache.lucene.util.mutable...
  [javadoc] Loading source files for package org.apache.lucene.util.packed...
  [javadoc] Constructing Javadoc information...
  [javadoc] Standard Doclet version 1.7.0_25
  [javadoc] Building tree for all the packages and classes...
  [javadoc] Generating 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/package-summary.html...
  [javadoc] Copying file 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-2.png
 to directory 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
  [javadoc] Copying file 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-1.png
 to directory 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
  [javadoc] Building index for all the packages and classes...
  [javadoc] Building index for all classes...
  [javadoc] Generating 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/help-doc.html...
  [javadoc] 1 warning

[...truncated 27 lines...]
  [javadoc] Generating Javadoc
  [javadoc] Javadoc execution
  [javadoc] warning: [options] bootstrap class path not set in conjunction with 
-source 1.6
  [javadoc] Loading source files for package org.apache.lucene.analysis.ar...
  [javadoc] Loading source files for package org.apache.lucene.analysis.bg...
  [javadoc] Loading source files for package org.apache.lucene.analysis.br...
  [javadoc] Loading source files for package org.apache.lucene.analysis.ca...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.charfilter...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cjk...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cn...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.commongrams...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound.hyphenation...
  [javadoc] Loading source files for package org.apache.lucene.analysis.core...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cz...
  [javadoc] Loading source files for package org.apache.lucene.analysis.da...
  [javadoc] Loading source files for package org.apache.lucene.analysis.de...
  [javadoc] Loading source files for package org.apache.lucene.analysis.el...
  [javadoc] Loading source files for package org.apache.lucene.analysis.en...
  [javadoc] Loading source files for package org.apache.lucene.analysis.es...
  [javadoc] Loading source files for package org.apache.lucene.analysis.eu...
  [javadoc] Loading source files for package org.apache.lucene.analysis.fa...
  [javadoc] Loading source files for package 

[jira] [Commented] (SOLR-3280) to many / sometimes stale CLOSE_WAIT connections from SnapPuller during / after replication

2013-08-15 Thread David Fu (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741426#comment-13741426
 ] 

David Fu commented on SOLR-3280:


I am facing the same problem. Any update on when this will be resolved?

 to many / sometimes stale CLOSE_WAIT connections from SnapPuller during / 
 after replication
 ---

 Key: SOLR-3280
 URL: https://issues.apache.org/jira/browse/SOLR-3280
 Project: Solr
  Issue Type: Bug
Affects Versions: 3.5, 3.6, 4.0-ALPHA
Reporter: Bernd Fehling
Assignee: Robert Muir
Priority: Minor
 Attachments: SOLR-3280.patch


 There are sometimes to many and also stale CLOSE_WAIT connections 
 during/after replication left over on SLAVE server.
 Normally GC should clean up this but this is not always the case.
 Also if a CLOSE_WAIT is hanging then the new replication won't load.
 Dirty work around so far is to fake a TCP connection as root to that 
 connection and close it. 
 After that the new replication will load, the old index and searcher released 
 and the system will
 return to normal operation.
 Background:
 The SnapPuller is using Apache httpclient 3.x and uses the 
 MultiThreadedHttpConnectionManager.
 The manager holds a connection in CLOSE_WAIT after its use for further 
 requests.
 This is done by calling releaseConnection. But if a connection is stuck it is 
 not available any more and a new
 connection from the pool is used.
 Solution:
 After calling releaseConnection clean up with closeIdleConnections(0).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-Tests-4.x-Java7 - Build # 1506 - Failure

2013-08-15 Thread Steve Rowe
The email notification regex didn't catch the cause of the build failure (I'll 
fix the regex):

---
-ecj-javadoc-lint-src:
 [ecj-lint] Compiling 617 source files
 [ecj-lint] invalid Class-Path header in manifest of jar file: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/solr/core/lib/org.restlet-2.1.1.jar
 [ecj-lint] invalid Class-Path header in manifest of jar file: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/solr/core/lib/org.restlet.ext.servlet-2.1.1.jar
 [ecj-lint] --
 [ecj-lint] 1. ERROR in 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/solr/core/src/java/org/apache/solr/search/EarlyTerminatingCollectorException.java
 (at line 39)
 [ecj-lint] * the index when collecting the {@see #getNumberCollected()} 
documents 
 [ecj-lint]   ^^^
 [ecj-lint] Javadoc: Unexpected tag
 [ecj-lint] --
 [ecj-lint] 1 problem (1 error)
---


On Aug 15, 2013, at 4:07 PM, Apache Jenkins Server jenk...@builds.apache.org 
wrote:

 Build: https://builds.apache.org/job/Lucene-Solr-Tests-4.x-Java7/1506/
 
 All tests passed
 
 Build Log:
 [...truncated 28316 lines...]
  [javadoc] Generating Javadoc
  [javadoc] Javadoc execution
  [javadoc] warning: [options] bootstrap class path not set in conjunction 
 with -source 1.6
  [javadoc] Loading source files for package org.apache.lucene...
  [javadoc] Loading source files for package org.apache.lucene.analysis...
  [javadoc] Loading source files for package 
 org.apache.lucene.analysis.tokenattributes...
  [javadoc] Loading source files for package org.apache.lucene.codecs...
  [javadoc] Loading source files for package 
 org.apache.lucene.codecs.compressing...
  [javadoc] Loading source files for package 
 org.apache.lucene.codecs.lucene3x...
  [javadoc] Loading source files for package 
 org.apache.lucene.codecs.lucene40...
  [javadoc] Loading source files for package 
 org.apache.lucene.codecs.lucene41...
  [javadoc] Loading source files for package 
 org.apache.lucene.codecs.lucene42...
  [javadoc] Loading source files for package 
 org.apache.lucene.codecs.perfield...
  [javadoc] Loading source files for package org.apache.lucene.document...
  [javadoc] Loading source files for package org.apache.lucene.index...
  [javadoc] Loading source files for package org.apache.lucene.search...
  [javadoc] Loading source files for package 
 org.apache.lucene.search.payloads...
  [javadoc] Loading source files for package 
 org.apache.lucene.search.similarities...
  [javadoc] Loading source files for package org.apache.lucene.search.spans...
  [javadoc] Loading source files for package org.apache.lucene.store...
  [javadoc] Loading source files for package org.apache.lucene.util...
  [javadoc] Loading source files for package 
 org.apache.lucene.util.automaton...
  [javadoc] Loading source files for package org.apache.lucene.util.fst...
  [javadoc] Loading source files for package org.apache.lucene.util.mutable...
  [javadoc] Loading source files for package org.apache.lucene.util.packed...
  [javadoc] Constructing Javadoc information...
  [javadoc] Standard Doclet version 1.7.0_25
  [javadoc] Building tree for all the packages and classes...
  [javadoc] Generating 
 /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/package-summary.html...
  [javadoc] Copying file 
 /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-2.png
  to directory 
 /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
  [javadoc] Copying file 
 /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-1.png
  to directory 
 /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
  [javadoc] Building index for all the packages and classes...
  [javadoc] Building index for all classes...
  [javadoc] Generating 
 /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/help-doc.html...
  [javadoc] 1 warning
 
 [...truncated 27 lines...]
  [javadoc] Generating Javadoc
  [javadoc] Javadoc execution
  [javadoc] warning: [options] bootstrap class path not set in conjunction 
 with -source 1.6
  [javadoc] Loading source files for package org.apache.lucene.analysis.ar...
  [javadoc] Loading source files for package org.apache.lucene.analysis.bg...
  [javadoc] Loading source files for package org.apache.lucene.analysis.br...
  [javadoc] Loading source files for package org.apache.lucene.analysis.ca...
  [javadoc] Loading source files for package 
 org.apache.lucene.analysis.charfilter...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cjk...
  [javadoc] Loading source 

[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741469#comment-13741469
 ] 

Hoss Man commented on LUCENE-5177:
--

I understand the value of the improved testing -- but is there any tangible 
benefit to users to converting the WeakHashMap - HashMap?



 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-Tests-4.x-Java7 - Build # 1506 - Failure

2013-08-15 Thread Chris Hostetter

:  [ecj-lint]   * the index when collecting the {@see #getNumberCollected()} 
documents 

Bah ... Sorry, fix in progress.


: On Aug 15, 2013, at 4:07 PM, Apache Jenkins Server 
jenk...@builds.apache.org wrote:
: 
:  Build: https://builds.apache.org/job/Lucene-Solr-Tests-4.x-Java7/1506/
:  
:  All tests passed
:  
:  Build Log:
:  [...truncated 28316 lines...]
:   [javadoc] Generating Javadoc
:   [javadoc] Javadoc execution
:   [javadoc] warning: [options] bootstrap class path not set in conjunction 
with -source 1.6
:   [javadoc] Loading source files for package org.apache.lucene...
:   [javadoc] Loading source files for package org.apache.lucene.analysis...
:   [javadoc] Loading source files for package 
org.apache.lucene.analysis.tokenattributes...
:   [javadoc] Loading source files for package org.apache.lucene.codecs...
:   [javadoc] Loading source files for package 
org.apache.lucene.codecs.compressing...
:   [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene3x...
:   [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene40...
:   [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene41...
:   [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene42...
:   [javadoc] Loading source files for package 
org.apache.lucene.codecs.perfield...
:   [javadoc] Loading source files for package org.apache.lucene.document...
:   [javadoc] Loading source files for package org.apache.lucene.index...
:   [javadoc] Loading source files for package org.apache.lucene.search...
:   [javadoc] Loading source files for package 
org.apache.lucene.search.payloads...
:   [javadoc] Loading source files for package 
org.apache.lucene.search.similarities...
:   [javadoc] Loading source files for package 
org.apache.lucene.search.spans...
:   [javadoc] Loading source files for package org.apache.lucene.store...
:   [javadoc] Loading source files for package org.apache.lucene.util...
:   [javadoc] Loading source files for package 
org.apache.lucene.util.automaton...
:   [javadoc] Loading source files for package org.apache.lucene.util.fst...
:   [javadoc] Loading source files for package 
org.apache.lucene.util.mutable...
:   [javadoc] Loading source files for package org.apache.lucene.util.packed...
:   [javadoc] Constructing Javadoc information...
:   [javadoc] Standard Doclet version 1.7.0_25
:   [javadoc] Building tree for all the packages and classes...
:   [javadoc] Generating 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/package-summary.html...
:   [javadoc] Copying file 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-2.png
 to directory 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
:   [javadoc] Copying file 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-1.png
 to directory 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/org/apache/lucene/search/doc-files...
:   [javadoc] Building index for all the packages and classes...
:   [javadoc] Building index for all classes...
:   [javadoc] Generating 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-Java7/lucene/build/docs/core/help-doc.html...
:   [javadoc] 1 warning
:  
:  [...truncated 27 lines...]
:   [javadoc] Generating Javadoc
:   [javadoc] Javadoc execution
:   [javadoc] warning: [options] bootstrap class path not set in conjunction 
with -source 1.6
:   [javadoc] Loading source files for package org.apache.lucene.analysis.ar...
:   [javadoc] Loading source files for package org.apache.lucene.analysis.bg...
:   [javadoc] Loading source files for package org.apache.lucene.analysis.br...
:   [javadoc] Loading source files for package org.apache.lucene.analysis.ca...
:   [javadoc] Loading source files for package 
org.apache.lucene.analysis.charfilter...
:   [javadoc] Loading source files for package 
org.apache.lucene.analysis.cjk...
:   [javadoc] Loading source files for package org.apache.lucene.analysis.cn...
:   [javadoc] Loading source files for package 
org.apache.lucene.analysis.commongrams...
:   [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound...
:   [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound.hyphenation...
:   [javadoc] Loading source files for package 
org.apache.lucene.analysis.core...
:   [javadoc] Loading source files for package org.apache.lucene.analysis.cz...
:   [javadoc] Loading source files for package org.apache.lucene.analysis.da...
:   [javadoc] Loading source files for package org.apache.lucene.analysis.de...
:   [javadoc] Loading source files for package org.apache.lucene.analysis.el...
:   [javadoc] Loading 

[jira] [Created] (SOLR-5162) Implicit core properties are no longer available

2013-08-15 Thread Alan Woodward (JIRA)
Alan Woodward created SOLR-5162:
---

 Summary: Implicit core properties are no longer available
 Key: SOLR-5162
 URL: https://issues.apache.org/jira/browse/SOLR-5162
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.5
Reporter: Alan Woodward
Assignee: Alan Woodward
Priority: Minor
 Fix For: 4.5


As noted by ~elyograg on IRC, implicit property substitution doesn't work any 
more in 4.5.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to ArithmeticException: / by zero

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741477#comment-13741477
 ] 

ASF subversion and git services commented on SOLR-5122:
---

Commit 1514494 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1514494 ]

SOLR-5122: fix javadocs

 spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
 to ArithmeticException: / by zero
 

 Key: SOLR-5122
 URL: https://issues.apache.org/jira/browse/SOLR-5122
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.4
Reporter: Hoss Man
Assignee: Hoss Man
 Fix For: 4.5, 5.0

 Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch


 As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
 and this (aparently) led to a failure in testEstimatedHitCounts.
 As far as i can tell: the test assumes that specific values would be returned 
 as the _estimated_ hits for a colleation, and it appears that the change in 
 MergePolicy however resulted in different segments with different term stats, 
 causing the estimation code to produce different values then what is expected.
 I made a quick attempt to improve the test to:
  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
 set such that the estimate' should actually be exact (ie: 
 collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
 docs in the index
  * randomize the values used for collateMaxCollectDocs and confirm that the 
 estimates are never more then the num docs in the index
 This lead to an odd ArithmeticException: / by zero error in the test, which 
 seems to suggest that there is a genuine bug in the code for estimating the 
 hits that only gets tickled in certain 
 mergepolicy/segment/collateMaxCollectDocs combinations.
 *Update:* This appears to be a general problem with collecting docs out of 
 order and the estimation of hits -- i believe even if there is no divide by 
 zero error, the estimates are largely meaningless since the docs are 
 collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5162) Implicit core properties are no longer available

2013-08-15 Thread Alan Woodward (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5162?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741482#comment-13741482
 ] 

Alan Woodward commented on SOLR-5162:
-

I managed to chop off the 'solr.' prefix of all these substitutions in 
SOLR-4914.  The tricksy part here is that these prefixes are not present in the 
core.properties file, so we need to add them in/strip them out at the 
appropriate times.

The only call where the 'solr.' prefix needs to be present is in the Properties 
object returned by CoreDescriptor.getCoreProperties(), and this method is only 
ever called when creating ResourceLoaders, so we should be able to safely 
isolate the changes here.

 Implicit core properties are no longer available
 

 Key: SOLR-5162
 URL: https://issues.apache.org/jira/browse/SOLR-5162
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.5
Reporter: Alan Woodward
Assignee: Alan Woodward
Priority: Minor
 Fix For: 4.5


 As noted by ~elyograg on IRC, implicit property substitution doesn't work any 
 more in 4.5.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to ArithmeticException: / by zero

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741477#comment-13741477
 ] 

ASF subversion and git services commented on SOLR-5122:
---

Commit 1514494 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1514494 ]

SOLR-5122: fix javadocs

 spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
 to ArithmeticException: / by zero
 

 Key: SOLR-5122
 URL: https://issues.apache.org/jira/browse/SOLR-5122
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.4
Reporter: Hoss Man
Assignee: Hoss Man
 Fix For: 4.5, 5.0

 Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch


 As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
 and this (aparently) led to a failure in testEstimatedHitCounts.
 As far as i can tell: the test assumes that specific values would be returned 
 as the _estimated_ hits for a colleation, and it appears that the change in 
 MergePolicy however resulted in different segments with different term stats, 
 causing the estimation code to produce different values then what is expected.
 I made a quick attempt to improve the test to:
  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
 set such that the estimate' should actually be exact (ie: 
 collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
 docs in the index
  * randomize the values used for collateMaxCollectDocs and confirm that the 
 estimates are never more then the num docs in the index
 This lead to an odd ArithmeticException: / by zero error in the test, which 
 seems to suggest that there is a genuine bug in the code for estimating the 
 hits that only gets tickled in certain 
 mergepolicy/segment/collateMaxCollectDocs combinations.
 *Update:* This appears to be a general problem with collecting docs out of 
 order and the estimation of hits -- i believe even if there is no divide by 
 zero error, the estimates are largely meaningless since the docs are 
 collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741491#comment-13741491
 ] 

Robert Muir commented on LUCENE-5177:
-

there need not be tangible benefits to users for us to make a change.

we can make a change because its refactoring our codebase and making it cleaner 
or many other reasons.

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741500#comment-13741500
 ] 

Hoss Man commented on LUCENE-5177:
--

bq. its refactoring our codebase and making it cleaner or many other reasons.

How does this do that?

I'm not trying to be confrontational, i'm genuinely not understanding what is 
improved by switching away from a WeakHashMap and i just want to make sure i'm 
not missunderstanding something about the big picture.

(If you proposed to get rid of the Map completely and have the Caches hang 
directly off the readers (something i remember discussing a LNG time ago 
that people seemed to think was a good idea but no one seemd to have bandwidth 
for) then i could totally understand arguments that doing so would be making 
the codebase cleaner -- but i'm not understanding what's clearner/better about 
using a global static HashMap instead of a WeakHashMap.)



 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-Tests-trunk-Java7 - Build # 4237 - Failure

2013-08-15 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-trunk-Java7/4237/

All tests passed

Build Log:
[...truncated 32721 lines...]
BUILD FAILED
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-trunk-Java7/build.xml:389:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-trunk-Java7/build.xml:60:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-trunk-Java7/solr/build.xml:583:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-trunk-Java7/lucene/common-build.xml:1657:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-trunk-Java7/lucene/common-build.xml:1691:
 Compile failed; see the compiler error output for details.

Total time: 76 minutes 10 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (SOLR-5122) spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead to ArithmeticException: / by zero

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741506#comment-13741506
 ] 

ASF subversion and git services commented on SOLR-5122:
---

Commit 1514504 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1514504 ]

SOLR-5122: fix javadocs (merge r1514494)

 spellcheck.collateMaxCollectDocs estimates seem to be meaninless -- can lead 
 to ArithmeticException: / by zero
 

 Key: SOLR-5122
 URL: https://issues.apache.org/jira/browse/SOLR-5122
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.4
Reporter: Hoss Man
Assignee: Hoss Man
 Fix For: 4.5, 5.0

 Attachments: SOLR-5122.patch, SOLR-5122.patch, SOLR-5122.patch


 As part of SOLR-4952 SpellCheckCollatorTest started using RandomMergePolicy, 
 and this (aparently) led to a failure in testEstimatedHitCounts.
 As far as i can tell: the test assumes that specific values would be returned 
 as the _estimated_ hits for a colleation, and it appears that the change in 
 MergePolicy however resulted in different segments with different term stats, 
 causing the estimation code to produce different values then what is expected.
 I made a quick attempt to improve the test to:
  * expect explicit exact values only when spellcheck.collateMaxCollectDocs is 
 set such that the estimate' should actually be exact (ie: 
 collateMaxCollectDocs  == 0 or collateMaxCollectDocs greater then the num 
 docs in the index
  * randomize the values used for collateMaxCollectDocs and confirm that the 
 estimates are never more then the num docs in the index
 This lead to an odd ArithmeticException: / by zero error in the test, which 
 seems to suggest that there is a genuine bug in the code for estimating the 
 hits that only gets tickled in certain 
 mergepolicy/segment/collateMaxCollectDocs combinations.
 *Update:* This appears to be a general problem with collecting docs out of 
 order and the estimation of hits -- i believe even if there is no divide by 
 zero error, the estimates are largely meaningless since the docs are 
 collected out of order.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5162) Implicit core properties are no longer available

2013-08-15 Thread Alan Woodward (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Woodward updated SOLR-5162:


Attachment: SOLR-5162.patch

Quick patch, which should solve the problem.  I need to think of a decent way 
of testing it, though - is there a reasonably non-invasive solrconfig parameter 
I can use here?

 Implicit core properties are no longer available
 

 Key: SOLR-5162
 URL: https://issues.apache.org/jira/browse/SOLR-5162
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.5
Reporter: Alan Woodward
Assignee: Alan Woodward
Priority: Minor
 Fix For: 4.5

 Attachments: SOLR-5162.patch


 As noted by ~elyograg on IRC, implicit property substitution doesn't work any 
 more in 4.5.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741514#comment-13741514
 ] 

Robert Muir commented on LUCENE-5177:
-

I don't think we should hang fieldcaches directly off readers.

Fieldcache is a broken design: users who want to sort/facet on a field should 
index their content correctly with docvalues instead.

Its fine for it to be wrapped underneath an UninvertingFilterReader, that also 
takes a MapString,DocValuesType up front though, but by no means should we 
have this broken shit tightly coupled with stuff in lucene/core

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-4.x-Linux (32bit/jdk1.6.0_45) - Build # 6944 - Still Failing!

2013-08-15 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6944/
Java: 32bit/jdk1.6.0_45 -client -XX:+UseParallelGC

3 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.client.solrj.impl.CloudSolrServerTest

Error Message:
2 threads leaked from SUITE scope at 
org.apache.solr.client.solrj.impl.CloudSolrServerTest: 1) Thread[id=15, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-SendThread(localhost.localdomain:58694),
 state=TIMED_WAITING, group=TGRP-CloudSolrServerTest] at 
java.lang.Thread.sleep(Native Method) at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:86) 
at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:937)
 at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)
2) Thread[id=16, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-EventThread,
 state=WAITING, group=TGRP-CloudSolrServerTest] at 
sun.misc.Unsafe.park(Native Method) at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:156) at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987)
 at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) 
at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:491)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: 2 threads leaked from SUITE 
scope at org.apache.solr.client.solrj.impl.CloudSolrServerTest: 
   1) Thread[id=15, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-SendThread(localhost.localdomain:58694),
 state=TIMED_WAITING, group=TGRP-CloudSolrServerTest]
at java.lang.Thread.sleep(Native Method)
at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:86)
at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:937)
at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)
   2) Thread[id=16, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-EventThread,
 state=WAITING, group=TGRP-CloudSolrServerTest]
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987)
at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:491)
at __randomizedtesting.SeedInfo.seed([9B250AF16AC5B6E4]:0)


FAILED:  
junit.framework.TestSuite.org.apache.solr.client.solrj.impl.CloudSolrServerTest

Error Message:
There are still zombie threads that couldn't be terminated:1) Thread[id=15, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-SendThread(localhost.localdomain:58694),
 state=TIMED_WAITING, group=TGRP-CloudSolrServerTest] at 
java.lang.Thread.sleep(Native Method) at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:86) 
at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:937)
 at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)

Stack Trace:
com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie 
threads that couldn't be terminated:
   1) Thread[id=15, 
name=TEST-CloudSolrServerTest.testDistribSearch-seed#[9B250AF16AC5B6E4]-SendThread(localhost.localdomain:58694),
 state=TIMED_WAITING, group=TGRP-CloudSolrServerTest]
at java.lang.Thread.sleep(Native Method)
at 
org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:86)
at 
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:937)
at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)
at __randomizedtesting.SeedInfo.seed([9B250AF16AC5B6E4]:0)


REGRESSION:  
org.apache.solr.client.solrj.impl.CloudSolrServerTest.testDistribSearch

Error Message:
java.util.concurrent.TimeoutException: Could not connect to ZooKeeper 
127.0.0.1:58694 within 3 ms

Stack Trace:
java.lang.RuntimeException: java.util.concurrent.TimeoutException: Could not 
connect to ZooKeeper 127.0.0.1:58694 within 3 ms
at 
__randomizedtesting.SeedInfo.seed([9B250AF16AC5B6E4:1AC384E91D9AD6D8]:0)
at 
org.apache.solr.common.cloud.SolrZkClient.init(SolrZkClient.java:130)
at 
org.apache.solr.common.cloud.SolrZkClient.init(SolrZkClient.java:93)
at 
org.apache.solr.common.cloud.SolrZkClient.init(SolrZkClient.java:84)
at 
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:89)
at 
org.apache.solr.cloud.AbstractZkTestCase.buildZooKeeper(AbstractZkTestCase.java:83)
at 

[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741547#comment-13741547
 ] 

Hoss Man commented on LUCENE-5177:
--

bq. by no means should we have this broken shit tightly coupled with stuff in 
lucene/core

ok fine, fieldcaches as a concept is fundementally broken, and the suggestion 
of hanging the caches of index readers is stupid in this day and age of 
docvalues -- forget it, then.  It was simply an aside comment.

Can you (or anyone else) please help me understand why keeping these broken 
caches in global static HashMaps is cleaner/better then keeping them in global 
static WeakHashMaps?



 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5163) edismax gobbles exceptions for unrelated issues

2013-08-15 Thread Steven Bower (JIRA)
Steven Bower created SOLR-5163:
--

 Summary: edismax gobbles exceptions for unrelated issues
 Key: SOLR-5163
 URL: https://issues.apache.org/jira/browse/SOLR-5163
 Project: Solr
  Issue Type: Bug
Reporter: Steven Bower


query:

q=foo AND bar
qf=field1
qf=field2
defType=edismax

Where field1 exists and field2 doesn't..

will treat the AND as a term vs and operator

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4956) make maxBufferedAddsPerServer configurable

2013-08-15 Thread Otis Gospodnetic (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741556#comment-13741556
 ] 

Otis Gospodnetic commented on SOLR-4956:


+1 for what [~gpendleb] wrote.  Let me be the master of my own destiny.

 make maxBufferedAddsPerServer configurable
 --

 Key: SOLR-4956
 URL: https://issues.apache.org/jira/browse/SOLR-4956
 Project: Solr
  Issue Type: Improvement
Affects Versions: 4.3, 5.0
Reporter: Erick Erickson

 Anecdotal user's list evidence indicates that in high-throughput situations, 
 the default of 10 docs/batch for inter-shard batching can generate 
 significant CPU load. See the thread titled Sharding and Replication on 
 June 19th, but the gist is below.
 I haven't poked around, but it's a little surprising on the surface that Asif 
 is seeing this kind of difference. So I'm wondering if this change indicates 
 some other underlying issue. Regardless, this seems like it would be good to 
 investigate.
 Here's the gist of Asif's experience from the thread:
 Its a completely practical problem - we are exploring Solr to build a real
 time analytics/data solution for a system handling about 1000 qps. We have
 various metrics that are stored as different collections on the cloud,
 which means very high amount of writes. The cloud also needs to support
 about 300-400 qps.
 We initially tested with a single Solr node on a 16 core / 24 GB box  for a
 single metric. We saw that writes were not a issue at all - Solr was
 handling it extremely well. We were also able to achieve about 200 qps from
 a single node.
 When we set up the cloud ( a ensemble on 6 boxes), we saw very high CPU
 usage on the replicas. Up to 10 cores were getting used for writes on the
 replicas. Hence my concern with respect to batch updates for the replicas.
 BTW, I altered the maxBufferedAddsPerServer to 1000 - and now CPU usage is
 very similar to single node installation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741555#comment-13741555
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. ok fine, fieldcaches as a concept is fundementally broken, and the 
suggestion of hanging the caches of index readers is stupid in this day and age 
of docvalues 

Meh... I happen to disagree with both of those assertions.

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741560#comment-13741560
 ] 

Robert Muir commented on LUCENE-5177:
-

Thats fine, users who want to have slow reopen times and inefficient use of 
memory can use an FilterReader that uninverts and exposes stuff via the 
AtomicReader docValues APIs.

Such a FilterReader is useful as a transition mechanism anyway: users could 
pass it to addIndexes to 'upgrade' to docvalues.

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741580#comment-13741580
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. users who want to have slow reopen times and inefficient use of memory can 
use an FilterReader that uninverts and exposes stuff via the AtomicReader 
docValues APIs

As long as it's not slower / more inefficient than the field cache stuff we 
have today, that's fine.  Just a different API to access it. 

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5164) Can not create a collection via collections API

2013-08-15 Thread Erick Erickson (JIRA)
Erick Erickson created SOLR-5164:


 Summary: Can not create a collection via collections API
 Key: SOLR-5164
 URL: https://issues.apache.org/jira/browse/SOLR-5164
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.5, 5.0
Reporter: Erick Erickson
Assignee: Erick Erickson
Priority: Blocker


When you try to create a collection in SolrCloud, the instanceDir that gets 
created has an extra solr in it which messes up the pathing for all the lib 
directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741583#comment-13741583
 ] 

Yonik Seeley commented on LUCENE-5177:
--

Hoss - wrt WeakHashMap, I guess if it's not needed, lookups could be very 
slightly faster, debugging maybe slightly easier (if you're looking at weak 
references on the heap for example), and the code easier to understand (since 
we are not in fact relying on a weak reference to clean anything up anymore).

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5175) Add parameter to lower-bound TF normalization for BM25 (for long documents)

2013-08-15 Thread Tom Burton-West (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741582#comment-13741582
 ] 

Tom Burton-West commented on LUCENE-5175:
-

Hi Robert,

I tried running luceneutils with the default wikimedium10m collection and 
tasks.   I ran it first on the DefaultSimilarity, which shouldn't be affected 
by the patch to BM25Similarity and it showed about -2.3% difference.  I'm 
guessing there is some inaccuracy in the tests.   When I changed 
DEFAULT_SIMILARITY to BM25Similarity, the worst change was a difference of 
-8.8%.  

Is there a separate mailing list for questions about luceneutils or should I 
write to the java-dev list? or directly to Mike or you?

Tom

 Add parameter to lower-bound TF normalization for BM25 (for long documents)
 ---

 Key: LUCENE-5175
 URL: https://issues.apache.org/jira/browse/LUCENE-5175
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/search
Reporter: Tom Burton-West
Priority: Minor
 Attachments: LUCENE-5175.patch


 In the article When Documents Are Very Long, BM25 Fails! a fix for the 
 problem is documented.  There was a TODO note in BM25Similarity to add this 
 fix. I will attach a patch that implements the fix shortly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741586#comment-13741586
 ] 

Robert Muir commented on LUCENE-5177:
-

Yes, i dont think we should make it slower. but with such a filterreader, it 
could be implemented cleaner/differently more easily, because people can access 
it thru DV apis rather than thru FieldCache.DEFAULT.xxx


 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5175) Add parameter to lower-bound TF normalization for BM25 (for long documents)

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741595#comment-13741595
 ] 

Robert Muir commented on LUCENE-5175:
-

Tom: there is some variation on the test.

in my localconstants.py i have the following which I found reduces variation 
significantly:
{noformat}
JAVA_COMMAND = 'java -Xms4g -Xmx4g -server'
SEARCH_NUM_THREADS = 1
{noformat}

As far as questions, just send them to the dev@lucene list i think? a lot of 
committers use this thing so its probably your best bet.

 Add parameter to lower-bound TF normalization for BM25 (for long documents)
 ---

 Key: LUCENE-5175
 URL: https://issues.apache.org/jira/browse/LUCENE-5175
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/search
Reporter: Tom Burton-West
Priority: Minor
 Attachments: LUCENE-5175.patch


 In the article When Documents Are Very Long, BM25 Fails! a fix for the 
 problem is documented.  There was a TODO note in BM25Similarity to add this 
 fix. I will attach a patch that implements the fix shortly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5164) Can not create a collection via collections API

2013-08-15 Thread Shawn Heisey (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5164?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741601#comment-13741601
 ] 

Shawn Heisey commented on SOLR-5164:


In a cloud (or manually sharded) environment, IMHO it's better to put jars in 
${solr.solr.home}/lib because they'll be automatically loaded at startup and 
available to all cores.

Because we can't assume symlink support, it's really difficult to use this 
method with the example.

 Can not create a collection via collections API
 ---

 Key: SOLR-5164
 URL: https://issues.apache.org/jira/browse/SOLR-5164
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.5, 5.0
Reporter: Erick Erickson
Assignee: Erick Erickson
Priority: Blocker

 When you try to create a collection in SolrCloud, the instanceDir that gets 
 created has an extra solr in it which messes up the pathing for all the 
 lib directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-5164) Can not create a collection via collections API

2013-08-15 Thread Shawn Heisey (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5164?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741601#comment-13741601
 ] 

Shawn Heisey edited comment on SOLR-5164 at 8/15/13 10:39 PM:
--

In a cloud (or manually sharded) environment, IMHO it's better to put jars in 
$solrhome/lib because they'll be automatically loaded at startup and available 
to all cores.

Because we can't assume symlink support, it's really difficult to use this 
method with the example.

  was (Author: elyograg):
In a cloud (or manually sharded) environment, IMHO it's better to put jars 
in ${solr.solr.home}/lib because they'll be automatically loaded at startup and 
available to all cores.

Because we can't assume symlink support, it's really difficult to use this 
method with the example.
  
 Can not create a collection via collections API
 ---

 Key: SOLR-5164
 URL: https://issues.apache.org/jira/browse/SOLR-5164
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.5, 5.0
Reporter: Erick Erickson
Assignee: Erick Erickson
Priority: Blocker

 When you try to create a collection in SolrCloud, the instanceDir that gets 
 created has an extra solr in it which messes up the pathing for all the 
 lib directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5152) Lucene FST is not immutable

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5152?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741604#comment-13741604
 ] 

ASF subversion and git services commented on LUCENE-5152:
-

Commit 1514520 from [~mikemccand] in branch 'dev/trunk'
[ https://svn.apache.org/r1514520 ]

LUCENE-5152: make assert less costly

 Lucene FST is not immutable
 ---

 Key: LUCENE-5152
 URL: https://issues.apache.org/jira/browse/LUCENE-5152
 Project: Lucene - Core
  Issue Type: Bug
  Components: core/FSTs
Affects Versions: 4.4
Reporter: Simon Willnauer
Priority: Blocker
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5152.patch, LUCENE-5152.patch, LUCENE-5152.patch, 
 LUCENE-5152.patch


 a spinnoff from LUCENE-5120 where the analyzing suggester modified a returned 
 output from and FST (BytesRef) which caused sideffects in later execution. 
 I added an assertion into the FST that checks if a cached root arc is 
 modified and in-fact this happens for instance in our MemoryPostingsFormat 
 and I bet we find more places. We need to think about how to make this less 
 trappy since it can cause bugs that are super hard to find.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5152) Lucene FST is not immutable

2013-08-15 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5152?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741609#comment-13741609
 ] 

ASF subversion and git services commented on LUCENE-5152:
-

Commit 1514521 from [~mikemccand] in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1514521 ]

LUCENE-5152: make assert less costly

 Lucene FST is not immutable
 ---

 Key: LUCENE-5152
 URL: https://issues.apache.org/jira/browse/LUCENE-5152
 Project: Lucene - Core
  Issue Type: Bug
  Components: core/FSTs
Affects Versions: 4.4
Reporter: Simon Willnauer
Priority: Blocker
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5152.patch, LUCENE-5152.patch, LUCENE-5152.patch, 
 LUCENE-5152.patch


 a spinnoff from LUCENE-5120 where the analyzing suggester modified a returned 
 output from and FST (BytesRef) which caused sideffects in later execution. 
 I added an assertion into the FST that checks if a cached root arc is 
 modified and in-fact this happens for instance in our MemoryPostingsFormat 
 and I bet we find more places. We need to think about how to make this less 
 trappy since it can cause bugs that are super hard to find.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Shawn Heisey (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741611#comment-13741611
 ] 

Shawn Heisey commented on LUCENE-5177:
--

How is Solr affected by the idea of not using fieldcache?  From what I 
understand, docValues have the same data that would go into a stored field, not 
the indexed field ... so they might not behave exactly the same when it comes 
to sorting.  Although it doesn't make any sense to sort on a fully tokenized 
field, it can make sense to sort (or facet) on a field that uses the keyword 
tokenizer, the lowercase filter, and the trim filter.  I don't think that's 
possible with docValues.  


 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741613#comment-13741613
 ] 

Robert Muir commented on LUCENE-5177:
-

Thats all solr issues, its not related here.

DocValues fields take a byte[]. you can put whatever you want in there, it 
doesnt have to be the same as what goes in a stored field, you can run it thru 
an analysis chain yourself in some solr document processor or something like 
that if you really want to do that... you don't need indexwriter's help.

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741622#comment-13741622
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. How is Solr affected by the idea of not using fieldcache?

AFAIK, fieldcache-type functionality isn't going away, and any Lucene API 
changes (like FieldCache.DEFAULT, etc) will be hidden by Solr.
DocValues adds additional functionality, and does not take away from anything 
we already have.

There are currently one or two disadvantages to using docvalues in solr 
currently.  The major disadvantage is being forced to specify a default value 
in Solr, thus making them impossible to use from dynamic fields.   Ideally we'd 
be able to specify a missing value (i.e. the value used when there is no 
value for that document... or rather the default at the DocValues layer), and 
that would allow us to remove the currently mandated default value at the Solr 
layer.

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741631#comment-13741631
 ] 

Robert Muir commented on LUCENE-5177:
-

But really this should just be an option on the solr fieldtype, you know when 
the value is missing, solr just puts that missing value in the dv field for 
that document (nowhere else like stored fields or anything like that).

same with if you want the fieldtype to run stuff thru an analyzer or anything 
like that. I dont think this stuff really has to do with lucene, it can just be 
options in solrs update process/fieldtypes.

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-4.x-MacOSX (64bit/jdk1.7.0) - Build # 715 - Still Failing!

2013-08-15 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-MacOSX/715/
Java: 64bit/jdk1.7.0 -XX:-UseCompressedOops -XX:+UseG1GC

All tests passed

Build Log:
[...truncated 28349 lines...]
  [javadoc] Generating Javadoc
  [javadoc] Javadoc execution
  [javadoc] warning: [options] bootstrap class path not set in conjunction with 
-source 1.6
  [javadoc] Loading source files for package org.apache.lucene...
  [javadoc] Loading source files for package org.apache.lucene.analysis...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.tokenattributes...
  [javadoc] Loading source files for package org.apache.lucene.codecs...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.compressing...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene3x...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene40...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene41...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.lucene42...
  [javadoc] Loading source files for package 
org.apache.lucene.codecs.perfield...
  [javadoc] Loading source files for package org.apache.lucene.document...
  [javadoc] Loading source files for package org.apache.lucene.index...
  [javadoc] Loading source files for package org.apache.lucene.search...
  [javadoc] Loading source files for package 
org.apache.lucene.search.payloads...
  [javadoc] Loading source files for package 
org.apache.lucene.search.similarities...
  [javadoc] Loading source files for package org.apache.lucene.search.spans...
  [javadoc] Loading source files for package org.apache.lucene.store...
  [javadoc] Loading source files for package org.apache.lucene.util...
  [javadoc] Loading source files for package org.apache.lucene.util.automaton...
  [javadoc] Loading source files for package org.apache.lucene.util.fst...
  [javadoc] Loading source files for package org.apache.lucene.util.mutable...
  [javadoc] Loading source files for package org.apache.lucene.util.packed...
  [javadoc] Constructing Javadoc information...
  [javadoc] Standard Doclet version 1.7.0_25
  [javadoc] Building tree for all the packages and classes...
  [javadoc] Generating 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/build/docs/core/org/apache/lucene/search/package-summary.html...
  [javadoc] Copying file 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-1.png
 to directory 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/build/docs/core/org/apache/lucene/search/doc-files...
  [javadoc] Copying file 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/core/src/java/org/apache/lucene/search/doc-files/nrq-formula-2.png
 to directory 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/build/docs/core/org/apache/lucene/search/doc-files...
  [javadoc] Building index for all the packages and classes...
  [javadoc] Building index for all classes...
  [javadoc] Generating 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-4.x-MacOSX/lucene/build/docs/core/help-doc.html...
  [javadoc] 1 warning

[...truncated 27 lines...]
  [javadoc] Generating Javadoc
  [javadoc] Javadoc execution
  [javadoc] warning: [options] bootstrap class path not set in conjunction with 
-source 1.6
  [javadoc] Loading source files for package org.apache.lucene.analysis.ar...
  [javadoc] Loading source files for package org.apache.lucene.analysis.bg...
  [javadoc] Loading source files for package org.apache.lucene.analysis.br...
  [javadoc] Loading source files for package org.apache.lucene.analysis.ca...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.charfilter...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cjk...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cn...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.commongrams...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound...
  [javadoc] Loading source files for package 
org.apache.lucene.analysis.compound.hyphenation...
  [javadoc] Loading source files for package org.apache.lucene.analysis.core...
  [javadoc] Loading source files for package org.apache.lucene.analysis.cz...
  [javadoc] Loading source files for package org.apache.lucene.analysis.da...
  [javadoc] Loading source files for package org.apache.lucene.analysis.de...
  [javadoc] Loading source files for package org.apache.lucene.analysis.el...
  [javadoc] Loading source files for package org.apache.lucene.analysis.en...
  [javadoc] Loading source files for package org.apache.lucene.analysis.es...
  [javadoc] Loading source files for package org.apache.lucene.analysis.eu...
  [javadoc] Loading source files for package org.apache.lucene.analysis.fa...
  [javadoc] Loading source files for package 

[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741703#comment-13741703
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. But really this should just be an option on the solr fieldtype, you know 
when the value is missing, solr just puts that missing value in the dv field 
for that document.

But doing it at the solr level means that you can't define a field without 
using it.
Defining a non-docvalues field in solr and not using it currently incurs no 
overhead.  This isn't the case with docvalues, and I don't believe docvalues 
allows one to currently specify a default (it always defaults to 0 for missing 
values?)

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741707#comment-13741707
 ] 

Robert Muir commented on LUCENE-5177:
-

why would have fields defined you arent using? I dont understnad the use case 
here.

and to say there is no overhead with fieldcache is wrong: it makes huge arrays 
even if one document has the field. so I'm really missing something: in both 
cases its a column-stride field, just one is built at index-time.

I dont understand why docvalues needs to allow you to specify a default, when 
you can just specify your own for the document with missing value (if you 
specify 0, its no different than if it fills that for you).

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5164) Can not create a collection via collections API (cloud mode)

2013-08-15 Thread Erick Erickson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erick Erickson updated SOLR-5164:
-

Summary: Can not create a collection via collections API (cloud mode)  
(was: Can not create a collection via collections API)

 Can not create a collection via collections API (cloud mode)
 

 Key: SOLR-5164
 URL: https://issues.apache.org/jira/browse/SOLR-5164
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.5, 5.0
Reporter: Erick Erickson
Assignee: Erick Erickson
Priority: Blocker

 When you try to create a collection in SolrCloud, the instanceDir that gets 
 created has an extra solr in it which messes up the pathing for all the 
 lib directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741714#comment-13741714
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. why would have fields defined you arent using?

Why not? Other fields take no resources if you don't use them, but docvalues do.
Dynamic fields represent an almost infinite number of fields you aren't using 
until you do.
BTW, this is why the only uses of docvalues in the example schema are commented 
out.  Who want's to incur index overhead for fields you aren't even using?

If we want anything other than 0 for a value for missing, we must do it in 
DocValues.


 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741715#comment-13741715
 ] 

Robert Muir commented on LUCENE-5177:
-

Thats so wrong: if you want a value other than 0 (say 20) you just set the 
missing value yourself by setting it in the o.a.l.Document you add to 
indexwriter.

There is absolutely no reason for indexwriter to do things for you that you can 
do yourself.

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (SOLR-3936) QueryElevationComponent: Wrong order when result grouping is activated

2013-08-15 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3936?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man reassigned SOLR-3936:
--

Assignee: Hoss Man

 QueryElevationComponent: Wrong order when result grouping is activated
 --

 Key: SOLR-3936
 URL: https://issues.apache.org/jira/browse/SOLR-3936
 Project: Solr
  Issue Type: Bug
  Components: SearchComponents - other
Affects Versions: 4.0
Reporter: Michael Berger
Assignee: Hoss Man
 Attachments: SOLR-3936.patch, SOLR-3936.patch


 When I use elevation together with grouping I got not the expected result 
 order.
 I tried it with the standard solr example:
 http://localhost:8983/solr/elevate?enableElevation=truefl=score%2C[elevated]%2Cid%2CnameforceElevation=truegroup.field=manugroup=onindent=onq=ipodwt=json
  
 but the results ignored the elevation: 
 { 
   responseHeader:{ 
 status:0, 
 QTime:2, 
 params:{ 
   enableElevation:true, 
   fl:score,[elevated],id,name, 
   indent:on, 
   q:ipod, 
   forceElevation:true, 
   group.field:manu, 
   group:on, 
   wt:json}}, 
   grouped:{ 
 manu:{ 
   matches:2, 
   groups:[{ 
   groupValue:belkin, 
   doclist:{numFound:1,start:0,maxScore:0.7698604,docs:[ 
   { 
 id:F8V7067-APL-KIT, 
 name:Belkin Mobile Power Cord for iPod w/ Dock, 
 score:0.7698604, 
 [elevated]:false}] 
   }}, 
 { 
   groupValue:inc, 
   doclist:{numFound:1,start:0,maxScore:0.28869766,docs:[ 
   { 
 id:MA147LL/A, 
 name:Apple 60 GB iPod with Video Playback Black, 
 score:0.28869766, 
 [elevated]:true}] 
   }}]}}}
 the elevate.xml defines the following rules :
 query text=ipod
doc id=MA147LL/A /  !-- put the actual ipod at the top --
doc id=IW-02 exclude=true / !-- exclude this cable --
  /query
  
 /elevate

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3936) QueryElevationComponent: Wrong order when result grouping is activated

2013-08-15 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3936?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-3936:
---

Attachment: SOLR-3936.patch

[~mgarski]: Thanks for tracking this down and providing a patch, and 
esspecially thanks for including tests!

Everything you have looks great -- i've just updated your patch to include 
additional permutations (forceElevate, group.sort, etc...)

I'll try to commit this tomorrow (running it through full regression now to 
make sure it doesn't break any other tests unexpectedly)

 QueryElevationComponent: Wrong order when result grouping is activated
 --

 Key: SOLR-3936
 URL: https://issues.apache.org/jira/browse/SOLR-3936
 Project: Solr
  Issue Type: Bug
  Components: SearchComponents - other
Affects Versions: 4.0
Reporter: Michael Berger
 Attachments: SOLR-3936.patch, SOLR-3936.patch


 When I use elevation together with grouping I got not the expected result 
 order.
 I tried it with the standard solr example:
 http://localhost:8983/solr/elevate?enableElevation=truefl=score%2C[elevated]%2Cid%2CnameforceElevation=truegroup.field=manugroup=onindent=onq=ipodwt=json
  
 but the results ignored the elevation: 
 { 
   responseHeader:{ 
 status:0, 
 QTime:2, 
 params:{ 
   enableElevation:true, 
   fl:score,[elevated],id,name, 
   indent:on, 
   q:ipod, 
   forceElevation:true, 
   group.field:manu, 
   group:on, 
   wt:json}}, 
   grouped:{ 
 manu:{ 
   matches:2, 
   groups:[{ 
   groupValue:belkin, 
   doclist:{numFound:1,start:0,maxScore:0.7698604,docs:[ 
   { 
 id:F8V7067-APL-KIT, 
 name:Belkin Mobile Power Cord for iPod w/ Dock, 
 score:0.7698604, 
 [elevated]:false}] 
   }}, 
 { 
   groupValue:inc, 
   doclist:{numFound:1,start:0,maxScore:0.28869766,docs:[ 
   { 
 id:MA147LL/A, 
 name:Apple 60 GB iPod with Video Playback Black, 
 score:0.28869766, 
 [elevated]:true}] 
   }}]}}}
 the elevate.xml defines the following rules :
 query text=ipod
doc id=MA147LL/A /  !-- put the actual ipod at the top --
doc id=IW-02 exclude=true / !-- exclude this cable --
  /query
  
 /elevate

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741728#comment-13741728
 ] 

Robert Muir commented on LUCENE-5177:
-

By the way: just to explain how it works in FieldCache when you supply these 
sort missing/first/last stuff. 
The way that works is with a separate fieldcache field (a bitset) which marks 
documents that have a value.
So its really a separate fieldcache'd boolean[maxdoc] telling you if there is 
anything there or not for the original field.

You can easily do the exact same parallel with docvalues yourself (e.g. in a 
solr fieldtype) if you want to support those options (a numericdvfield only 1 
or 0, takes 1 bit per document, same situation).


 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3936) QueryElevationComponent: Wrong order when result grouping is activated

2013-08-15 Thread Michael Garski (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3936?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741729#comment-13741729
 ] 

Michael Garski commented on SOLR-3936:
--

Great! Thanks [~hossman]!

 QueryElevationComponent: Wrong order when result grouping is activated
 --

 Key: SOLR-3936
 URL: https://issues.apache.org/jira/browse/SOLR-3936
 Project: Solr
  Issue Type: Bug
  Components: SearchComponents - other
Affects Versions: 4.0
Reporter: Michael Berger
Assignee: Hoss Man
 Attachments: SOLR-3936.patch, SOLR-3936.patch


 When I use elevation together with grouping I got not the expected result 
 order.
 I tried it with the standard solr example:
 http://localhost:8983/solr/elevate?enableElevation=truefl=score%2C[elevated]%2Cid%2CnameforceElevation=truegroup.field=manugroup=onindent=onq=ipodwt=json
  
 but the results ignored the elevation: 
 { 
   responseHeader:{ 
 status:0, 
 QTime:2, 
 params:{ 
   enableElevation:true, 
   fl:score,[elevated],id,name, 
   indent:on, 
   q:ipod, 
   forceElevation:true, 
   group.field:manu, 
   group:on, 
   wt:json}}, 
   grouped:{ 
 manu:{ 
   matches:2, 
   groups:[{ 
   groupValue:belkin, 
   doclist:{numFound:1,start:0,maxScore:0.7698604,docs:[ 
   { 
 id:F8V7067-APL-KIT, 
 name:Belkin Mobile Power Cord for iPod w/ Dock, 
 score:0.7698604, 
 [elevated]:false}] 
   }}, 
 { 
   groupValue:inc, 
   doclist:{numFound:1,start:0,maxScore:0.28869766,docs:[ 
   { 
 id:MA147LL/A, 
 name:Apple 60 GB iPod with Video Playback Black, 
 score:0.28869766, 
 [elevated]:true}] 
   }}]}}}
 the elevate.xml defines the following rules :
 query text=ipod
doc id=MA147LL/A /  !-- put the actual ipod at the top --
doc id=IW-02 exclude=true / !-- exclude this cable --
  /query
  
 /elevate

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4734) FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight

2013-08-15 Thread Jack Krupansky (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741735#comment-13741735
 ] 

Jack Krupansky commented on LUCENE-4734:


Looking at some highlighter code, I see this constructor in 
org.apache.lucene.search.vectorhighlight.FieldPhraseList.java of branch_4x:

{code}
/**
 * a constructor.
 * 
 * @param fieldTermStack FieldTermStack object
 * @param fieldQuery FieldQuery object
 * @param phraseLimit maximum size of phraseList
 */
public FieldPhraseList( FieldTermStack fieldTermStack, FieldQuery fieldQuery, 
int phraseLimit ){
  final String field = fieldTermStack.getFieldName();

  QueryPhraseMap qpm = fieldQuery.getRootMap(field);
  if (qpm != null) {
LinkedListTermInfo phraseCandidate = new LinkedListTermInfo();
extractPhrases(fieldTermStack.termList, qpm, phraseCandidate, 0);
assert phraseCandidate.size() == 0;
  }
}
{code}

Clearly phraseLimit is no longer used. Is it being deprecated, or is this 
simply work in progress that will use it again eventually?

This parameter is passed over several layers of code, ultimately it is set up 
in Solr using the hl.phraseLimit parameter.

Seems like a dead parameter that should be cleaned up now or deprecated for 
future cleanup, but I can't say that I have been able to follow all of the work 
that has transpired in the highlighters.

The change occurred in Revision 1505732 (related to this Jira.) Before then, 
this parameter was used.

Comments? Or should this be a separate Jira issue?


 FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight
 

 Key: LUCENE-4734
 URL: https://issues.apache.org/jira/browse/LUCENE-4734
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/highlighter
Affects Versions: 4.0, 4.1, 5.0
Reporter: Ryan Lauck
Assignee: Adrien Grand
  Labels: fastvectorhighlighter, highlighter
 Fix For: 5.0, 4.5

 Attachments: LUCENE-4734-2.patch, lucene-4734.patch, LUCENE-4734.patch


 If a proximity phrase query overlaps with any other query term it will not be 
 highlighted.
 Example Text:  A B C D E F G
 Example Queries: 
 B E~10 D
 (D will be highlighted instead of B C D E)
 B E~10 C F~10
 (nothing will be highlighted)
 This can be traced to the FieldPhraseList constructor's inner while loop. 
 From the first example query, the first TermInfo popped off the stack will be 
 B. The second TermInfo will be D which will not be found in the submap 
 for B E~10 and will trigger a failed match.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741738#comment-13741738
 ] 

Yonik Seeley commented on LUCENE-5177:
--

bq. Thats so wrong: if you want a value other than 0 (say 20) you just set the 
missing value yourself by setting it in the o.a.l.Document you add to 
indexwriter.

And how exactly could we do that with dynamic fields?
Face it - if we want any other defaults than 0 (for numerics) it needs to 
somehow be configurable.

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741747#comment-13741747
 ] 

Robert Muir commented on LUCENE-5177:
-

It doesnt need to be configurable. its not configurable in fieldcache today!
It has a separate bitset cached on fieldcache indicating 1 or 0.

so you just do the same thing with docvalues, as I explained earlier.


 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5164) Can not create a collection via collections API (cloud mode)

2013-08-15 Thread Erick Erickson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erick Erickson updated SOLR-5164:
-

Attachment: SOLR-5164.patch

[~romseygeek] Well, this has to be the most time I've spent changing about 4 
lines.

If you can please take a glance at it. These seem fairly safe changes, they 
both just bring back what was in the code a while ago. Besides the solr/solr 
issue, when creating a properties file in Cloud mode, it appears that the 
parent directory hasn't been created first, so that step was failing.

If you don't apply it (and backport it to 4.4) I'll do it in the morning. But 
you have a 5 hour head start G...
 

 Can not create a collection via collections API (cloud mode)
 

 Key: SOLR-5164
 URL: https://issues.apache.org/jira/browse/SOLR-5164
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.5, 5.0
Reporter: Erick Erickson
Assignee: Erick Erickson
Priority: Blocker
 Attachments: SOLR-5164.patch


 When you try to create a collection in SolrCloud, the instanceDir that gets 
 created has an extra solr in it which messes up the pathing for all the 
 lib directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Yonik Seeley (JIRA)
Yonik Seeley created LUCENE-5178:


 Summary: doc values should allow configurable defaults
 Key: LUCENE-5178
 URL: https://issues.apache.org/jira/browse/LUCENE-5178
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Yonik Seeley


DocValues should somehow allow a configurable default per-field.
Possible implementations include setting it on the field in the document or 
registration of an IndexWriter callback.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5177) remove fieldcache weakmap or at least see what relies on GC for purging today

2013-08-15 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741770#comment-13741770
 ] 

Yonik Seeley commented on LUCENE-5177:
--

I've opened LUCENE-5178 for this.  It's not strictly related to the FieldCache.

 remove fieldcache weakmap or at least see what relies on GC for purging today
 -

 Key: LUCENE-5177
 URL: https://issues.apache.org/jira/browse/LUCENE-5177
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-5177.patch


 If we are registering close listeners why does this need to be weak?
 But really i dont care about that, here is what i said to Hoss on the solr 
 mailing list:
 {quote}
  (In any case: it looks like a WeakHashMap is still used in case the
  listeners never get called, correct?)
 
 I think it might be the other way around: i think it was weakmap
 before always, the close listeners were then added sometime in 3.x
 series, so we registered purge events as an optimization.
 But one way to look at it is: readers should really get closed, so why
 have the weak map and not just a regular hashmap.
 Even if we want to keep the weak map (seriously i dont care, and i
 dont want to be the guy fielding complaints on this), I'm going to
 open with an issue with a patch that removes it and fails tests in
 @afterclass if there is any entries. This way its totally clear
 if/when/where anything is relying on GC today here and we can at
 least look at that.
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Yonik Seeley (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-5178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonik Seeley updated LUCENE-5178:
-

Description: 
DocValues should somehow allow a configurable default per-field.
Possible implementations include setting it on the field in the document or 
registration of an IndexWriter callback.

If we don't make the default configurable, then another option is to have 
DocValues fields keep track of whether a value was indexed for that document or 
not.

  was:
DocValues should somehow allow a configurable default per-field.
Possible implementations include setting it on the field in the document or 
registration of an IndexWriter callback.


 doc values should allow configurable defaults
 -

 Key: LUCENE-5178
 URL: https://issues.apache.org/jira/browse/LUCENE-5178
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Yonik Seeley

 DocValues should somehow allow a configurable default per-field.
 Possible implementations include setting it on the field in the document or 
 registration of an IndexWriter callback.
 If we don't make the default configurable, then another option is to have 
 DocValues fields keep track of whether a value was indexed for that document 
 or not.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5164) Can not create a collection via collections API (cloud mode)

2013-08-15 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5164?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741783#comment-13741783
 ] 

Erick Erickson commented on SOLR-5164:
--

[~elyograg] It's not a matter of putting the jars in the right place at all. 
The core was just being created in the wrong place so the relative pathing from 
the stock solrconfig.xml couldn't work. The Carrotsearch classes that couldn't 
be found are part of the stock solrconfig distribution


 Can not create a collection via collections API (cloud mode)
 

 Key: SOLR-5164
 URL: https://issues.apache.org/jira/browse/SOLR-5164
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.5, 5.0
Reporter: Erick Erickson
Assignee: Erick Erickson
Priority: Blocker
 Attachments: SOLR-5164.patch


 When you try to create a collection in SolrCloud, the instanceDir that gets 
 created has an extra solr in it which messes up the pathing for all the 
 lib directives in solrconfig.xml as they're all relative.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741786#comment-13741786
 ] 

Robert Muir commented on LUCENE-5178:
-

I dont think lucene needs to do anything here. As i explained on LUCENE-5177 
(but you refused to listen), you just have a separate numericdocvalues field 
just like fieldcache does (it has a separate bitset).

So you write a 1 there, when the document has a value for the field. Otherwise 
it will contain a 0 (because it gets filled with that).

This is very easy to do and will use ~ 1 bit per document just like fieldcache.

Then its just a matter of making it easy to pass this NumericDV to 
FIeldComparator (currently: it always pulls a Bits directly from FC). this can 
be something like:
{code}
if (ndv instanceof Bits) { // codec already specializes here
  docsWithField = (Bits) ndv;
} else {
  docsWithField = new Bits() {
boolean get(int index) {
  return ndv.get(index) != 0;
}
}
{code}


 doc values should allow configurable defaults
 -

 Key: LUCENE-5178
 URL: https://issues.apache.org/jira/browse/LUCENE-5178
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Yonik Seeley

 DocValues should somehow allow a configurable default per-field.
 Possible implementations include setting it on the field in the document or 
 registration of an IndexWriter callback.
 If we don't make the default configurable, then another option is to have 
 DocValues fields keep track of whether a value was indexed for that document 
 or not.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5178) doc values should allow configurable defaults

2013-08-15 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741786#comment-13741786
 ] 

Robert Muir commented on LUCENE-5178:
-

I dont think lucene needs to do anything here. As i explained on LUCENE-5177 
(but you refused to listen), you just have a separate numericdocvalues field 
just like fieldcache does (it has a separate bitset).

So you write a 1 there, when the document has a value for the field. Otherwise 
it will contain a 0 (because it gets filled with that).

This is very easy to do and will use ~ 1 bit per document just like fieldcache.

Then its just a matter of making it easy to pass this NumericDV to 
FIeldComparator (currently: it always pulls a Bits directly from FC). this can 
be something like:
{code}
if (ndv instanceof Bits) { // codec already specializes here
  docsWithField = (Bits) ndv;
} else {
  docsWithField = new Bits() {
boolean get(int index) {
  return ndv.get(index) != 0;
}
}
{code}


 doc values should allow configurable defaults
 -

 Key: LUCENE-5178
 URL: https://issues.apache.org/jira/browse/LUCENE-5178
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Yonik Seeley

 DocValues should somehow allow a configurable default per-field.
 Possible implementations include setting it on the field in the document or 
 registration of an IndexWriter callback.
 If we don't make the default configurable, then another option is to have 
 DocValues fields keep track of whether a value was indexed for that document 
 or not.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



  1   2   >