[jira] Commented: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Yonik Seeley (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-505?page=comments#action_12368447 ] Yonik Seeley commented on LUCENE-505: - >We're still using TermScorer, which generates the fakeNorms() regardless of >omitNorms on or off. Let me focus on that point for t

[jira] Commented: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Steven Tamm (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-505?page=comments#action_12368443 ] Steven Tamm commented on LUCENE-505: We're still using TermScorer, which generates the fakeNorms() regardless of omitNorms on or off. ConstantTermScorer is a step in the

[jira] Commented: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Yonik Seeley (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-505?page=comments#action_12368440 ] Yonik Seeley commented on LUCENE-505: - > One can now omit norms when indexing, and, if such a field is searched with a > normal query then fakeNorms will be used. > But a

[jira] Commented: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Yonik Seeley (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-505?page=comments#action_12368437 ] Yonik Seeley commented on LUCENE-505: - > I made the change less for MultiReader, but to prevent the instantiation of > the fakeNorms array (which is an extra 1MB of useles

[jira] Updated: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Steven Tamm (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-505?page=all ] Steven Tamm updated LUCENE-505: --- Attachment: LazyNorms.patch Here's a patch where if you turn LOAD_NORMS_INTO_MEM to false, it will instead load the norms from the disk all the time. When combi

[jira] Created: (LUCENE-508) SegmentTermEnum.next() doesn't maintain prevBuffer at end

2006-03-01 Thread Steven Tamm (JIRA)
SegmentTermEnum.next() doesn't maintain prevBuffer at end - Key: LUCENE-508 URL: http://issues.apache.org/jira/browse/LUCENE-508 Project: Lucene - Java Type: Bug Components: Index Versions: 1.9, 2.0

[jira] Commented: (LUCENE-507) CLONE -[PATCH] remove unused variables

2006-03-01 Thread Steven Tamm (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-507?page=comments#action_12368403 ] Steven Tamm commented on LUCENE-507: In Lucene 1.9, there are a lot of local variable and unused import warnings. > CLONE -[PATCH] remove unused variables > --

[jira] Created: (LUCENE-507) CLONE -[PATCH] remove unused variables

2006-03-01 Thread Steven Tamm (JIRA)
CLONE -[PATCH] remove unused variables -- Key: LUCENE-507 URL: http://issues.apache.org/jira/browse/LUCENE-507 Project: Lucene - Java Type: Improvement Components: Search Versions: unspecified Environment: Operating Sy

[jira] Updated: (LUCENE-506) Optimize Memory Use for Short-Lived Indexes (Do not load TermInfoIndex if you know the queries ahead of time)

2006-03-01 Thread Steven Tamm (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-506?page=all ] Steven Tamm updated LUCENE-506: --- Attachment: Prefetching.patch This also includes two additional test cases. The public exposure to the prefetching is controlled solely by IndexReader.open(Dire

[jira] Created: (LUCENE-506) Optimize Memory Use for Short-Lived Indexes (Do not load TermInfoIndex if you know the queries ahead of time)

2006-03-01 Thread Steven Tamm (JIRA)
Optimize Memory Use for Short-Lived Indexes (Do not load TermInfoIndex if you know the queries ahead of time) - Key: LUCENE-506 URL: http://issues.apache.org/jira/browse/L

[jira] Commented: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Steven Tamm (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-505?page=comments#action_12368389 ] Steven Tamm commented on LUCENE-505: > I also worry about performance with this change. Have you benchmarked this > while searching large indexes? yes. see below. > Fo

[jira] Updated: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Doug Cutting (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-505?page=all ] Doug Cutting updated LUCENE-505: Version: 2.0 (was: 1.9) I don't see how the memory requirements of MultiReader are twice that of SegmentReader. MultiReader does not call nor

[jira] Commented: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Steven Tamm (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-505?page=comments#action_12368381 ] Steven Tamm commented on LUCENE-505: I made the change less for MultiReader, but to prevent the instantiation of the fakeNorms array (which is an extra 1MB of useless memo

[jira] Commented: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Yonik Seeley (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-505?page=comments#action_12368378 ] Yonik Seeley commented on LUCENE-505: - > MultiReader.norms() is very inefficient: it has to construct a byte array > that's as long as all the documents in every > segment

[jira] Updated: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Steven Tamm (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-505?page=all ] Steven Tamm updated LUCENE-505: --- Attachment: NormFactors.patch Sorry, I didn't remove whitespace in the previous patch. This one's easier to read. svn diff --diff-cmd diff -x "-b -u" works bet

[jira] Updated: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Steven Tamm (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-505?page=all ] Steven Tamm updated LUCENE-505: --- Attachment: NormFactors.patch This patch doesn't include my previous change to TermScorer. It passes all of the lucene unit tests in addition to our set of tes

[jira] Created: (LUCENE-505) MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object

2006-03-01 Thread Steven Tamm (JIRA)
MultiReader.norm() takes up too much memory: norms byte[] should be made into an Object --- Key: LUCENE-505 URL: http://issues.apache.org/jira/browse/LUCENE-505 Project: Lucene - Java

Re: Changes.txt for contrib

2006-03-01 Thread Chris Hostetter
: distribution, we should start documenting their changes. I suggest that we : add a file contrib/CHANGES.txt. This way we don't pollute the top-level : changes file. Having one changes file per contrib project on the other : hand makes it more difficult to get an overview, so one in contrib seems

[jira] Commented: (LUCENE-500) Lucene 2.0 requirements - Remove all deprecated code

2006-03-01 Thread Doug Cutting (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-500?page=comments#action_12368358 ] Doug Cutting commented on LUCENE-500: - +1 This looks like a good start towards 2.0. > Lucene 2.0 requirements - Remove all deprecated code > -

Changes.txt for contrib

2006-03-01 Thread Daniel Naber
Hi, as most parts of the contrib area are now part of the official Lucene distribution, we should start documenting their changes. I suggest that we add a file contrib/CHANGES.txt. This way we don't pollute the top-level changes file. Having one changes file per contrib project on the other ha

Re: Index compatibility question

2006-03-01 Thread Daniel Naber
On Mittwoch 01 März 2006 16:21, DM Smith wrote: > I find that 1.9 reads my 1.4.3 built indexes just fine. But not the > other way around. That's exactly how it is supposed to be. Regards Daniel -- http://www.danielnaber.de -

Re: version issue ?

2006-03-01 Thread Andi Vajda
On Wed, 1 Mar 2006, Doug Cutting wrote: It does correspond to a precise SVN version, but what we prefer is a tag. The tag for 1.9-final is: http://svn.apache.org/repos/asf/lucene/java/tags/lucene_1_9_final/ Tags should never be revised. If you're paranoid, then you could also note the revi

Re: version issue ?

2006-03-01 Thread Doug Cutting
Andi Vajda wrote: It would seem to me that the source code snapshot that is made to release 'official' source and binary tarballs on the Lucene website should correspond to a precise svn version It does correspond to a precise SVN version, but what we prefer is a tag. The tag for 1.9-final i

[jira] Updated: (LUCENE-504) FuzzyQuery produces a "java.lang.NegativeArraySizeException" in PriorityQueue.initialize if I use Integer.MAX_VALUE as BooleanQuery.MaxClauseCount

2006-03-01 Thread Joerg Henss (JIRA)
[ http://issues.apache.org/jira/browse/LUCENE-504?page=all ] Joerg Henss updated LUCENE-504: --- Attachment: TestFuzzyQueryError.java Simple test showing the error > FuzzyQuery produces a "java.lang.NegativeArraySizeException" in > PriorityQueue.initialize

Index compatibility question

2006-03-01 Thread DM Smith
I have just upgraded to 1.9-final and am now testing my use of it. One question regarding compatibility. Does 1.4.3 search 1.9-final built indexes? I find that 1.9 reads my 1.4.3 built indexes just fine. But not the other way around. -

[jira] Created: (LUCENE-504) FuzzyQuery produces a "java.lang.NegativeArraySizeException" in PriorityQueue.initialize if I use Integer.MAX_VALUE as BooleanQuery.MaxClauseCount

2006-03-01 Thread Joerg Henss (JIRA)
FuzzyQuery produces a "java.lang.NegativeArraySizeException" in PriorityQueue.initialize if I use Integer.MAX_VALUE as BooleanQuery.MaxClauseCount --

Re: FuzzyQuery / PriorityQueue BUG

2006-03-01 Thread Bernhard Messer
Jörg, could you please add this to JIRA so that things don't get lost. If you have a patch and/or a testcase showing the problem, it would be great if you append it to JIRA also. thanks, Bernhard Jörg Henß wrote: Hi, FuzzyQuery produces a "java.lang.NegativeArraySizeException" in PriorityQ

FuzzyQuery / PriorityQueue BUG

2006-03-01 Thread Jörg Henß
Hi, FuzzyQuery produces a "java.lang.NegativeArraySizeException" in PriorityQueue.initialize if I use Integer.MAX_VALUE as BooleanQuery.MaxClauseCount. This is because it adds 1 to MaxClauseCount and tries to allocate an Array of this Size (I think it overflows to MIN_VALUE). Usually nobody needs s

[jira] Created: (LUCENE-503) Contrib: ThaiAnalyzer to enable Thai full-text search in Lucene

2006-03-01 Thread Samphan Raruenrom (JIRA)
Contrib: ThaiAnalyzer to enable Thai full-text search in Lucene --- Key: LUCENE-503 URL: http://issues.apache.org/jira/browse/LUCENE-503 Project: Lucene - Java Type: New Feature Components: Analysis