Re: Exceptions during batch indexing

2014-11-10 Thread Peter Keegan
--Original Message- From: Peter Keegan > Sent: Thursday, November 6, 2014 3:21 PM > To: java-user > Subject: Exceptions during batch indexing > > > How are folks handling Solr exceptions that occur during batch indexing? > Solr (4.6) stops parsing the docs stream when an er

Re: Exceptions during batch indexing

2014-11-08 Thread Jack Krupansky
Oops... you sent this to the wrong list - this is the Lucene user list, send it to the Solr user list. -- Jack Krupansky -Original Message- From: Peter Keegan Sent: Thursday, November 6, 2014 3:21 PM To: java-user Subject: Exceptions during batch indexing How are folks handling Solr

Exceptions during batch indexing

2014-11-06 Thread Peter Keegan
How are folks handling Solr exceptions that occur during batch indexing? Solr (4.6) stops parsing the docs stream when an error occurs (e.g. a doc with a missing mandatory field), and stops indexing. The bad document is not identified, so it would be hard for the client to recover by skipping over

RE: Batch Indexing - best practice?

2010-03-17 Thread Murdoch, Paul
ks again, Paul -Original Message- From: java-user-return-45439-paul.b.murdoch=saic@lucene.apache.org [mailto:java-user-return-45439-paul.b.murdoch=saic@lucene.apache.org ] On Behalf Of Erick Erickson Sent: Monday, March 15, 2010 12:45 PM To: java-user@lucene.apache.org Subject: Re: Batch

Re: Batch Indexing - best practice?

2010-03-15 Thread Erick Erickson
-Original Message- > From: java-user-return-45433-paul.b.murdoch=saic@lucene.apache.org > [mailto:java-user-return-45433-paul.b.murdoch=saic@lucene.apache.org > ] On Behalf Of Mark Miller > Sent: Monday, March 15, 2010 10:48 AM > To: java-user@lucene.apache.org > Subject:

Re: Batch Indexing - best practice?

2010-03-15 Thread Mark Miller
:48 AM To: java-user@lucene.apache.org Subject: Re: Batch Indexing - best practice? On 03/15/2010 10:41 AM, Murdoch, Paul wrote: Hi, I'm using Lucene 2.9.2. Currently, when creating my index, I'm calling indexWriter.addDocument(doc) for each Document I want to ind

RE: Batch Indexing - best practice?

2010-03-15 Thread Murdoch, Paul
:java-user-return-45433-paul.b.murdoch=saic@lucene.apache.org ] On Behalf Of Mark Miller Sent: Monday, March 15, 2010 10:48 AM To: java-user@lucene.apache.org Subject: Re: Batch Indexing - best practice? On 03/15/2010 10:41 AM, Murdoch, Paul wrote: > Hi, > > > > I'm using

Re: Batch Indexing - best practice?

2010-03-15 Thread Ian Lea
ble.  I have the merge factor set to > 1000.  Do you have any suggestions for batch indexing?  Is there > something like indexWriter.addDocuments(Document[] docs) in the API? > > > > Thanks. > > Paul > > > > -

Re: Batch Indexing - best practice?

2010-03-15 Thread Mark Miller
econds. I'd like to try and speed this upunless 90 seconds for 500 Documents is reasonable. I have the merge factor set to 1000. Do you have any suggestions for batch indexing? Is there something like indexWriter.addDocuments(Document[] docs) in the API? Thanks. Paul You shoul

Batch Indexing - best practice?

2010-03-15 Thread Murdoch, Paul
unless 90 seconds for 500 Documents is reasonable. I have the merge factor set to 1000. Do you have any suggestions for batch indexing? Is there something like indexWriter.addDocuments(Document[] docs) in the API? Thanks. Paul

Re: bad index by batch indexing

2008-08-06 Thread yanyanzeng
ot figure out where it >> went >> wrong, and how come this strange behavior: 520 documents but 0 terms. >> >> I would be very grateful if anyone could advise. THanks very much. >> >> yanyan >> >> >> -- >> View this message in conte

Re: bad index by batch indexing

2008-08-06 Thread Mark Miller
g, and how come this strange behavior: 520 documents but 0 terms. I would be very grateful if anyone could advise. THanks very much. yanyan -- View this message in context: http://www.nabble.com/bad-index-by-batch-indexing-tp18862037p18862037.html Sent from the Lucene - Java Users mailing list a

Re: bad index by batch indexing

2008-08-06 Thread Anshum
d be very grateful if anyone could advise. THanks very much. > > yanyan > > > -- > View this message in context: > http://www.nabble.com/bad-index-by-batch-indexing-tp18862037p18862037.html > Sent from the Lucene - Java Users mailing list archive at Nabble.com. > > &g

bad index by batch indexing

2008-08-06 Thread yanyanzeng
very grateful if anyone could advise. THanks very much. yanyan -- View this message in context: http://www.nabble.com/bad-index-by-batch-indexing-tp18862037p18862037.html Sent from the Lucene - Java Users mailing list archive at Nabble

Re: batch indexing

2007-05-02 Thread Erick Erickson
For some intermediate period of time when indexing, the document are buffered in RAM. There is a complex interplay between several of the parameters to an IndexWriter that govern how many documents are indexed in RAM before being flushed to the FSDirectory. Of course, if you specify a RAMdirector

Re: batch indexing

2007-05-02 Thread Chris
Sorry , ask a question. You say FSDirectory is RADdirectory as least until it flushes. I cannot understand your means . May you please teach me what it means? FSDirectory stored in filesystem , and RADdirectory stored in RAM. MergeFactor and MaxBufferedDocs settings are limited and controlled

Re: batch indexing

2007-04-29 Thread Erick Erickson
MAIL PROTECTED]> wrote: Thanks Erik , so FSDirectory seems better option than RAMDirectory ? Also I think O.S can cache files in which case FSDirectory may not be bad , your thoughts ? -Original Message- From: Erick Erickson [mailto:[EMAIL PROTECTED] Sent: Sunday, April 29, 2007 7:07

RE: batch indexing

2007-04-29 Thread Chandan Tamrakar
@lucene.apache.org Subject: Re: batch indexing As I understand it, FSDirectory *is* RAMdirectory, at least until it flushes. There have been several discussions of this, search the mail archive for things like MergeFactor, MaxBufferedDocs and the like. You'll find quite a bit of information about how

Re: batch indexing

2007-04-29 Thread Erick Erickson
As I understand it, FSDirectory *is* RAMdirectory, at least until it flushes. There have been several discussions of this, search the mail archive for things like MergeFactor, MaxBufferedDocs and the like. You'll find quite a bit of information about how these parameters interact. Particularly, s

batch indexing

2007-04-29 Thread Chandan Tamrakar
I am trying to index a huge documents on batches . Batch size is parameterized to the application say X docs , that means it will hold X no. of Docs in the RAM before I flush to file system using IndexWriter.addIndexes(Directory[]) method My question is : Do I need to set mergefactor ?

Re: batch indexing using RAMDirectory

2006-06-28 Thread James Pine
Hey Eric, I think you want: fsWriter.addIndexes(Directory[] {ramDir}); to be: fsWriter.addIndexes(new Directory[]{ramDir}); JAMES --- zheng <[EMAIL PROTECTED]> wrote: > I am a novice in lucene. I write some code to do > batch indexing using > RAMDirectory according to the c

batch indexing using RAMDirectory

2006-06-28 Thread zheng
I am a novice in lucene. I write some code to do batch indexing using RAMDirectory according to the code provided in lucene in action, which is something like FSDirectory fsDir = FSDirectory.getDirectory("/tmp/index", true); RAMDirectory ramDir = new RAMDirectory(); IndexWrite