> From: Michal Plechawski
>
> I think that Brian's idea is more flexible and extendable. In my
> application, I need three or more kinds of analyzers: for 
> counting tfidf
> statistics, for indexing (compute more, e.g. summaries) and 
> for document
> classification (compute document-to-class assignment and 
> store outside the
> index) and for some minor things.
> My experience shows that in complex Lucene applications there is a
> substantial need for many different Analyzers or - better 
> solution - many
> faces of the same Analyzer in the same time. Something should be done
> here.

Currently it is easy to use different analyzers for different purposes, no?
I'm not sure how Brian's proposal (bi-modal analyzers: tokenize only &
tokenize+normalize) addresses your needs.

> Another story is - why did you put document deletion to 
> IndexReader? I guess
> the main reason was the implementation, but from the API 
> point of view it is
> horrible.

Yes, sorry.  I wonder if it would have been better to instead call
IndexWriter IndexAdder or something, to make clear that it can only add
documents.  Perhaps someday this can be fixed.

> Last thing - did you ever think about adding transactions to 
> Lucene? May be
> very simple exclusive-write transactions - e.g. reads are not 
> transacted nor
> isolated, and writes are done in such a way - the write is 
> exclusive (I
> guess it is in 1.2, I use 1.0), and one may commit/rollback 
> all changes made
> during last session. Would it be hard?

That is in fact what is done in 1.2.

Doug

--
To unsubscribe, e-mail:   <mailto:[EMAIL PROTECTED]>
For additional commands, e-mail: <mailto:[EMAIL PROTECTED]>

Reply via email to