Grant Ingersoll wrote:
If I'm understanding correctly...

What about a SinkTokenizer that is backed by a Reader/Field instead of the current one that stores it all in a List? This is more or less the use case for the Tee/Sink implementations, w/ the exception that we didn't plan for the Sink being too large, but that is easily overcome, IMO.

That is, you use a TeeTokenFilter that adds to your Sink, which serializes to some storage, and then your SinkTokenizer just unserializes. No need to change Fieldable at all or anything else

Or maybe just a Tokenizer that is backed by a Field would work and uses a TermEnum on the Field to serve up next() for the TokenStream.

Just thinking out loud...

Actually, the scenario is more complicated, because I need to implement this as a Solr FieldType ... besides, wouldn't this mean that I can't store the original value, because I'm setting the tokenStream on a Field (which automatically makes it un-stored)?

Anyway, thanks for the hint, I'll check if I can do it this way. Other points about the new Analyzer API - I still think it would offer more flexibility than the current API, for a minimal cost in compatibility, and likely no cost in performance.

--
Best regards,
Andrzej Bialecki     <><
 ___. ___ ___ ___ _ _   __________________________________
[__ || __|__/|__||\/|  Information Retrieval, Semantic Web
___|||__||  \|  ||  |  Embedded Unix, System Integration
http://www.sigram.com  Contact: info at sigram dot com


---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to