Please check the Lucene's jGuru FAQ, your question is answered there.

Otis

--- Flavio Eduardo de Cordova <[EMAIL PROTECTED]> wrote:
> People...
> 
>       I've created a custom analyser that uses the StandardTokenizer class
> to get the tokens from the reader.
>       It seemed to work fine but I just noticed that some large documents
> are not having all their content properly indexed, but just [the
> starting]
> part of them.
>       After some debuging I've found out that StandardTokenizer reads up
> to 10001 tokens from the reader.
> 
>       Have anybody went through something like that before ? What should I
> do as a workaround ?
> 
> Thanks !
> 
> Flavio
> 
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: [EMAIL PROTECTED]
> For additional commands, e-mail: [EMAIL PROTECTED]
> 


__________________________________
Do you Yahoo!?
SBC Yahoo! DSL - Now only $29.95 per month!
http://sbc.yahoo.com

---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to