Hi,
While trying to play with the CompoundWordTokenFilterBase I noticed that
the behavior is to include the original token together with the new
sub-tokens.
I assume this is expected (haven't found any relevant docs on this), but I
was wondering if it's a hard requirement or can I propose a small
Hi,
I was investigating some performance issues and during profiling I noticed
that there is a significant amount of time being spent decompressing fields
which are unrelated to the actual field I'm trying to load from the lucene
documents. In our benchmark doing mostly a simple full-test search,
gt; fully fits in the filesystem cache, this stored fields format might
> indeed have non-negligible overhead.
>
>
> On Wed, Apr 9, 2014 at 9:17 PM, Alex Parvulescu
> wrote:
> > Hi,
> >
> > I was investigating some performance issues and during profiling I
> noticed
&