bq: Does the new compressed stored field format in Solr 4.1 do anything to
reduce the number of disk seeks required to retrieve all document fields?

Probably, but I doubt by a whole lot. Although I confess I really don't
know the guts. Let's assume that all the stored content for a doc is
contiguous. The odds of having more than one doc in a block read from disk
goes up with compression which would reduce the number of seeks. But, the
odds of any of the top 20 docs in a corpus of, say, 20M docs being close
enough for this to happen is probably pretty small.

But read Uwe's excellent blog on MMapDirectory here:
http://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html

Best,
Erick


On Tue, Feb 19, 2013 at 11:24 PM, Shawn Heisey <s...@elyograg.org> wrote:

> On 2/19/2013 6:47 PM, Erick Erickson wrote:
>
>> It Depends (tm). Storing data in a Solr index pretty much just consumes
>> disk space, the *.fdt and *.fdx files aren't really germane to the amount
>> of memory needed for search. There will be some additional memory
>> requirements for the document cache though. And you'll also have resources
>> consumed if you use the &fl=*, there's more disk seeking going on to fetch
>> the fields.
>>
>
> Does the new compressed stored field format in Solr 4.1 do anything to
> reduce the number of disk seeks required to retrieve all document fields?
>
> Thanks,
> Shawn
>
>

Reply via email to