On 30 Jul 2013, at 20:03, Shane Johnson <[email protected]> wrote:

> One option might be to use a fix key set size and simply increment the value 
> for each key by X every time it is written. Sort of like an object with a 
> collection and every time a nested object is added to the collection, the 
> parent object is written to the cache.

In this example the aggregated objects should hold a foreign key to the 
aggregator, otherwise the object would grow indefinitely causing OOMs.
But good point nevertheless: if the size of every object varies from 
1k,2k,3k..Nk circularly, then the total disk capacity allocated for storing an 
entry is ((1+N)*N)/2.
So for storing 100MB of data you'd end up with a file with size 5GB. On top of 
that the memory consumption grows proportionally, as we keep in memory 
information about all the allocated segments on disk.  

Cheers,
-- 
Mircea Markus
Infinispan lead (www.infinispan.org)





_______________________________________________
infinispan-dev mailing list
[email protected]
https://lists.jboss.org/mailman/listinfo/infinispan-dev

Reply via email to