Yeah, still need to check that one.

On Nov 12, 2009, at 10:43 AM, Hans Pikkemaat wrote:

Hi,

Yes, the paginated query would indeed be the only way for me to go forward.
The problem however is that I get the exception I posted earlier.

tx

Hans

Andrus Adamchik wrote:
For paginated queries we contemplated a strategy of a list with constant size of fully resolved objects. I.e. when a page is swapped in, some other (LRU?) page is swapped out. We decided against it, as in a general case it is hard to consistently predict which page should be swapped out.

However it should be rather easy to write such a list for a specific case with a known access order (e.g. a standard iteration order). In fact I would vote to even include such implementation in Cayenne going forward.

More specifically, you can extend IncrementalFaultList [1], overriding 'resolveInterval' to swap out previously read pages, turning them back into ids. And the good part is that you can use your extension directly without any need to modify the rest of Cayenne.

Andrus


[1] 
http://cayenne.apache.org/doc/api/org/apache/cayenne/access/IncrementalFaultList.html


On Nov 12, 2009, at 10:07 AM, Hans Pikkemaat wrote:

Hi,

So this means that if I use a generic query that the query results are always stored completely in the object store (or the query cache if I configure it).

Objects are returned in a list so as long I have a reference to this list (because I'm
traversing it) these objects are not garbage collected.

If I use the query cache the full query results are cached. This means that I can only
tell it to remove the whole query.

Effectively this means I'm unable to run a big query and process the results as a stream. So I cannot process the first results and then somehow make them available for
garbage collection.

The only option I have would be the iterated query but this is only usefull for queries one 1 table without any relations because it is not possible to use prefetching nor is
it possible to manually construct relations between obects.

My conclusion here is that cayenne is simply not suitable for doing large batch wise
query processing because of the memory implications.

tx

HPI

Andrus Adamchik wrote:

As mentioned in the docs, individual objects and query lists are
cached independently. Of course query lists contain a subset of cached object store objects inside the lists. An object won't get gc'd if it
is also stored in the query list.

Now list cache expiration is controlled via query cache factory. By
default this is an LRU map, so as long as the map has enough space to hold lists (its capacity == # of lists, not # of objects), the objects
won't get gc'd.

You can explicitly remove entries from the cache via QueryCache remove and removeGroup methods. Or you can use a different QueryCacheFactory
that implements some custom expiration/cleanup mechanism.

Andrus

On Nov 11, 2009, at 3:43 PM, Hans Pikkemaat wrote:



Hi,

I use the latest version of cayenne, 3.0b and am experimenting with
the object caching features.

The documentation states that committed objects are purged from the
cache because it uses weak references.
(http://cayenne.apache.org/doc/individual-object-caching.html)

If I however run a query using SQLTemplate which caches the objects
into the dataContext local cache (objectstore),
the objects don't seem to be purged at all. If I simply run the
query dump the contents using an iterator on the resulting
List then the nr of registered objects in the objectstore stays the
same (dataContext.getObjectStore().registeredObjectsCount()).
Even if I manually run System.gc() I don't see any changes (I know
this can be normal as gc() doesn't guarantee anything)

What am I doing wrong? Under which circumstances will cayenne purge
the cache?

tx

Hans








Reply via email to