http://www.google.com/search?q=datastore+garbage+collection
Typically, it is done on a scheduled basis. Justin On Apr 29, 2011, at 8:16 AM, Fabián Mandelbaum <[email protected]> wrote: > How do I do such a thing? If I have to do it every time an item (file) > is deleted, wouldn't it be too resource-costly? > > On Fri, Apr 29, 2011 at 9:10 AM, Justin Edelson > <[email protected]> wrote: >> Did you do a datastore garbage collection? >> >> Justin >> >> On Apr 29, 2011, at 8:03 AM, Fabián Mandelbaum <[email protected]> wrote: >> >>> Hello, >>> >>> while doing some tests yesterday I've noticed the following >>> (unexpected) behaviour with JackRabbit 2.2.5 (probably 1.6.x behaves >>> the same): >>> >>> I have a repository with lots of data (mostly files and folders), the >>> on-disk repository folder takes 10GB. >>> >>> I remove most of the data with standard JCR APIs, the on-disk >>> repository folder still takes 10GB. >>> >>> I use the backup feature of the jackrabbit-standalone-2.2.5.jar >>> (slightly modified by me to include the H2 jar files, because I use H2 >>> as storage backend) to copy the 10GB repo folder on a new folder, the >>> on-disk copied repository folder with the data left after removal on >>> the original repo takes (the expected) 1GB. >>> >>> Why isn't the datastore removing the files belonging to the removed >>> data (I use the standard repository.xml configuration file, just >>> replacing Derby by H2)? >>> >>> Hope to have been clear enough, waiting for an answer, thanks in advance. >>> >>> -- >>> Fabián Mandelbaum >>> IS Engineer >> > > > > -- > Fabián Mandelbaum > IS Engineer
