I read the BerkeleyDB documentation at Oracle last year, IIRC Oracle did not ask for shutdown of the process that uses BerkeleyDB (in our case, that's qmaster) before db_archive can be run:
http://docs.oracle.com/cd/E17076_02/html/api_reference/C/db_archive.html Since you are not using the BDB RPC server, then you don't need to run db_checkpoint (bdb_checkpoint.sh calls db_checkpoint), and can simply run "db_archive -d" in a weekly crontab. Rayson On Fri, May 18, 2012 at 2:39 PM, Simon Matthews <[email protected]> wrote: > Thanks for pointing this out to me > > The documentation says that it should be used every minute if the > configuration uses a BDB server. I don't use a BDB server, but the storage > method I use is BDB (not flat files). If I should use this checkppoint > script, how often should I run it, and should I shut down the qmaster to run > it? > >> >> >> Rayson >> >> >> >> On Fri, May 18, 2012 at 1:17 PM, Simon Matthews >> <[email protected]> wrote: >> > After SGE was killed by the OOM killed, the file (a berkely db file) in >> > my >> > cluster was 1.4GB. I did a db_dump and db_load, on this file, resulting >> > in a >> > much smaller file. >> > >> > However, this then raised the question -- how is this file maintained? >> > Presumably, it holds the information on jobs in all states (queued, >> > running >> > and finished). How do the finished jobs get removed from this file? >> > Obviously, I don't want the file to grow without limit. >> > >> > We are now putting about 50k jobs into our small cluster every day (many >> > finish running in a fraction of a second). >> > >> > Simon >> > >> > _______________________________________________ >> > users mailing list >> > [email protected] >> > https://gridengine.org/mailman/listinfo/users >> > > > _______________________________________________ users mailing list [email protected] https://gridengine.org/mailman/listinfo/users
