Bruce Momjian wrote:
> > What's the advantage of using pg_dumpall over tar/gzip for backup?
>
> pg_dumpall grabs a constent snapshot of the data. tar/gzip is just
> backing up the files, so you can get some data in some table that is
> committed, but miss data in another table that is part of the same
> transaction. Of course, if the database is not being used, this is not
> an issue.
>
> --
> Bruce Momjian | http://www.op.net/~candle
> [EMAIL PROTECTED] | (610) 853-3000
> + If your life is a hard drive, | 830 Blythe Avenue
> + Christ can be your backup. | Drexel Hill, Pennsylvania 19026
Thank you, Bruce.
I was using pg_dumpall nightly (with the help of cron) to backup and
suddenly my backup fails. The following was what I get from my backup file:
\connect template1
select datdba into table tmp_pg_shadow from pg_database where datname
= 'template1';
delete from pg_shadow where usesysid <> tmp_pg_shadow.datdba;
drop table tmp_pg_shadow;
copy pg_shadow from stdin;
officeuser 41 t t f t bb \N
\.
\connect template1 postgres
create database cpos;
\connect cpos postgres
I don't know why it stopped there. The size of the database is about
350MB. I allocated about 6 hours for the backup and it still fails. Is
there a limit to the size of the database for pg_dumpall or is it because it
takes much longer time than 6 hours for 350MB?
Regards,
Thomas.