Well, the biggest problem we have to answer for the clients is the following: 1. Backup method that doesn't take long and don't impact system 2. Restore needs to be done on a quick as possible way in order to minimize downtime.
The one client is running master - master replication with master server in usa, and slave in south africa. They need master backup to be done in the states. Sent via my BlackBerry from Vodacom - let your email find you! -----Original Message----- From: Reindl Harald <h.rei...@thelounge.net> Date: Thu, 01 Nov 2012 16:49:45 To: mysql@lists.mysql.com<mysql@lists.mysql.com> Subject: Re: Mysql backup for large databases good luck i would call snapshots on a running system much more dumb than "innodb_flush_log_at_trx_commit = 2" on systems with 100% stable power instead waste IOPS on shared storages Am 01.11.2012 16:45, schrieb Singer Wang: > Assuming you're not doing dumb stuff like innodb_flush_log_at_tx=0 or 2 and > etc, you should be fine. We have been > using the trio: flush tables with read lock, xfs_freeze, snapshot for months > now without any issues. And we test > the backups (we load the backup into a staging once a day, and dev once a > week) > > On Thu, Nov 1, 2012 at 11:41 AM, Reindl Harald <h.rei...@thelounge.net > <mailto:h.rei...@thelounge.net>> wrote: > > > Why do you need downtime? > > because mysqld has many buffers in memory and there > is no atomic "flush buffers in daemon and freeze backend FS" > > short ago there was a guy on this list which had to realize > this the hard way with a corrupt slave taken from a snapshot > > that's why i would ALWAYS do master/slave what means ONE time > down (rsync; stop master; rsync; start master) for a small > timewindow and after that you can stop the slave, take a > 100% consistent backup of it's whole datadir and start > the slave again which will do all transactions from the > binarylog happened in the meantime