Hello,

I'm looking to back up a CouchDB server with multiple databases. Currently
1,400, but it fluctuates up and down throughout the day as new databases
are added and old ones deleted. ~10% of the databases are written to within
any 5 minute period of time.

Goals
 - Maintain a continual off-site snapshot of all databases, preferably no
older than a few seconds (or minutes)
 - Be efficient with bandwidth (i.e. not copy the whole database file for
every backup run)

My current solution watches the global _changes feed and fires up a
continuous replication to an off-site server whenever it sees a change. If
it doesn't see a change from a database for 10 minutes, it kills that
replication. This means I only have ~150 active replications running on
average at any given time.

I thought this was a pretty clever approach, but I can't stabilize it.
Replications hang frequently with crashes in the log file. I haven't yet
tracked down the source of the crashes. I'm running the official 1.6.1
docker image as of yesterday so I don't think it would be an erlang issue.

Rather than keep banging my head against these stability issues, I thought
I'd ask to see if anyone else has come up with a clever backup solution
that meets the above goals?

  Nick

Reply via email to