mark k wrote:
Here is how I have mine setup, note the /' = at the top, backing
up 30 + systems this way with nor errors.
$Conf{BackupFilesExclude} = {
'/' = [
'/proc/*',
'/sys/*',
'/var/run/*',
'/dev/*',
If you use '/' as the key, then these excludes only apply to
Here is how I have mine setup, note the /' = at the top, backing up 30 +
systems this way with nor errors.
$Conf{BackupFilesExclude} = {
'/' = [
'/proc/*',
'/sys/*',
'/var/run/*',
'/dev/*',
-
This
Craig Barratt wrote:
Steven writes:
Running: /usr/bin/ssh -q -x -l netbackup freedom.rapidxdev.com
/usr/bin/sudo /usr/bin/rsync --server --sender --numeric-ids --perms
--owner --group --links --times --block-size=2048 --recursive -D
--bwlimit=200 --ignore-times . /
Yes, you can
Networking issue most likely. Sometimes hardware fails for no good reason.
If you have an extra ethernet card lying around I suggest you try that.
Alternatively, has anyone moved the computer recently? did the cable get
crunched? Is it near a microwave or do the cables run by a microwave? did
Hi can someone please help!..
Backuppc 3.1.0
Suse 10.3
2.6.22.16-0.2-default
rsync 2.6.9 on server and clients
I removed all the arguments from the command inorder to better troubleshoot
the issue. (i can just copy them back from the restore settings later)
but this is killing... any ideas?
Hi
Can someone help me please, I have just downloading Backuppc from the
syncatic for Ubuntu 7.10 and told it to load apache2. When I put in the url
for it, it gives me the following:
http://www.backuppc.com/
This domain is missing from the Web server configuration
The domain name is correctly
PROBLEM: Backup Hangs when using BackupPC / rsync over ssh to a Windows 2003
server.
WORKAROUND SUMMARY: Backup a Windows 2003 server using by using BackupPC's
Pre and Post commands to establish a forwarding ssh tunnel and a locally
bound Windows rsyncd service. I know what you are thinking - I
This is a great piece of knowledge, I encourage you to put this on the wiki.
also note that this can be done in reverse to have the remote machine create
the tunnel and issue a command over ssh 'BackupPC_servermsg or
BackupPC_dump' allowing remote clients to backup on their own schedule
without
Joe, what filesystem are you running on the downloading machine? is it
FAT32? When I read this I immediately thought FAT32 filesystem, FAT32 can't
handle 4GB+ files.
On Feb 19, 2008 10:54 AM, Joe Krahn [EMAIL PROTECTED] wrote:
Mirco Piccin wrote:
Hi and thanks for reply.
I'm trying to
Mirco Piccin wrote:
Hi and thanks for reply.
I'm trying to get a backup of a file.
It's size is about 12 GB.
I try to download it directly from Browse Backup, but the first time
download freeze at about 3,99 GB, and the second time at about 2,76
GB.
Maybe there's a timeout for the
Hi,
I'm using BackupPC version 3.0.0 and samba for backing up several
Windows and Linux clients..
and have spent lots of time (without working solution) trying to figure
out if there is some way to exclude following files:
A) directories which name starts with dot. ( ex. /dir/.obj/ )
B)
On 2008/02/19 17:04, dan wrote:
Networking issue most likely. Sometimes hardware fails for no good
reason. If you have an extra ethernet card lying around I suggest you
try that.
Alternatively, has anyone moved the computer recently? did the cable
get crunched? Is it near a microwave
Carl Wilhelm Soderstrom wrote:
On 11/28 09:39 , Tim Hall wrote:
Are there any known backuppc tweaks/settings that
are proven to increase transfer performance over
wan links? Specifically with using rsyncd or rsync
as the transfer method.
. . . .
If you're running =v3 the following
If you're running =v3 the following option will make all the incrementals
sync against the previous incremental, instead of the last full. This keeps
them from growing quite as quickly. (It's the behavior you expect from
rsync).
$Conf{IncrLevels} = [1, 2, 3, 4, 5, 6];
I was under
do you use a headset with your phone? like a plantronics? i have
plantronics cs50 and cs55 headsets that have to be about 6 away from an
ethernet cable or they distrupt the signal. typically this just causes the
network to fall back to 10Mbps but can cause all kinds of funky errors.
On Feb 19,
On 02/19 05:53 , Raman Gupta wrote:
Carl Wilhelm Soderstrom wrote:
If you're running =v3 the following option will make all the incrementals
sync against the previous incremental, instead of the last full. This
keeps
them from growing quite as quickly. (It's the behavior you expect from
i looked at my archive history hear and i have a number of hosts than do
incrementals take like 6 minutes and fulls like 46 minutes
On Feb 19, 2008 4:07 PM, Carl Wilhelm Soderstrom [EMAIL PROTECTED]
wrote:
On 02/19 05:53 , Raman Gupta wrote:
Carl Wilhelm Soderstrom wrote:
If you're
no, incrementals are more efficient on bandwidth. they do a less strenuous
test to determine if a file has changed.
at the expense of CPU power on both sides, you can compress the rsync
traffic either with rsync -z or if you are using ssh then with ssh's
compression. if you REALLY wanted to go
Hello,
It would be quite interesting to install BackupPC on a Linux WLAN Router.
For me it would be the AVM Fritz Box. It has a Ti-AR7 Processor with 32MB
Ram.
Now, someone successfully installed a Debian system on it, so I wonder, if
it would be useful to install BackupPC on it.
What would be
dan wrote:
no, incrementals are more efficient on bandwidth. they do a less
strenuous test to determine if a file has changed.
at the expense of CPU power on both sides, you can compress the rsync
traffic either with rsync -z
Have you tried rsync -z? Last I heard, BackupPC's rsync
Rich Rauenzahn wrote:
dan wrote:
no, incrementals are more efficient on bandwidth. they do a less
strenuous test to determine if a file has changed.
at the expense of CPU power on both sides, you can compress the rsync
traffic either with rsync -z
Have you tried rsync -z? Last I
rsync alone makes this box pretty much incapable of running backuppc. 32MB
of ram minus running system will give you at most 20MB usable under BEST
CASE scenario, which is about 150,000 files on a client MAXIMUM, which is
just not enough for many many clients. also, that CPU is just not
Steven writes:
Yes, I know that seems like the likely answer, but unfortunately it
isn't. This happens on all of the hosts, regardless of whether or not
there are per host configuration files.
Well, if you are confident the likely explanation is wrong, I'd
recommend adding some debug
23 matches
Mail list logo