Re: [BackupPC-users] avoidable failure?
Hi, Cristian Tibirna wrote on 14.01.2007 at 13:16:29 [Re: [BackupPC-users] avoidable failure?]: On 14 January 2007 00:34, Holger Parplies wrote: [...] Might you simply need to increase your $Conf{ClientTimeout}? It would make sense that your backups take longer with busy client machines than with idle ones, after all. Interesting suggestion. I will try to investigate more in this direction. I don't know exactly was should be done as a matter of test though, as the errors aren't reproduceable, as I mentioned in the beginning. yes, there are always some things you can't really test :-(. You could look at the logs of your failing backups though and check whether the time they ran seems to correspond with your current (resp. former) setting of $Conf{ClientTimeout}. If that is set to 7200 s (2 h) and you have failed backups running 30 min, 43 min and 22 min and good ones running 15 min, 35 min and 65 min, then that's obviously the wrong track. If your good backups are comparatively short though and the others fail after roughly 2 hours, I'd simply try maybe doubling the value and seeing if the failures go away. As I understand it, $Conf{ClientTimeout} is not a value that needs to be fine tuned to be only slightly larger than your longest backup, but rather a measure to eventually detect and kill hung backups. The only really problematic 'resource' a hung backup seems to consume is that it counts in terms of $Conf{MaxBackups}, thus preventing or postponing running other backups (well, yes, a hung rsync might consume a considerable amount of (swappable) memory). I've read of people using timeout values of 72000 (20h) or more. Regards, Holger - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
Re: [BackupPC-users] BackupPC_dump Segmentation Fault
Art writes: I have been running BackupPC for over two years on a Mepis Linux box. All was well until early in the month when one of the WinXP boxes mysteriously stopped backing up. I have upgraded BackupPC to 2.1.2pl1 (using the Mepis Kpackage utility). I have used CPAN up re-install File::RsyncP. I have replaced the rsyncd on the PC with cygwin-rsyncd-2.6.8_0.zip from SourceForge. I ran BackupPC_dump manually (with -v) and all went well until... same 644 400/401 5304636 old pictures/2006/2006-07-08/2006-07-08 15-29-04.JPG same 644 400/401 5124612 old pictures/2006/2006-07-08/2006-07-08 15-29-05.JPG same 644 400/401 5098832 old pictures/2006/2006-07-08/2006-07-08 15-30-07.JPG same 644 400/401 5422218 old pictures/2006/2006-07-08/2006-07-08 15-30-09.JPG Segmentation fault The old pictures directory has 35,377 files encompassing 93.6 GB. Have I hit a wall or is there a workaround??? Have you installed that latest File::RsyncP (0.68)? Sounds like you have, but I wanted to confirm. What are your perl and Compress::Zlib versions? Any chance you are out of memory? Also, the next step is to get a stack trace out of perl. It's important to know which process it is and where it is failing. Craig - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
Re: [BackupPC-users] reinstalling backuppc
David writes: I would like to reinstall Linux on new hard disks and copy the existing backuppc archives from the old disk to the new. What are the potential pitfalls? Can I just copy the whole directory of backuppc recursively? How can I copy and preserve hard links accross hard disks? There is a lot of discussion of these issues on the list. It is quite time consuming to copy the pool because of all the hardlinks. The easiest approach is to start fresh with a new BackupPC install and keep the old disks around for a while in case you need one of the old backups. Craig - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
Re: [BackupPC-users] avoidable failure?
Cristian writes: First, I'd like to thank Craig and his collaborators, who gave us this great tool that simplifies our lives greatly. I use BackupPC for many years already, in many settings, and I couldn't think of a better way of dealing with this thorny requirement. Thanks. So, once in a while, I get errors like this: - Xfer PIDs are now 9356,9357 [ skipped 6674 lines ] finish: removing in-process file ctibirna-work/MEF/CVS-HEAD/GIREF/src/commun/Adaptation/.makedep [ skipped 39 lines ] Done: 15 files, 106665 bytes Got fatal error during xfer (aborted by signal=ALRM) Backup aborted by user signal --- It is failing because an ALRM (alarm) signal got delivered to the process. You should try increasing $Conf{ClientTimeout} significantly (eg: 10x). Craig - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
[BackupPC-users] New version of BackupPC 3?
Hello! I was wondering if there will be a new release of the BackupPC 3 code? I know that there was at least one change (the permissions on web-edited files), and at least one bug (the GUI brings up the oldest log file when you click on the LOG link) that I am specifically interested in. Are we near to another beta release, or will there be a RC release (or even 3.0.0) in the near future? I must say that I have been *extremely* happy with BackupPC 3 so far. I am excited to see it in an offical release version! I do not want to start updating my production servers with beta code... :) Tim Massey - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
Re: [BackupPC-users] [Semi-OT] Encrypting backup partitions?
On 01/11 08:20 , Bradley Alexander wrote: I (currently) have a 200GB partition for backups, and I was considering using Truecrypt on-the-fly encryption. I'm still on the fence regarding whether to set it up as individual containers or one large one. I don't know anything about this encryption tool; but I would suspect that you'll need to set it all up as one encrypted store. Otherwise the hardlinks in the pool won't work correctly. -- Carl Soderstrom Systems Administrator Real-Time Enterprises www.real-time.com - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
Re: [BackupPC-users] Email reports
On Thu, 2007-01-11 at 10:30 +0100, Francisco Daniel Igual Peña wrote: Hi, Is it possible that Backuppc sends weekly (or daily, dont mind) reports to a specific email even though there are no errors with the backups?. I want something like that, but I dont know how to do it. I have a daily status email and RSS feed. You can edit out the RSS stuff if you want just email. The status email address is configurable. I wrote up a little howto for myself (I am running backuppc-2.1.1, so adjustments may be necessary if using a newer version): BackupPC RSS feed and email status HOWTO --- 1. I created a script[see step 5] called BackupPC_statusUpdate modeled on BackupPC_sendEmail. The script parses the backup status of each host, creates an RSS feed and also sends the information by email. BackupPC_statusUpdate resides in $BinDir (/usr/lib/backuppc/bin/ in my case) and runs once each night. 2. Added $Conf{EMailStatusUserName} to the main config file /var/lib/backuppc/conf/config.pl for email address(es) to receive nightly status emails: # $Conf{EMailFromUserName} = 'backuppc'; +# +# Destination address for daily positive status email. +# +$Conf{EMailStatusUserName} = '[EMAIL PROTECTED]'; # # Destination address to an administrative user who will receive a 3. Added a call to BackupPC_statusUpdate in BackupPC_nightly (note the addition of the semicolon on the first system command below): if ( $opts{m} ) { print(log BackupPC_nightly now running BackupPC_sendEmail\n); !system($BinDir/BackupPC_sendEmail); +# RSS and positive status email +# +print(log BackupPC_nightly now running BackupPC_statusUpdate\n); +system($BinDir/BackupPC_statusUpdate); } 4. Added header (to advertise feed to RSS readers e.g. Firefox) on my backup server documentation webpage (this can be any spot viewable from your intranet) at /var/www/localhost/htdocs/index.html. This is an optional step. The link path is the place in the webroot that the main script writes the xml file. /style +link rel=alternate type=application/rss+xml + href=backuppc/backuppc_status.xml title=BackupPC RSS feed /head 5. BackupPC_statusUpdate #!/usr/bin/perl #= -*-perl-*- # # BackupPC_statusUpdate # # DESCRIPTION # # This script implements a positive status email and an RSS feed. # # The script is called from BackupPC_nightly. # # AUTHOR # Travis Fraser [EMAIL PROTECTED] # # Credit to Rich Duzenbury for the original idea. # # # Requires XML::RSS # # Edit the variable $serverName to suit depending on DNS status on your # network. # Edit the use lib ... in the 3rd line of code below. # Edit the $base_url in the RSS section to reflect the correct path to # the cgi page. # Edit the $rss-save ... line near the end of the script to suit. # # use strict; no utf8; # # The lib path needs to match that in the stock backuppc files. # use lib /usr/lib/backuppc/lib; use BackupPC::Lib; use XML::RSS; use Data::Dumper; use Getopt::Std; use DirHandle (); use vars qw($Lang $TopDir $BinDir %Conf); # # Variables # my($fullTot, $fullSizeTot, $incrTot, $incrSizeTot, $str, $mesg, $strNone, $strGood, $hostCntGood, $hostCntNone); $hostCntGood = $hostCntNone = 0; my $serverName = '192.168.1.3'; # # Initialize # die(BackupPC::Lib-new failed\n) if ( !(my $bpc = BackupPC::Lib-new) ); $TopDir = $bpc-TopDir(); $BinDir = $bpc-BinDir(); %Conf = $bpc-Conf(); $Lang = $bpc-Lang(); $bpc-ChildInit(); my $err = $bpc-ServerConnect($Conf{ServerHost}, $Conf{ServerPort}); if ( $err ) { print(Can't connect to server ($err)\n); exit(1); } # # Retrieve status of hosts # my $reply = $bpc-ServerMesg(status hosts); $reply = $1 if ( $reply =~ /(.*)/s ); my(%Status, %Info, %Jobs, @BgQueue, @UserQueue, @CmdQueue); eval($reply); # # Ignore status related to admin and trash jobs foreach my $host ( grep(/admin/, keys(%Status)) ) { delete($Status{$host}) if ( $bpc-isAdminJob($host) ); } delete($Status{$bpc-trashJob}); # # Set up RSS feed # my $now = $bpc-timeStamp(time); # # The cgi page in this case is over HTTPS # my $base_url = 'https://' . $serverName . '/cgi-bin/BackupPC_Admin'; my $rss = new XML::RSS (version = '2.0', encoding = 'ISO-8859-1');
Re: [BackupPC-users] schedule problem since using latest beta
Brad writes: On all of the servers we have installed the latest beta I have seen a problem with the blackout periods. They appear to be ignored or interpreted incorrectly. I am using the default blackout periods and schedules in all my configs but all the incremental backups are starting in the middle of the day, 11am mostly. By default, Blackouts only apply after a host has been successfully pinged 7 times in a row, which could be ~1 week of elapsed time. A sequence of 3 failures (eg: host down) resets the count. Look in the host summary page and you should see text like: - Pings to HOST have succeeded 77 consecutive times. - Because HOST has been on the network at least 7 consecutive times, it will not be backed up from 1:00 to 8:00 on Sun, Mon, Tue, Wed, Thu, Fri, Sat. What does your's say? Finally, please check your settings of $Conf{BlackoutPeriods} $Conf{BlackoutBadPingLimit} $Conf{BlackoutGoodCnt} Craig - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
Re: [BackupPC-users] Proper way to schedule Archive jobs
Tim writes: The first is that it overwrites the previous archive daily. Is it possible to get ArchiveHost/TarCreate to use the backup number in the file name even when you use -1 as the backup job number? You could modify the code to use the date in the file name. Is there a way to launch an archive on a regular basis where the jobs are recorded and managed within BackupPC? The proper way to do it is to create an archive request file (see the CGI code to see how to do that) and to use BackupPC_serverMesg to send an archive request (again, see the CGI code to see how to do that) to BackupPC. BackupPC will then run the archive and keep the log information about what it did. Craig - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
Re: [BackupPC-users] Email reports
Travis Fraser wrote: On Thu, 2007-01-11 at 10:30 +0100, Francisco Daniel Igual Peña wrote: Hi, Is it possible that Backuppc sends weekly (or daily, dont mind) reports to a specific email even though there are no errors with the backups?. I want something like that, but I dont know how to do it. I have a daily status email and RSS feed. You can edit out the RSS stuff if you want just email. The status email address is configurable. I wrote up a little howto for myself (I am running backuppc-2.1.1, so adjustments may be necessary if using a newer version): Travis, that looks like some fantastic work!! Certainly something that i think is worthwhile. I only wish i knew perl so i could knock out stuff like this ;) I haven't tested myself, but if you have a sample email/rss output could you post so we can see what it looks like? Has this been added to any feature requests? implemented in version 3? If not i would like to see it become a feature. A daily email to send to an admin about the status of the system and good/ bad backups etc is worthwhile. In my experience sometimes the end user in charge of monitoring the state of hosts being backed up is a little slack. Regards, Les BackupPC RSS feed and email status HOWTO --- 1. I created a script[see step 5] called BackupPC_statusUpdate modeled on BackupPC_sendEmail. The script parses the backup status of each host, creates an RSS feed and also sends the information by email. BackupPC_statusUpdate resides in $BinDir (/usr/lib/backuppc/bin/ in my case) and runs once each night. 2. Added $Conf{EMailStatusUserName} to the main config file /var/lib/backuppc/conf/config.pl for email address(es) to receive nightly status emails: # $Conf{EMailFromUserName} = 'backuppc'; +# +# Destination address for daily positive status email. +# +$Conf{EMailStatusUserName} = '[EMAIL PROTECTED]'; # # Destination address to an administrative user who will receive a 3. Added a call to BackupPC_statusUpdate in BackupPC_nightly (note the addition of the semicolon on the first system command below): if ( $opts{m} ) { print(log BackupPC_nightly now running BackupPC_sendEmail\n); !system($BinDir/BackupPC_sendEmail); +# RSS and positive status email +# +print(log BackupPC_nightly now running BackupPC_statusUpdate\n); +system($BinDir/BackupPC_statusUpdate); } 4. Added header (to advertise feed to RSS readers e.g. Firefox) on my backup server documentation webpage (this can be any spot viewable from your intranet) at /var/www/localhost/htdocs/index.html. This is an optional step. The link path is the place in the webroot that the main script writes the xml file. /style +link rel=alternate type=application/rss+xml + href=backuppc/backuppc_status.xml title=BackupPC RSS feed /head 5. BackupPC_statusUpdate #!/usr/bin/perl #= -*-perl-*- # # BackupPC_statusUpdate # # DESCRIPTION # # This script implements a positive status email and an RSS feed. # # The script is called from BackupPC_nightly. # # AUTHOR # Travis Fraser [EMAIL PROTECTED] # # Credit to Rich Duzenbury for the original idea. # # # Requires XML::RSS # # Edit the variable $serverName to suit depending on DNS status on your # network. # Edit the use lib ... in the 3rd line of code below. # Edit the $base_url in the RSS section to reflect the correct path to # the cgi page. # Edit the $rss-save ... line near the end of the script to suit. # # use strict; no utf8; # # The lib path needs to match that in the stock backuppc files. # use lib /usr/lib/backuppc/lib; use BackupPC::Lib; use XML::RSS; use Data::Dumper; use Getopt::Std; use DirHandle (); use vars qw($Lang $TopDir $BinDir %Conf); # # Variables # my($fullTot, $fullSizeTot, $incrTot, $incrSizeTot, $str, $mesg, $strNone, $strGood, $hostCntGood, $hostCntNone); $hostCntGood = $hostCntNone = 0; my $serverName = '192.168.1.3'; # # Initialize # die(BackupPC::Lib-new failed\n) if ( !(my $bpc = BackupPC::Lib-new) ); $TopDir = $bpc-TopDir(); $BinDir = $bpc-BinDir(); %Conf = $bpc-Conf(); $Lang = $bpc-Lang(); $bpc-ChildInit(); my $err = $bpc-ServerConnect($Conf{ServerHost}, $Conf{ServerPort}); if ( $err ) { print(Can't connect to server ($err)\n); exit(1); } # # Retrieve status of hosts # my $reply = $bpc-ServerMesg(status hosts); $reply = $1 if (
Re: [BackupPC-users] Email reports
Francisco writes: Is it possible that Backuppc sends weekly (or daily, dont mind) reports to a specific email even though there are no errors with the backups?. I want something like that, but I dont know how to do it. I really don't like programs that annoy you with emails. That said, I agree there hasn't been a way to be sure BackupPC is running ok. In BackupPC 3.0.0beta there is a new option to BackupPC_sendEmail: -c check if BackupPC is alive and send an email if not That allows you to run BackupPC_sendEmail -c from cron and it will send an email to $Conf{EMailAdminUserName} if it can't contact the BackupPC server. The existing email options should handle the cases where BackupPC is running but backups are failing. Craig - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
Re: [BackupPC-users] blackout fulls incrs differently?
brien writes: I'd like to run full backups at night (say, 10pm-2am), but run incrementals every 2 hours from 6am-6pm. There doesn't seem to be any way to do this. Unless, maybe I can use a predump script to test the time and $type and abort fulls that try to run during the day? It would be annoying to see a lot of bogus errors, though. Any ideas? You could disable automatic backups and schedule eveything from cron using BackupPC_serverMesg. Or you could - set $Conf{FullPeriod} to, say, 0.9, - set $Conf{IncrPeriod} to 2/24 = 0.08, - use blackouts on the midnight - 6am and 6pm-10pm windows - kick things off by starting the first full backup at 10pm. The drawback with this approach is there no guarantee that the full backup won't shift to during the day (eg: if a server is down). Craig - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
[BackupPC-users] searching backups
I am looking for a file that I think is in the backups somewhere. Is there a way to search for files saved in backups? - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
[BackupPC-users] CGI scripts to manage removable hard drive media.
Hello! Here are the scripts I use to manage the removable hard drive media used to store daily archives of my backup servers via the GUI, instead of from the command line. Briefly, the system is set up like this: BackupPC's pool is stored on a large internal hard drive. Every day at a little after 7:00 A.M., the backup server starts an archive of each host, which is stored on a second hard drive that is mounted in a removable tray. Once this is complete, the user can shut down the server, remove the hard drive, replace it with a different one, and turn the server back on. Once the new drive is in place, it is repartitioned, reformatted and remounted in place, ready for the next archive. There are two scripts that make this happen. The first one simply shuts the server down. The second one handles the repartitioning, reformatting and remounting. There is absolutely no reason why this couldn't be handed by simply ssh'ing into the server. Except that these servers are destined for network administrators for whom the command line is a tremendously evil thing, and if you try to sell them a solution that contains instructions like Use putty, log into the server and type this command, they will say no. Hence, the CGI scripts... Because the scripts will be run by the webserver, they will be run with its permissions, which likely do not include the ability to shut down the server, or other such commands. The way I have done this is to use sudo, with the proper lines in sudoers. I've tried to make the commands as specific as posssible, to avoid possible security issues. To partition the drive, I am echoing responses to the fdisk command. I looked into parted, but I could not find a clean way of getting it to create a single large partition without knowing how big the partition was. Seeing as this will be used with drives of different sizes, I decided to stick with fdisk. Also, the HTML files that are cat to the user are created simply by saving any old BackupPC HTML page to a file, and chopping the part before the main body DIV into the top file, and the part after the main body DIV into the bottom file. If you have any suggestions as to how to make this script better, I would be happy to hear them. Otherwise, I hope they are useful to someone else. Tim Massey #!/bin/sh # shutdown.cgi - Shut down server echo Content-type: text/html echo cat bpc_top.html echo div id=\Content\ div class=\h1\Shut Down Server/div ph2The system is being shut down!/h2/p pThis will take approximately 60 seconds. Do not remove the drive before the system has powered itself off./p cat bpc_bottom.html sudo /sbin/shutdown -h now /dev/null 21 exit #!/bin/sh # instmedia.cgi - Install new media for BackupPC Archive echo Content-type: text/html echo cat bpc_top.html echo div id=\Content\ div class=\h1\Initialize Removable Media/div ph2Initializing Removable Media/h2/p pThis will take approximately 10 minutes to complete, depending upon the size of the removable drive. Do not naviagate away from this page./p echo PBUnmounting removable drive./B/P sudo /bin/umount /var/lib/BackupPC/removable 21 if [ `sudo /bin/df /var/lib/BackupPC/removable | grep /var/lib/BackupPC/removable | wc -l` = 1 ]; then echo PH2Error: drive did not unmount./H2/P cat bpc_bottom.html exit fi echo PBCreating proper partition on drive./B/P PThis will take approximately 45 seconds. Please wait./P # Pipe responses for fdisk command via echo. # This does the following: # The first series of lines will delete all partitions on a drive with up to 9 # partitions. It does this by having pairs of delete commands: d9, d8, etc. # until it gets to the end. When there's just one partition, fdisk # doesn't ask for a number, so the last one is just a d. # This will actually generate lots of errors in practice: when there are no # more partitions left, the d's will generate an error saying that there are # no partitions to delete, and the numbers are interpreted as nonsense # commands. However, this is harmless. # It then goes through the sequence to create a new partition: # n (New partition) # p (Primary parition) # 1 (First partition) # ENTER (Default starting cylinder is the first one) # ENTER (Default ending cylinder is the last one) # w (Write the changes to disk and exit) # Several newlines are added at the end in case something goes wrong. # Three newlines in a row is interpreted by the fdisk command by # exiting immediately. echo d 9 d 8 d 7 d 6 d 5 d 4 d 3 d 2 d n p 1 w | sudo /sbin/fdisk /dev/hdc /dev/null echo PBFormatting partition for use./B/P PThis can take up to 10 minutes. Please wait./P sudo /sbin/mke2fs -j -m 1 -LRemovableData /dev/hdc1 /dev/null echo PBMounting drive./B/P sudo /bin/mount /var/lib/BackupPC/removable 21 if [ `sudo /bin/df /var/lib/BackupPC/removable | grep /var/lib/BackupPC/removable | wc -l` = 0 ]; then echo PH2Error:
Re: [BackupPC-users] searching backups
On 01/15 02:09 , Krsnendu dasa wrote: I am looking for a file that I think is in the backups somewhere. Is there a way to search for files saved in backups? find | grep filename -- Carl Soderstrom Systems Administrator Real-Time Enterprises www.real-time.com - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
[BackupPC-users] What happens, if client rsync process crashes?
Hi all, I am very new to backupPC - and I like it more more and more every day. ;-) I made a direcory with four 1MB files and some smaller text files for test purposes. Today, I tried some error cases and killed the rsync process on my client machine while syncing with xfermethod=rsync. After killing the process, the cgi interface showed this backup run as OK, but most of my test files were not backed up. The files transfered until my interrupt seem to be in the archive, all other files lack. Is this the expected behaviour? If yes: How can I distinguish between an correctly backed up archive and an archive, that is half backed up due to a process crash on the client machine? Thanks a lot, Clemente Btw: I use backuppc as packaged by debian maintainers (sarge). It is v2.1.1 -- Clemens von Musil [EMAIL PROTECTED] Je weniger die Leute davon wissen, wie Würste und Gesetze gemacht werden, desto besser schlafen sie. (Bismarck) - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
Re: [BackupPC-users] Email reports
On Tue, 2007-01-16 at 08:24 +1100, Les Stott wrote: Travis Fraser wrote: On Thu, 2007-01-11 at 10:30 +0100, Francisco Daniel Igual Peña wrote: Hi, Is it possible that Backuppc sends weekly (or daily, dont mind) reports to a specific email even though there are no errors with the backups?. I want something like that, but I dont know how to do it. I have a daily status email and RSS feed. You can edit out the RSS stuff if you want just email. The status email address is configurable. I wrote up a little howto for myself (I am running backuppc-2.1.1, so adjustments may be necessary if using a newer version): Travis, that looks like some fantastic work!! Certainly something that i think is worthwhile. I only wish i knew perl so i could knock out stuff like this ;) I haven't tested myself, but if you have a sample email/rss output could you post so we can see what it looks like? A typical email looks like so (for an RSS screenshot, I can email that later): To: [EMAIL PROTECTED] Subject: BackupPC status: 4 hosts with good backups Date: Mon, 15 Jan 2007 01:01:01 -0500 Host: crescent Full Count: 2Full age/days: 435.6 Full Size/GB: 0.15 Speed MB/sec: 3.03 Incremental Count: 0 Incremental Age/Days: State: idle Last Attempt: nothing to do Host: marmolata Full Count: 9Full age/days: 3.2 Full Size/GB: 0.59 Speed MB/sec: 3.66 Incremental Count: 6 Incremental Age/Days: 0.2 State: idle Last Attempt: nothing to do Host: pigwin Full Count: 5Full age/days: 53.4 Full Size/GB: 1.28 Speed MB/sec: 6.30 Incremental Count: 1 Incremental Age/Days: 212.3 State: backup starting Last Attempt: no ping (host not found) Host: sweetpea Full Count: 9Full age/days: 3.2 Full Size/GB: 0.94 Speed MB/sec: 1.91 Incremental Count: 6 Incremental Age/Days: 0.2 State: backup starting Last Attempt: nothing to do -- Travis Fraser [EMAIL PROTECTED] - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
Re: [BackupPC-users] Email reports
Hi, Travis Fraser wrote: On Thu, 2007-01-11 at 10:30 +0100, Francisco Daniel Igual Peña wrote: Hi, Is it possible that Backuppc sends weekly (or daily, dont mind) reports to a specific email even though there are no errors with the backups?. I want something like that, but I dont know how to do it. I have a daily status email and RSS feed. You can edit out the RSS stuff if you want just email. The status email address is configurable. I wrote up a little howto for myself (I am running backuppc-2.1.1, so adjustments may be necessary if using a newer version): Travis, that looks like some fantastic work!! Certainly something that i think is worthwhile. I only wish i knew perl so i could knock out stuff like this ;) I haven't tested myself, but if you have a sample email/rss output could you post so we can see what it looks like? Has this been added to any feature requests? implemented in version 3? If not i would like to see it become a feature. A daily email to send to an admin about the status of the system and good/ bad backups etc is worthwhile. In my experience sometimes the end user in charge of monitoring the state of hosts being backed up is a little slack. Maybe a project like this can be integrated into BackupPC: http://sourceforge.net/projects/backupmon ? Michael. - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/
[BackupPC-users] Switching cipher
Hi, Just wanted to share this interesting test result I found about switching the ssh cipher. I am running the 3.0.0 beta3 in production on a fairly small centos4 server (basic spec but with a 1.5TB raid5 setup) and backing up a variety of windows, linux and solaris across a 100mbit lan. I wanted to see whether I could speed up the data transfer without resorting to rsh or applying the no-encryption patch for ssh, so I did some tests and found that arcfour was the fastest cipher for me (seems to be a common conclusion), by a factor of about 2. Interestingly when I start using this in the backuppc ssh config it has made huge improvements to the backup completion time by a factor of up to 10. Here are stats showing the difference, the cipher was switched after backup 21: Backup#TypeFilled Level Start Date Duration/mins Age/days Server Backup Path 15 fullyes 0 12/28 20:00 811.7 18.5 /home/backuppc/data/pc/oddball/15 18 incrno 1 1/1 01:00 124.9 15.3 /home/backuppc/data/pc/oddball/18 19 incrno 1 1/2 01:00 137.2 14.3 /home/backuppc/data/pc/oddball/19 20 incrno 1 1/3 01:00 124.2 13.3 /home/backuppc/data/pc/oddball/20 21 incrno 1 1/4 01:00 145.2 12.3 /home/backuppc/data/pc/oddball/21 22 fullyes 0 1/13 12:43 212.4 2.8 /home/backuppc/data/pc/oddball/22 23 incrno 1 1/14 12:14 7.7 1.8 /home/backuppc/data/pc/oddball/23 24 incrno 1 1/15 12:14 7.5 0.8 /home/backuppc/data/pc/oddball/24 Totals Existing Files New Files Backup# Type#Files Size/MB MB/sec #Files Size/MB #Files Size/MB 15 full269365 5971.7 0.1299242 863.5 180563 5115.6 18 incr2151256.2 0.03133521.2855 235.1 19 incr2161258.2 0.03150925.1686 233.2 20 incr2167260.2 0.03151826.2681 234.0 21 incr2173262.1 0.03124426.9962 235.3 22 full268940 5993.7 0.47268640 5722.1 11914 279.1 23 incr2136257.2 0.56159818.12965239.3 24 incr2142259.1 0.57149322.1682 237.1 I'd be interested to know if others have seen similar results. Phil - Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT business topics through brief surveys - and earn cash http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV ___ BackupPC-users mailing list BackupPC-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/backuppc-users http://backuppc.sourceforge.net/