Re: [BackupPC-users] How do I use an external USB drive as backup target?

2010-02-14 Thread Jeffrey J. Kosowsky
Note excuse my top posting since my comments reference the entire
thread here...

John, may I politely suggest that you will get more help and more
sympathy if you focus your time asking precise, relevant, and
well-formed questions rather than repeatedly criticizing the
documentation, rehashing FAQ's, and asking broad hypotheticals that
are not relevant to most people's conception of backups.

This is an open source project. If you think the documentation or
anything else is lacking, you are more than welcome to contribute
enhancements.

John Hudak wrote at about 14:32:51 -0500 on Sunday, February 14, 2010:
 > my comments are interspersed below...
 > 
 > On Sat, Feb 13, 2010 at 6:40 PM, Jeffrey J. Kosowsky
 > wrote:
 > 
 > > John Hudak wrote at about 12:06:29 -0500 on Saturday, February 13, 2010:
 > >  > Hello:
 > >  > I am considering using an external USB drive as the storage for my
 > > backups.
 > >  > I am running backup pc under Debian 5.0.
 > > External USB drives are a *BAD* idea for multiple reasons:
 > > - Slow
 > > - Unreliable
 > > - Subject to being disconnected
 > > etc.
 > >
 > > Yes, I know 1, and 2 is 'it depends', and 3 is exactly the reason why I
 > want to use usb drives
 > 
 > 
 > >  > Part 1
 > >  > What do I need to do to configure the USB disk as the target? (e.g. how
 > > do I
 > >  > do it?)
 > >  > The USB disk is currently formatted as a NTFS file system.  Do I *need*
 > > to
 > >  > reformat it to ext3? or other?
 > > - NTFS is not usually used - need to check whether it supports the
 > > types of hard links required for BackupPC
 > >
 > >  >
 > >  > Part 2
 > >  > Assume I am crazy paranoid about preserving backup data and I get a
 > > second
 > >  > USB drive to serve as a backup to the first USB drive.
 > >  > Also assume that I am not concerned about the bandwidth across the
 > > network
 > >  > or the various buses.
 > >  >
 > >  > >From a data reliability standpoint, is it better to run a backup
 > > session to
 > >  > USB drive 1, and then repeat the backup to USB drive 2? OR
 > >  > run a backup session to USB drive 1, and then copy the backup
 > > directories to
 > >  > USB drive 2???
 > > Look at the archives and FAQ - this has been discussed *many* times so
 > > no point in wasting peoples time in rehashing.
 > >
 > > I did a quick search of the archives before asking - I did not find a
 > definitive answer...
 > 
 > 
 > >  > The first approach could have errors in different backed up files on
 > > disk 1
 > >  > or 2 but given the odds, very unlikely that the same exact error would
 > > show
 > >  > up
 > >  > in the same exact way in the same file across both USB disks.
 > >  > OTOH, the second approach would allow the exact error in the backup on
 > > USB
 > >  > disk 1 to be copied to USB disk 2.
 > >  >
 > >  > I am leaning towards repeating the backup on two drives.
 > >  >
 > >  > My understanding is that files that are backed up (using either rsync or
 > >  > smb) are 'encrypted' (for lack of a better word), and to view them I
 > > need to
 > >  > use zcat.-True?
 > >
 > > There is a better word -- *compressed*
 > >
 > So that is the word that is not clearly used in the documentation.  There
 > are many ways that backups can be manipulated: stored in a completely
 > nonstandard/proprietary file system and protocol such as z-san, they can be
 > encrypted, and they can be compressed.  The backup PC doc talks about using
 > compression, but does not state if any compression is used in the default
 > configuration.  Compression is often configuration parameter.  It does not
 > make sense to compress many audio and video formats.  If the data to be
 > backed up consists predominantly of these types of files, then it makes no
 > sense to waste CPU cycles applying compression to get < 5% compression.
 > 
 > 
 > >  > Also, can the backup profile be specified to perform complete data
 > > copies
 > >  > periodically, as opposed to a baseline and then periodic incrementals?
 > >
 > > Read the documentation and FAQ.
 > >
 > I have read the doc, (where IMHO) this should have been clearly stated.
 > Instead the doc frequently introduces a topic with 1-2 sentences, then goes
 > off on a tangent for 1-3 paragraphs about how things were done in a previous
 > verson (completely irrevelant), or talks about what will be comming (again,
 > irrelevant), or points one to another section of the doc, in the middle of
 > some other thread that is related to the topic but does not address the
 > topic at hand, or,  introduces a topic, then talks about 3-4 other ways to
 > accomplish the same thing, without telling the reader exactly how to do the
 > initial topic to begin with.  So my fault...I need to also read FAQs.
 > 
 > >
 > >  > Lastly, does anyone have a statistical number that represents the
 > >  > probability of a backup file (e.g. on the target backup disk) containing
 > > an
 > >  > error introduced
 > >  > by the backup procedure?  I know there are error probabilities for

Re: [BackupPC-users] How do I use an external USB drive as backup target?

2010-02-14 Thread Les Mikesell
John Hudak wrote:
>  External USB drives are a *BAD* idea for multiple reasons:
> - Slow
> - Unreliable
> - Subject to being disconnected
> etc.
> 
> Yes, I know 1, and 2 is 'it depends', and 3 is exactly the reason why I 
> want to use usb drives

Note that there are other ways to disconnect drives that don't share the USB 
disadvantages, like hot-swap SATA enclosures, or ESATA connections.

>  > >From a data reliability standpoint, is it better to run a backup
> session to
>  > USB drive 1, and then repeat the backup to USB drive 2? OR
>  > run a backup session to USB drive 1, and then copy the backup
> directories to
>  > USB drive 2???
> Look at the archives and FAQ - this has been discussed *many* times so
> no point in wasting peoples time in rehashing.
> 
> I did a quick search of the archives before asking - I did not find a 
> definitive answer...

I doubt if there is a definitive answer. It will depend on the scale and your 
circumstances.  I'm more comfortable with an always-mirrored internal pair plus 
offsite copies.

>  > My understanding is that files that are backed up (using either
> rsync or
>  > smb) are 'encrypted' (for lack of a better word), and to view
> them I need to
>  > use zcat.-True?
> 
> There is a better word -- *compressed*
> 
> So that is the word that is not clearly used in the documentation.  

Filename mangling isn't optional, compression is, as
http://backuppc.sourceforge.net/faq/BackupPC.html#configuration_parameters
should make clear.  See the entry for $Conf{CompressLevel}.

> There are many ways that backups can be manipulated: stored in a 
> completely nonstandard/proprietary file system and protocol such as 
> z-san, they can be encrypted, and they can be compressed.  The backup PC 
> doc talks about using compression, but does not state if any compression 
> is used in the default configuration.

??? The defaults are shown, but with a packaged version you need to check to 
see 
if the packager made his own changes.

> Compression is often 
> configuration parameter.  It does not make sense to compress many audio 
> and video formats.  If the data to be backed up consists predominantly 
> of these types of files, then it makes no sense to waste CPU cycles 
> applying compression to get < 5% compression.

Note that backuppc doesn't compress/uncompress every time you touch a file so 
the tradeoffs may be different than you expect. Once the initial version of a 
file has been compressed in the pool, matching copies from the same or other 
backups won't have the operation repeated.

> 
>  > Also, can the backup profile be specified to perform complete
> data copies
>  > periodically, as opposed to a baseline and then periodic
> incrementals?
> 
> Read the documentation and FAQ.
> 
> I have read the doc, (where IMHO) this should have been clearly stated.  
> Instead the doc frequently introduces a topic with 1-2 sentences, 

I'm not sure what you are reading here, but set the first 3 variables under the 
'What to backup and when to do it' section if you don't like the default 
once-a-week full:
$Conf{FullPeriod} = 6.97;
$Conf{IncrPeriod} = 0.97;
$Conf{FullKeepCnt} = 1;
The rest are for special cases that you can figure out if you need them.

 > then
> goes off on a tangent for 1-3 paragraphs about how things were done in a 
> previous verson (completely irrevelant),

Irrelevant to you, perhaps.  Not so for people who understood the importance of 
backups a few years ago and now are upgrading.

> or talks about what will be 
> comming (again, irrelevant), or points one to another section of the 
> doc, in the middle of some other thread that is related to the topic but 
> does not address the topic at hand, or,  introduces a topic, then talks 
> about 3-4 other ways to accomplish the same thing, without telling the 
> reader exactly how to do the initial topic to begin with.  So my 
> fault...I need to also read FAQs.

Just look at what the config options let you specify.

> The probability is either 0 if no bugs in the software (or your
> configuration of it) or 100% if bugs in the software and your dataset
> triggers the bug. Your question is not very well-framed and pretty
> meaningless. I suggest you learn a bit more about backup in general
> and backuppc in particular. There is a lot of good documentation on
> BackupPC in the Wikki and in the archives, I suggest you reference it...
> 
> 
> Well, in the extremely simplistic and ideal case, it is 0 or 1.  In the 
> real world, where algorithms are badly designed, or implemented, or 
> both, then problems arise due to things not originally considered in the 
> development which cause errors to arise in the backup procedure that is 
> not about the conversion of 0/1's in computer memory or the transferring 
> of them to tape, disk, etc.

It is still 0 or 1, and if you knew what was wrong with a proces

Re: [BackupPC-users] Odd 'unexpected repeated share name error'

2010-02-14 Thread Craig Barratt
John writes:

> The reason for the current layout is to allow the existing critical
> backups to complete even if it takes days to backup the newer shares
> (filesystems). So it's a temporary thing, but I could see somebody
> wanting to put /home first to prioritize it's backup over say /.
> Is the bug going to be fixed for the beta3 or final release?

Yes.

> Also on another topic, is the elimination of zombie processes on the
> todo list for the 3.2 release?

Yes.

If you want to try the two fixes, here's a patch against 3.2.0beta1
(note: I haven't tested this yet - just did this sitting on a plane).

Craig

--- bin/BackupPC_dump   2010-01-24 17:30:43.0 -0800
+++ bin/BackupPC_dump   2010-02-14 12:02:14.859375000 -0800
@@ -623,6 +623,7 @@
 #
 # Now backup each of the shares
 #
+my $shareDuplicate = {};
 for my $shareName ( @$ShareNames ) {
 local(*RH, *WH);
 
@@ -632,11 +633,17 @@
 $shareName = encode("utf8", $shareName);
 $stat{xferOK} = $stat{hostAbort} = undef;
 $stat{hostError} = $stat{lastOutputLine} = undef;
-if ( -d "$Dir/new/$shareName" ) {
+if ( $shareName eq "" ) {
+print(LOG $bpc->timeStamp,
+  "unexpected empty share name skipped\n");
+next;
+}
+if ( $shareDuplicate->{$shareName} ) {
 print(LOG $bpc->timeStamp,
   "unexpected repeated share name $shareName skipped\n");
 next;
 }
+$shareDuplicate->{$shareName} = 1;
 
 UserCommandRun("DumpPreShareCmd", $shareName);
 if ( $? && $Conf{UserCmdCheckStatus} ) {
@@ -915,6 +922,10 @@
 #
 last;
 }
+#
+# Wait for any child processes to exit
+#
+1 while ( wait() >= 0 );
 }
 
 #

--
SOLARIS 10 is the OS for Data Centers - provides features such as DTrace,
Predictive Self Healing and Award Winning ZFS. Get Solaris 10 NOW
http://p.sf.net/sfu/solaris-dev2dev
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
List:https://lists.sourceforge.net/lists/listinfo/backuppc-users
Wiki:http://backuppc.wiki.sourceforge.net
Project: http://backuppc.sourceforge.net/


Re: [BackupPC-users] BackupPC not cleaning out cpool.

2010-02-14 Thread Craig Barratt
Joe,

Yes, you have the IO::Dirent problem, and it sounds like it is
fixed, since BackupPC_nightly reports non-zero information.
However, it doesn't report that it removed any files, so every
cpool file (at least for the output you included) has at least
2 links.

However, BackupPC_trashClean (which is started once when BackupPC
first starts) will also have the same problem.  Did you re-start
BackupPC after patching $IODirentOk?

You should check to see if /var/data/backuppc/trash has a lot
of files.  If so, that's the problem.  After those files are
removed, then BackupPC_nightly will find files that only have
one link, and will remove them.

Craig

--
SOLARIS 10 is the OS for Data Centers - provides features such as DTrace,
Predictive Self Healing and Award Winning ZFS. Get Solaris 10 NOW
http://p.sf.net/sfu/solaris-dev2dev
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
List:https://lists.sourceforge.net/lists/listinfo/backuppc-users
Wiki:http://backuppc.wiki.sourceforge.net
Project: http://backuppc.sourceforge.net/


Re: [BackupPC-users] How do I use an external USB drive as backup target?

2010-02-14 Thread John Hudak
my comments are interspersed below...

On Sat, Feb 13, 2010 at 6:40 PM, Jeffrey J. Kosowsky
wrote:

> John Hudak wrote at about 12:06:29 -0500 on Saturday, February 13, 2010:
>  > Hello:
>  > I am considering using an external USB drive as the storage for my
> backups.
>  > I am running backup pc under Debian 5.0.
> External USB drives are a *BAD* idea for multiple reasons:
> - Slow
> - Unreliable
> - Subject to being disconnected
> etc.
>
> Yes, I know 1, and 2 is 'it depends', and 3 is exactly the reason why I
want to use usb drives


>  > Part 1
>  > What do I need to do to configure the USB disk as the target? (e.g. how
> do I
>  > do it?)
>  > The USB disk is currently formatted as a NTFS file system.  Do I *need*
> to
>  > reformat it to ext3? or other?
> - NTFS is not usually used - need to check whether it supports the
> types of hard links required for BackupPC
>
>  >
>  > Part 2
>  > Assume I am crazy paranoid about preserving backup data and I get a
> second
>  > USB drive to serve as a backup to the first USB drive.
>  > Also assume that I am not concerned about the bandwidth across the
> network
>  > or the various buses.
>  >
>  > >From a data reliability standpoint, is it better to run a backup
> session to
>  > USB drive 1, and then repeat the backup to USB drive 2? OR
>  > run a backup session to USB drive 1, and then copy the backup
> directories to
>  > USB drive 2???
> Look at the archives and FAQ - this has been discussed *many* times so
> no point in wasting peoples time in rehashing.
>
> I did a quick search of the archives before asking - I did not find a
definitive answer...


>  > The first approach could have errors in different backed up files on
> disk 1
>  > or 2 but given the odds, very unlikely that the same exact error would
> show
>  > up
>  > in the same exact way in the same file across both USB disks.
>  > OTOH, the second approach would allow the exact error in the backup on
> USB
>  > disk 1 to be copied to USB disk 2.
>  >
>  > I am leaning towards repeating the backup on two drives.
>  >
>  > My understanding is that files that are backed up (using either rsync or
>  > smb) are 'encrypted' (for lack of a better word), and to view them I
> need to
>  > use zcat.-True?
>
> There is a better word -- *compressed*
>
So that is the word that is not clearly used in the documentation.  There
are many ways that backups can be manipulated: stored in a completely
nonstandard/proprietary file system and protocol such as z-san, they can be
encrypted, and they can be compressed.  The backup PC doc talks about using
compression, but does not state if any compression is used in the default
configuration.  Compression is often configuration parameter.  It does not
make sense to compress many audio and video formats.  If the data to be
backed up consists predominantly of these types of files, then it makes no
sense to waste CPU cycles applying compression to get < 5% compression.


>  > Also, can the backup profile be specified to perform complete data
> copies
>  > periodically, as opposed to a baseline and then periodic incrementals?
>
> Read the documentation and FAQ.
>
I have read the doc, (where IMHO) this should have been clearly stated.
Instead the doc frequently introduces a topic with 1-2 sentences, then goes
off on a tangent for 1-3 paragraphs about how things were done in a previous
verson (completely irrevelant), or talks about what will be comming (again,
irrelevant), or points one to another section of the doc, in the middle of
some other thread that is related to the topic but does not address the
topic at hand, or,  introduces a topic, then talks about 3-4 other ways to
accomplish the same thing, without telling the reader exactly how to do the
initial topic to begin with.  So my fault...I need to also read FAQs.

>
>  > Lastly, does anyone have a statistical number that represents the
>  > probability of a backup file (e.g. on the target backup disk) containing
> an
>  > error introduced
>  > by the backup procedure?  I know there are error probabilities for both
> disk
>  > and tape reads/writes failures, but am wondering if anything like that
>  > exists for the backup software.  (A group I used to work with did this
> sort
>  > of testing, and actually had some statistics on the reliability of
> backup
>  > programs, wrt types of files, sizes, w/wo compression, and the types of
>  > compression.   Not sure the open source community would go through this
> type
>  > of assessment - but thought I'd ask.
>
> The probability is either 0 if no bugs in the software (or your
> configuration of it) or 100% if bugs in the software and your dataset
> triggers the bug. Your question is not very well-framed and pretty
> meaningless. I suggest you learn a bit more about backup in general
> and backuppc in particular. There is a lot of good documentation on
> BackupPC in the Wikki and in the archives, I suggest you reference it...


Well, in the extremely simplistic and ideal c

Re: [BackupPC-users] BackupPC not cleaning out cpool.

2010-02-14 Thread Matthias Meyer
Joseph Holland wrote:

> # du -xsmh /var/data/backuppc/
> 544G/var/data/backuppc/
> 
> 
> Joe.
> 
> On 11/02/2010 22:45, Matthias Meyer wrote:
>> Joseph Holland wrote:
>>
>>
>>> We are having a problem with many of our BackupPC servers in our company
>>> at the moment.  We are running Debian Lenny and BackupPC 3.1.0.  The
>>> data volumes since we upgraded to this version have seemed to just be
>>> filling up at a constant rate.
>>>
>>> The web interface is saying that the pool is 136.98GB but when you do a
>>> "du -smh" on the topdir "/var/data/backuppc" it returns 543GB used.
>>>  
>> Try du -xsmh
>>
>> br
>> Matthias
>>
> 
strangely, indeed.
du delivers a roughly estimation, not an exact value. But we are speaking
about 140GB vs 540GB.

1) "du" and GUI use the same directory? I have /var/lib/backuppc and
not /var/data/backuppc.

2) There are no extra files from you on your /var/data/backuppc? I don't
know how the GUI calculates the usage of space.

Sorry
Matthias
-- 
Don't Panic


--
SOLARIS 10 is the OS for Data Centers - provides features such as DTrace,
Predictive Self Healing and Award Winning ZFS. Get Solaris 10 NOW
http://p.sf.net/sfu/solaris-dev2dev
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
List:https://lists.sourceforge.net/lists/listinfo/backuppc-users
Wiki:http://backuppc.wiki.sourceforge.net
Project: http://backuppc.sourceforge.net/


Re: [BackupPC-users] Trying to get backuppc working

2010-02-14 Thread Tyler J. Wagner
Chris,

What happened when you ran the commands I sent?  Do you have output?  It 
sounds like something caused dpkg to stop during a configuration step, which 
might be resolved with:

apt-get -f install
apt-get dist-upgrade

I'll try to test a karmic install later.

I can say certainly that backuppc works just fine on Ubuntu hardy (the last LTS 
release), either with the stock packages or the ones from my repository.

Regards,
Tyler

On Sunday 14 February 2010 01:49:47 bigonroad wrote:
> I don't really know what to do now. After a day on it, I still haven't got
>  backuppc running on Ubuntu 9.10 server.
> 
> Should I try a different variant, or can you advise me what to do? I am
>  doing a fresh install, and will start from the beginning again.
> 
> I will install everything apache, samba, samba-common-bin, and everything
>  else before backuppc. However, if that doesn't work, I am out of ideas!
> 
> Bless,
> Chris
> 
> +--
> 
> |This was sent by bigonr...@gmail.com via Backup Central.
> |Forward SPAM to ab...@backupcentral.com.
> 
> +--
> 
> 
> 
> ---
> --- SOLARIS 10 is the OS for Data Centers - provides features such as
>  DTrace, Predictive Self Healing and Award Winning ZFS. Get Solaris 10 NOW
> http://p.sf.net/sfu/solaris-dev2dev
> ___
> BackupPC-users mailing list
> BackupPC-users@lists.sourceforge.net
> List:https://lists.sourceforge.net/lists/listinfo/backuppc-users
> Wiki:http://backuppc.wiki.sourceforge.net
> Project: http://backuppc.sourceforge.net/
> 
-- 
"Every normal man must be tempted at times to spit on his hands, hoist
the black flag, and begin to slit throats."
   -- H. L. Mencken

--
SOLARIS 10 is the OS for Data Centers - provides features such as DTrace,
Predictive Self Healing and Award Winning ZFS. Get Solaris 10 NOW
http://p.sf.net/sfu/solaris-dev2dev
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
List:https://lists.sourceforge.net/lists/listinfo/backuppc-users
Wiki:http://backuppc.wiki.sourceforge.net
Project: http://backuppc.sourceforge.net/


[BackupPC-users] Trying to get backuppc working

2010-02-14 Thread bigonroad

I don't really know what to do now. After a day on it, I still haven't got 
backuppc running on Ubuntu 9.10 server.

Should I try a different variant, or can you advise me what to do? I am doing a 
fresh install, and will start from the beginning again.

I will install everything apache, samba, samba-common-bin, and everything else 
before backuppc. However, if that doesn't work, I am out of ideas!

Bless,
Chris

+--
|This was sent by bigonr...@gmail.com via Backup Central.
|Forward SPAM to ab...@backupcentral.com.
+--



--
SOLARIS 10 is the OS for Data Centers - provides features such as DTrace,
Predictive Self Healing and Award Winning ZFS. Get Solaris 10 NOW
http://p.sf.net/sfu/solaris-dev2dev
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
List:https://lists.sourceforge.net/lists/listinfo/backuppc-users
Wiki:http://backuppc.wiki.sourceforge.net
Project: http://backuppc.sourceforge.net/