[Bacula-users] BigInt hit max value in mysql... what now?

2010-08-30 Thread Uwe Schuerkamp
Hi folks,

over the weekend, our mysql bacula db seems to have hit the maximum
possible auto_increment counter for an unsigned int(10) as I'm getting
duplicate key errors on insert's into the File table. 

Is there a way to increase the maximum value or do I need to rebuild
the bacula db from the last working dump?`The file table is around 15G
at the moment, we have 67657255 files with a total db size around 23G.

will

alter table File change FileId FileId unsigned int(20) 

do anything useful? 

Thanks in advance for your comments, 

Uwe 

-- 
NIONEX ist ein Unternehmen der DirectGroup Germany www.directgroupgermany.de

--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Verify differences: SHA1 sum doesn't match but it should

2010-08-30 Thread Henrik Johansen
'Paul Mather' wrote:
On Aug 28, 2010, at 7:12 AM, Steve Costaras wrote:

Could be due to a transient error (transmission or wild/torn read at
time of calculation).  I see this a lot with integrity checking of
files here (50TiB of storage).

Only way to get around this now is to do a known-good sha1/md5 hash of
data (2-3 reads of the file make sure that they all match and that the
file is not corrupted) save that as a baseline and then when doing
reads/compares if one fails do another re-read and see if the first one
was in error and compare that with your baseline.  This is one reason
why I'm switching to the new generation of sas drives that have ioecc
checks on READS not just writes to help cut down on some of this.

Corruption does occur as well and is more probable with the higher the
capacity of the drive.  Ideally you would have a drive that would do
ioecc on reads, plus using T10 PI extensions (DIX/DIF) from drive to
controller up to your file system layer.  It won't always prevent it by
itself but would allow if you have a raid setup to do some self-healing
when a drive reports a non transient (i.e. corrupted sector of data).

However the T10 PI extensions are only on sas/fc drives (520/528 byte
blocks) and so far as I can tell only the new LSI hba's support a small
subset of this (no hardware raid controllers I can find) and have not
seen any support up to the OS/filesystem level.  SATA is not included
at all as the T13 group opted not to include it in the spec.

You could also stick with your current hardware and use a file system
that emphasises end-to-end data integrity like ZFS.  ZFS checksums at
many levels, and has a don't trust the hardware mentality.  It can
detect silent data corruption and automatically self-heal where
redundancy permits.

ZFS also supports pool scrubbing---akin to the patrol reading of many
RAID controllers---for proactive detection of silent data corruption.
With drive capacities becoming very large, the probability of an
unrecoverable read becomes very high.  This becomes very significant
even in redundant storage systems because a drive failure necessitates
a lengthy rebuild period during which the storage array lacks any
redundancy (in the case of RAID-5).  It is for this reason that RAID-6
(ZFS raidz2) is becoming de rigeur for many-terabyte arrays using large
drives, and, specifically, the reason ZFS garnered its triple-parity
raidz3 pool type (in ZFS pool version 17).

Have you ever tried scrubbing a 40+ TB pool ?

I believe Btrfs intends to bring many ZFS features to Linux.

Cheers,

Paul.

--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users
worldwide. Take advantage of special opportunities to increase revenue and
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d

___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


-- 
Med venlig hilsen / Best Regards

Henrik Johansen
hen...@scannet.dk
Tlf. 75 53 35 00

ScanNet Group
A/S ScanNet 

--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Verify differences: SHA1 sum doesn't match but it should

2010-08-30 Thread Steve Costaras
  A little mis-quoted there:

On 2010-08-30 02:59, Henrik Johansen wrote:
 On Aug 28, 2010, at 7:12 AM, Steve Costaras wrote:

 Could be due to a transient error (transmission or wild/torn read at
 time of calculation).  I see this a lot with integrity checking of
 files here (50TiB of storage).

 Only way to get around this now is to do a known-good sha1/md5 hash of
 data (2-3 reads of the file make sure that they all match and that the
 file is not corrupted) save that as a baseline and then when doing
 reads/compares if one fails do another re-read and see if the first one
 was in error and compare that with your baseline.  This is one reason
 why I'm switching to the new generation of sas drives that have ioecc
 checks on READS not just writes to help cut down on some of this.

 Corruption does occur as well and is more probable with the higher the
 capacity of the drive.  Ideally you would have a drive that would do
 ioecc on reads, plus using T10 PI extensions (DIX/DIF) from drive to
 controller up to your file system layer.  It won't always prevent it by
 itself but would allow if you have a raid setup to do some self-healing
 when a drive reports a non transient (i.e. corrupted sector of data).

 However the T10 PI extensions are only on sas/fc drives (520/528 byte
 blocks) and so far as I can tell only the new LSI hba's support a small
 subset of this (no hardware raid controllers I can find) and have not
 seen any support up to the OS/filesystem level.  SATA is not included
 at all as the T13 group opted not to include it in the spec.

 You could also stick with your current hardware and use a file system
 that emphasises end-to-end data integrity like ZFS.  ZFS checksums at
 many levels, and has a don't trust the hardware mentality.  It can
 detect silent data corruption and automatically self-heal where
 redundancy permits.


'Paul Mather' wrote:
 ZFS also supports pool scrubbing---akin to the patrol reading of many
 RAID controllers---for proactive detection of silent data corruption.
 With drive capacities becoming very large, the probability of an
 unrecoverable read becomes very high.  This becomes very significant
 even in redundant storage systems because a drive failure necessitates
 a lengthy rebuild period during which the storage array lacks any
 redundancy (in the case of RAID-5).  It is for this reason that RAID-6
 (ZFS raidz2) is becoming de rigeur for many-terabyte arrays using large
 drives, and, specifically, the reason ZFS garnered its triple-parity
 raidz3 pool type (in ZFS pool version 17).


On 2010-08-30 02:59, Henrik Johansen wrote:
 Have you ever tried scrubbing a 40+ TB pool ?

If the  question was to me, then yes, I have but with the comment that I 
am working with SANs and otherwise
redundant luns/disks that I run ZFS on top of.So the availability 
portion of the disk subsystem is pretty stable
already.  I use ZFS mainly to check/verify data integrity as well as for 
volume management functions.For
performance reasons I am mainly using mirroring.   When pool sizes get 
large 50, 100, or more TiB the problem
is the time it takes to do a scrub and the cpu  i/o costs are high. 
For ~50TiB I would say you would want
to have a subsystem that is capable of 2-3GiB/s. And then increase 
that in proportion with larger sets.
Even then it takes a toll on a system that the primary job is NOT disk 
integrity but to run X application.


--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] BigInt hit max value in mysql... what now?

2010-08-30 Thread Uwe Schuerkamp
On Mon, Aug 30, 2010 at 08:45:50AM +0200, Uwe Schuerkamp wrote:
 Hi folks,
 
 over the weekend, our mysql bacula db seems to have hit the maximum
 possible auto_increment counter for an unsigned int(10) as I'm getting
 duplicate key errors on insert's into the File table. 
 
 Is there a way to increase the maximum value or do I need to rebuild
 the bacula db from the last working dump?`The file table is around 15G
 at the moment, we have 67657255 files with a total db size around 23G.
 
 will
 
 alter table File change FileId FileId unsigned int(20) 
 

Hi folks, 

I was able to extend the value range by issuing 


 ALTER TABLE File CHANGE COLUMN FileId FileId bigint unsigned;

I should have mentioned I'm running bacula 3.x on that server, bacula
5.x seems to use a bigint type value by default. 

All the best, Uwe 

-- 
NIONEX ist ein Unternehmen der DirectGroup Germany www.directgroupgermany.de

--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] BigInt hit max value in mysql... what now?

2010-08-30 Thread Uwe Schuerkamp
Hi folks,

looks like extending the value range did not fix the error
entirely. When running an incr. backup, I receive the following error
message: 

Fatal error: Can't fill File table Query failed: INSERT
INTO File (FileIndex, JobId, PathId, FilenameId, LStat, MD5)SELECT
batch.FileIndex,
batch.JobId, Path.PathId, Filename.FilenameId,batch.LStat, batch.MD5
FROM batch JOIN Path ON
(batch.Path = Path.Path) JOIN Filename ON (batch.Name =
Filename.Name): ERR=Duplicate entry
'0' for key 1

I've tried deleting those rows with 0 from File (1 Row), but
apparently that did not fix the problem. 

Will I need to upgrade to bacula 5.0 in order to resolve the issue? Is
3.x limited to 32bit auto_inc counters or something?

All the best, 

Uwe 
-- 
NIONEX ist ein Unternehmen der DirectGroup Germany www.directgroupgermany.de

--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Verify differences: SHA1 sum doesn't match but it should

2010-08-30 Thread Henrik Johansen
'Steve Costaras' wrote:
  A little mis-quoted there:

On 2010-08-30 02:59, Henrik Johansen wrote:
 On Aug 28, 2010, at 7:12 AM, Steve Costaras wrote:

 Could be due to a transient error (transmission or wild/torn read at
 time of calculation).  I see this a lot with integrity checking of
 files here (50TiB of storage).

 Only way to get around this now is to do a known-good sha1/md5 hash of
 data (2-3 reads of the file make sure that they all match and that the
 file is not corrupted) save that as a baseline and then when doing
 reads/compares if one fails do another re-read and see if the first one
 was in error and compare that with your baseline.  This is one reason
 why I'm switching to the new generation of sas drives that have ioecc
 checks on READS not just writes to help cut down on some of this.

 Corruption does occur as well and is more probable with the higher the
 capacity of the drive.  Ideally you would have a drive that would do
 ioecc on reads, plus using T10 PI extensions (DIX/DIF) from drive to
 controller up to your file system layer.  It won't always prevent it by
 itself but would allow if you have a raid setup to do some self-healing
 when a drive reports a non transient (i.e. corrupted sector of data).

 However the T10 PI extensions are only on sas/fc drives (520/528 byte
 blocks) and so far as I can tell only the new LSI hba's support a small
 subset of this (no hardware raid controllers I can find) and have not
 seen any support up to the OS/filesystem level.  SATA is not included
 at all as the T13 group opted not to include it in the spec.

 You could also stick with your current hardware and use a file system
 that emphasises end-to-end data integrity like ZFS.  ZFS checksums at
 many levels, and has a don't trust the hardware mentality.  It can
 detect silent data corruption and automatically self-heal where
 redundancy permits.


'Paul Mather' wrote:
 ZFS also supports pool scrubbing---akin to the patrol reading of many
 RAID controllers---for proactive detection of silent data corruption.
 With drive capacities becoming very large, the probability of an
 unrecoverable read becomes very high.  This becomes very significant
 even in redundant storage systems because a drive failure necessitates
 a lengthy rebuild period during which the storage array lacks any
 redundancy (in the case of RAID-5).  It is for this reason that RAID-6
 (ZFS raidz2) is becoming de rigeur for many-terabyte arrays using large
 drives, and, specifically, the reason ZFS garnered its triple-parity
 raidz3 pool type (in ZFS pool version 17).


On 2010-08-30 02:59, Henrik Johansen wrote:
 Have you ever tried scrubbing a 40+ TB pool ?

If the  question was to me, then yes, I have but with the comment that I
am working with SANs and otherwise
redundant luns/disks that I run ZFS on top of.So the availability
portion of the disk subsystem is pretty stable
already.  I use ZFS mainly to check/verify data integrity as well as for
volume management functions.For
performance reasons I am mainly using mirroring.   When pool sizes get
large 50, 100, or more TiB the problem
is the time it takes to do a scrub and the cpu  i/o costs are high.
For ~50TiB I would say you would want
to have a subsystem that is capable of 2-3GiB/s. And then increase
that in proportion with larger sets.
Even then it takes a toll on a system that the primary job is NOT disk
integrity but to run X application.

My point exactly.

Scrubbing is problematic when dealing with large datasets that have a
limited lifetime and / or a high changerate (eg D2D backups). There is a
hefty cost associated with scrubbing both in terms of IOPS / CPU cyles
and still you may not be able to cover your entire dataset before data
in it gets removed or changed.

Once a fix for CR6730306 is integrated it may become feasible to
schedule scrubbing operations during off-hours though.

Like most ZFS related stuff it all sounds (and looks) extremely easy but
in reality it is not quite so simple.




--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users
worldwide. Take advantage of special opportunities to increase revenue and
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users

-- 
Med venlig hilsen / Best Regards

Henrik Johansen
hen...@scannet.dk
Tlf. 75 53 35 00

ScanNet Group
A/S ScanNet 

--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 

[Bacula-users] Run backup job from command line in a non-interactive mode

2010-08-30 Thread Sergio Belkin
Hi,

Let's say that you had a problem with last scheduled job, and you need
to backup tonight. Idon't want to create a new job on bacula-dir.conf
only for use once. I'd love to add an entry on crontab file that does
it. But I don't know how to do it (is it possible). For example I can
run:

echo st dir | bconsole

But how should I do in order to run a backup job?

Thanks in advance!!
-- 
--
Sergio Belkin http://www.sergiobelkin.com
Watch More TV http://sebelk.blogspot.com
Sergio Belkin -

--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


[Bacula-users] FileSet for Windows 7

2010-08-30 Thread Ralph Kutschera
Hallo!

   I'd like to kindly ask if someone could post a reasonable Windows 7 
FileSet!

Sincerely,
   Ralph


--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Run backup job from command line in a non-interactive mode

2010-08-30 Thread me
Hello,
*help run 

  Command   Description   

  ===   ===   

  run   Run a job 

  

Arguments:

job=job-name client=client-name   

fileset=FileSet-name level=level-keyword  

storage=storage-namewhere=directory-prefix

when=universal-time-specification   

comment=text yes

  

When at a prompt, entering a period cancels the command.  

   

and then pipe it into bconsole, like you did with the status command.

hth

On Mon, 30 Aug 2010 08:48:48 -0300, Sergio Belkin seb...@gmail.com
wrote:
 Hi,
 
 Let's say that you had a problem with last scheduled job, and you need
 to backup tonight. Idon't want to create a new job on bacula-dir.conf
 only for use once. I'd love to add an entry on crontab file that does
 it. But I don't know how to do it (is it possible). For example I can
 run:
 
 echo st dir | bconsole
 
 But how should I do in order to run a backup job?
 
 Thanks in advance!!


--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


[Bacula-users] Problem with offline clients

2010-08-30 Thread Alex2398
Thank you, the runbefore script solved the problem!!

Truly appreciated.

+--
|This was sent by informat...@dhul.es via Backup Central.
|Forward SPAM to ab...@backupcentral.com.
+--



--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


[Bacula-users] Bacula VMs

2010-08-30 Thread Dan Langille
This just came up on IRC.  As an aid to newcomers, a ready-made VM may  
be useful. A potential user could download and run the VM, without  
having to go through the install process.

If you use VMs (eg VMWare, VirtualBox), then I think this would be a  
great way for you to contribute to the project. And help to lower the  
barrier to entry. Sometimes, the hardest step can be installation and  
just getting it running.

:)

-- 
Dan Langille
http://langille.org/


--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Verify differences: SHA1 sum doesn't match but it should

2010-08-30 Thread Paul Mather
On Aug 30, 2010, at 6:41 AM, Henrik Johansen wrote:

 Like most ZFS related stuff it all sounds (and looks) extremely easy but
 in reality it is not quite so simple.

Yes, but does ZFS makes things easier or harder?

Silent data corruption won't go away just because your pool is large. :-)

(But, this is all getting a bit off-topic for Bacula-users.)

Cheers,

Paul.


--
Sell apps to millions through the Intel(R) Atom(Tm) Developer Program
Be part of this innovative community and reach millions of netbook users 
worldwide. Take advantage of special opportunities to increase revenue and 
speed time-to-market. Join now, and jumpstart your future.
http://p.sf.net/sfu/intel-atom-d2d
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


[Bacula-users] Question about jobs and pools

2010-08-30 Thread Romer Ventura

Hello,

At the moment i have the following pools:
- FileServer-Pool-full
- FileServer-Pool-Incr
- FileServer-Pool-Diff

- MailServer-Pool-fulll
- MailServer-Pool-Incr
- MailServer-Pool-Diff

I have 7 tapes in my library and all of them were put in the Scratch  
pool. I have calculated that a full backup for the FileServer  will  
comfortable fit in 2 tapes and for the MailServer 1 tape.


Now, i would like the Incremental and Differential jobs for the  
FileServer to write to 1 tape and incremental and differential jobs  
for MailServer to write to another tape. This will end up with  
MailServer using 2 tapes and FileServer using 3 tapes. Leaving me 2  
tapes for other backups.


But it is not happening like that, each pool just pulls a new tape  
out of scratch and writes to it, if the job has run, it will append  
to it, but incremental and differential seem to not be able to  
coexists on the same tape.


My question is, what can i do to have incremental and differential  
jobs to be written to 1 tape?


After a week cycle i ended up with:
Pool: FileServPool-Full-tape
+-++---+-+- 
+--+--+-+--+---+--- 
+-+
| MediaId | VolumeName | VolStatus | Enabled | VolBytes|  
VolFiles | VolRetention | Recycle | Slot | InChanger | MediaType |  
LastWritten |
+-++---+-+- 
+--+--+-+--+---+--- 
+-+
|  15 | CNH895 | Full  |   1 | 165,219,038,208 |   
166 |2,592,000 |   1 |1 | 1 | tape  |  
2010-08-25 14:31:09 |
|  16 | CNH896 | Used  |   1 | 139,932,721,152 |   
140 |2,592,000 |   1 |2 | 1 | tape  |  
2010-08-25 18:27:05 |
+-++---+-+- 
+--+--+-+--+---+--- 
+-+

Pool: FileServPool-Diff-tape
+-++---+-+-+-- 
+--+-+--+---+--- 
+-+
| MediaId | VolumeName | VolStatus | Enabled | VolBytes| VolFiles  
| VolRetention | Recycle | Slot | InChanger | MediaType |  
LastWritten |
+-++---+-+-+-- 
+--+-+--+---+--- 
+-+
|  20 | CNH900 | Append|   1 | 778,982,400 |1  
|2,592,000 |   1 |6 | 1 | tape  | 2010-08-29  
18:14:51 |
+-++---+-+-+-- 
+--+-+--+---+--- 
+-+

Pool: FileServPool-Incre-tape
+-++---+-+-+-- 
+--+-+--+---+--- 
+-+
| MediaId | VolumeName | VolStatus | Enabled | VolBytes| VolFiles  
| VolRetention | Recycle | Slot | InChanger | MediaType |  
LastWritten |
+-++---+-+-+-- 
+--+-+--+---+--- 
+-+
|  17 | CNH897 | Append|   1 | 801,239,040 |3  
|  432,000 |   1 |3 | 1 | tape  | 2010-08-27  
20:14:59 |
+-++---+-+-+-- 
+--+-+--+---+--- 
+-+

Pool: MailPool-Full-tape
+-++---+-+- 
+--+--+-+--+---+--- 
+-+
| MediaId | VolumeName | VolStatus | Enabled | VolBytes|  
VolFiles | VolRetention | Recycle | Slot | InChanger | MediaType |  
LastWritten |
+-++---+-+- 
+--+--+-+--+---+--- 
+-+
|  18 | CNH898 | Used  |   1 | 113,341,132,800 |   
114 |2,592,000 |   1 |4 | 1 | tape  |  
2010-08-25 23:50:11 |
+-++---+-+- 
+--+--+-+--+---+--- 
+-+

Pool: MailPool-Incre-tape
+-++---+-+-+-- 
+--+-+--+---+--- 
+-+
| MediaId | VolumeName | VolStatus | Enabled | VolBytes| VolFiles  
| VolRetention | Recycle | Slot | InChanger | MediaType |  
LastWritten |
+-++---+-+-+-- 
+--+-+--+---+--- 
+-+
|  19 | CNH899 | Append|   1 | 

Re: [Bacula-users] Verify differences: SHA1 sum doesn't match but it should

2010-08-30 Thread Tobias Brink
Steve Costaras stev...@chaven.com writes:

  Could be due to a transient error (transmission or wild/torn read at
 time of calculation).   I see this a lot with integrity checking of
 files here (50TiB of storage).

 Only way to get around this now is to do a known-good sha1/md5 hash of
 data (2-3 reads of the file make sure that they all match and that the
 file is not corrupted) save that as a baseline and then when doing
 reads/compares if one fails do another re-read and see if the first
 one was in error and compare that with your baseline. This is one
 reason why I'm switching to the new generation of sas drives that have
 ioecc checks on READS not just writes to help cut down on some of
 this.

 Corruption does occur as well and is more probable with the higher the
 capacity of the drive. Ideally you would have a drive that would
 do ioecc on reads, plus using T10 PI extensions (DIX/DIF) from drive
 to controller up to your file system layer.It won't always prevent
 it by itself but would allow if you have a raid setup to do some
 self-healing when a drive reports a non transient (i.e. corrupted
 sector of data).

First off, thanks for the answers.  The thing is that I am well aware of
the reliability problems of hard drives and I would love to use some
advanced file system like ZFS or btrfs, but I am on Debian and I will
stay on Debian.  And btrfs is not mature enough to be used in production
at the moment.  The other thing is that I do not think that this is an
issue of corruption of the data itself!  As I said I checked the files
against backups and MD5 sums supplied by Debian (several times and from
cold cache) and the data seems to be OK.  The executables that are
reported by Bacula to have changed continue to work well and bug-free
just as before.

So I think this is a problem/bug with either the Postgresql database or
Bacula, not with my hard drives.  I just wonder how something like this
could happen and how I could avoid this.  I'm also not willing to do
additional checksums with other programs (AIDE or similar) because they
take _lots_ of time to run.  With Bacula I get the checksums for free.
I just want to use them to detect corruption on disk from time to time
and because I use VirtualFull and want to know if my differential
backups have missed something.

So I still don't know how to proceed.  Apart from that I will try to
upgrade my director and sd to 5.0.2 as soon as Debian backports are
available and see if the problem goes away.  I will also re-run the
DiskToCatalog after my next differential backup and see if something
is different.

Thanks,
Tobias

--
This SF.net Dev2Dev email is sponsored by:

Show off your parallel programming skills.
Enter the Intel(R) Threading Challenge 2010.
http://p.sf.net/sfu/intel-thread-sfd
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Verify differences: SHA1 sum doesn't match but it should

2010-08-30 Thread Henrik Johansen
'Paul Mather' wrote:
On Aug 30, 2010, at 6:41 AM, Henrik Johansen wrote:

 Like most ZFS related stuff it all sounds (and looks) extremely easy but
 in reality it is not quite so simple.

Yes, but does ZFS makes things easier or harder?

It hides a lot of the complexity involved. In ZFS it is either
super-easy or super-hard (once super-easy fails to apply).

Silent data corruption won't go away just because your pool is large. :-)

No, but a large pool combined with a high changerate renders scrubbing
somewhat useless.

(But, this is all getting a bit off-topic for Bacula-users.)

Agreed.

Cheers,

Paul.


-- 
Med venlig hilsen / Best Regards

Henrik Johansen
hen...@scannet.dk
Tlf. 75 53 35 00

ScanNet Group
A/S ScanNet 

--
This SF.net Dev2Dev email is sponsored by:

Show off your parallel programming skills.
Enter the Intel(R) Threading Challenge 2010.
http://p.sf.net/sfu/intel-thread-sfd
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Verify differences: SHA1 sum doesn't match but it should

2010-08-30 Thread Bruno Friedmann
On 08/30/2010 06:35 PM, Tobias Brink wrote:
 Steve Costaras stev...@chaven.com writes:
 
  Could be due to a transient error (transmission or wild/torn read at
 time of calculation).   I see this a lot with integrity checking of
 files here (50TiB of storage).

 Only way to get around this now is to do a known-good sha1/md5 hash of
 data (2-3 reads of the file make sure that they all match and that the
 file is not corrupted) save that as a baseline and then when doing
 reads/compares if one fails do another re-read and see if the first
 one was in error and compare that with your baseline. This is one
 reason why I'm switching to the new generation of sas drives that have
 ioecc checks on READS not just writes to help cut down on some of
 this.

 Corruption does occur as well and is more probable with the higher the
 capacity of the drive. Ideally you would have a drive that would
 do ioecc on reads, plus using T10 PI extensions (DIX/DIF) from drive
 to controller up to your file system layer.It won't always prevent
 it by itself but would allow if you have a raid setup to do some
 self-healing when a drive reports a non transient (i.e. corrupted
 sector of data).
 
 First off, thanks for the answers.  The thing is that I am well aware of
 the reliability problems of hard drives and I would love to use some
 advanced file system like ZFS or btrfs, but I am on Debian and I will
 stay on Debian.  And btrfs is not mature enough to be used in production
 at the moment.  The other thing is that I do not think that this is an
 issue of corruption of the data itself!  As I said I checked the files
 against backups and MD5 sums supplied by Debian (several times and from
 cold cache) and the data seems to be OK.  The executables that are
 reported by Bacula to have changed continue to work well and bug-free
 just as before.
 
 So I think this is a problem/bug with either the Postgresql database or
 Bacula, not with my hard drives.  I just wonder how something like this
 could happen and how I could avoid this.  I'm also not willing to do
 additional checksums with other programs (AIDE or similar) because they
 take _lots_ of time to run.  With Bacula I get the checksums for free.
 I just want to use them to detect corruption on disk from time to time
 and because I use VirtualFull and want to know if my differential
 backups have missed something.
 
 So I still don't know how to proceed.  Apart from that I will try to
 upgrade my director and sd to 5.0.2 as soon as Debian backports are
 available and see if the problem goes away.  I will also re-run the
 DiskToCatalog after my next differential backup and see if something
 is different.
 
 Thanks,
 Tobias
 

Tobias, I use this little python script to extract information which I used to 
track duplicates files
(users are users :-)

Hope this could help you a bit to have inspiration and decode the lstat column.
(If I remember, someone has also do the same in pl/pgsql: check the archives 
list)

#!/usr/bin/python
# -*- coding: utf-8 -*-
#
#   call it with a jobid and pipe it to csv file
#
import sys
import time
import MySQLdb

jobid = sys.argv[1]

def base64_decode_lstat(record, position):
b64 = ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/
val = 0
size = record.split(' ')[position]
for i in range(len(size)):
val += (b64.find(size[i])) * (pow(64,(len(size)-i)-1))
return val

# Adjust localhost,username,passwd,dbname

db = MySQLdb.connect(host=localhost, user=bacula, passwd=bacula, 
db=bacula)
db.set_character_set('utf8')
cursor = db.cursor()
cursor.execute('SET NAMES utf8;')
cursor.execute('SET CHARACTER SET utf8;')
cursor.execute('SET character_set_connection=utf8;')
cursor.execute(SELECT File.MD5 as cheksum, convert(Filename.Name using utf8) 
as filename, convert(Path.Path using utf8) as
path, File.LStat as lstat\
FROM File, Filename, Path \
WHERE File.JobId = '%s' \
AND Filename.FilenameId = File.FilenameId \
AND Path.PathId = File.PathId\
ORDER BY 
File.MD5,Filename.Name,Path.Path % jobid)
result = cursor.fetchall()

# no headerprint 
'checksum;filename;path;lstat;gid;uid;bytes;blocksize;blocks_allocated;atime;mtime;ctime'

for record in result:

print '%s;%s;%s;%s;%s;%s;%s;%s;%s;%s;%s;%s' % (
record[0] , record[1] , record[2] , record[3] , \
base64_decode_lstat(record[3],5) , \
base64_decode_lstat(record[3],6) , \
base64_decode_lstat(record[3],7) , \
base64_decode_lstat(record[3],8) , \
base64_decode_lstat(record[3],9) , \
base64_decode_lstat(record[3],10) , \
base64_decode_lstat(record[3],11) , \
base64_decode_lstat(record[3],12) \
)

# no empty line at end print





-- 

Bruno Friedmann  br...@ioda-net.ch

Ioda-Net Sàrl www.ioda-net.ch

  openSUSE Member
User www.ioda.net/r/osu
Blog 

Re: [Bacula-users] Bacula VMs

2010-08-30 Thread Josh Fisher
  I assume you mean SD and Dir in a VM. I have been using Bacula in a 
qemu-kvm guest for some time now. Dir and database run beautifully in a 
VM. SD is a bit trickier. SD works well with disk storage. All works as 
expected even with USB disk, but only at USB 1 speeds. That is only due 
to qemu-kvm not having EHCI (USB 2.0) device emulation, however I hear 
that is forthcoming. The problem lies in using SCSI or SAS tape drives. 
Passing particular SCSI or SATA/SAS devices through to the VM is 
accomplished by passing through the PCI / PCIe controller hardware, 
rather than the drive itself. Setting up qemu-kvm to attach anything 
other than disk storage devices to the VM is more complex, and I'm not 
sure it would make it any easier. I've never tried with VMWare or 
VirtualBox, though.

On 8/30/2010 10:50 AM, Dan Langille wrote:
 This just came up on IRC.  As an aid to newcomers, a ready-made VM may
 be useful. A potential user could download and run the VM, without
 having to go through the install process.

 If you use VMs (eg VMWare, VirtualBox), then I think this would be a
 great way for you to contribute to the project. And help to lower the
 barrier to entry. Sometimes, the hardest step can be installation and
 just getting it running.

 :)


--
This SF.net Dev2Dev email is sponsored by:

Show off your parallel programming skills.
Enter the Intel(R) Threading Challenge 2010.
http://p.sf.net/sfu/intel-thread-sfd
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


[Bacula-users] Duplicate job not allowed during copy job.

2010-08-30 Thread Brian Debelius
  Bacula 5.0.3 Ubuntu 10.04 64bit MySQL

Since upgrading to 5.0.3 from 3.0.x I get the following duplicate job 
error when running from a schedule.  If I run the copy job manually, it 
runs without error.  The following was typed out by hand, there may be 
some typos.

Brian-


25-Aug 09:02 marlin2-dir JobId 10782:  The following six JobIds were 
chosen to be copied:  10775,10776,10777,10778,10779,10781
25-Aug 09:02 marlin2-dir JobId 10782:  Job queued. JobId=10783
25-Aug 09:02 marlin2-dir JobId 10782:  Copying JobId 10783 started.
25-Aug 09:02 marlin2-dir JobId 10782:  Job queued. JobId=10785
25-Aug 09:02 marlin2-dir JobId 10782:  Copying JobId 10785 started.
25-Aug 09:02 marlin2-dir JobId 10782:  Job queued. JobId=10785
25-Aug 09:02 marlin2-dir JobId 10782:  Copying JobId 10785 started.
25-Aug 09:02 marlin2-dir JobId 10782:  Job queued. JobId=10787
25-Aug 09:02 marlin2-dir JobId 10782:  Copying JobId 10787 started.
25-Aug 09:02 marlin2-dir JobId 10782:  Job queued. JobId=10789
25-Aug 09:02 marlin2-dir JobId 10782:  Copying JobId 10789 started.
25-Aug 09:02 marlin2-dir JobId 10782:  Job queued. JobId=10791
25-Aug 09:02 marlin2-dir JobId 10782:  Copying JobId 10791 started.


25-Aug 09:02 marlin2-dir JobId 10783:  Fatal Error: JobId 10782 already 
running.  Duplicate job not allowed.
25-Aug 09:02 marlin2-dir JobId 10783:  Fatal Error: JobId 10785 already 
running.  Duplicate job not allowed.
25-Aug 09:02 marlin2-dir JobId 10783:  Fatal Error: JobId 10791 already 
running.  Duplicate job not allowed.
25-Aug 09:02 marlin2-dir JobId 10783:  Fatal Error: JobId 10787 already 
running.  Duplicate job not allowed.
25-Aug 09:02 marlin2-dir JobId 10783:  Fatal Error: JobId 10789 already 
running.  Duplicate job not allowed.

## Disk Full Pool
Pool {
   Name = Disk-Full
   Pool Type= Backup
   Storage  = Disk
   Volume Retention = 4 Weeks
   Auto Prune   = yes
   Recycle  = yes
   Label Format = vol-
   Maximum Jobs = 1
   Maximum Volume Bytes = 4G
   Scratch Pool = Disk-Scratch
   Recycle Pool = Disk-Scratch
   Volume Use Duration = 1 days
   Action On Purge  = Truncate
   Next Pool= LTO3-Copy_Full
}

## LTO-3 Pool
Pool {
   Name   = LTO3-Copy_Full
   Pool Type  = Backup
   Storage= LTO-3
   Volume Retention = 16 Weeks
   Auto Prune = yes
   Recycle= yes
   Volume Use Duration = 8 days
}


## Copy Job to LTO
Job {
Name  = LTO3-Copy_Full
Type  = Copy
Schedule  = LTO3-Copy_Full
Client= None
File Set  = None
Pool  = Disk-Full
Messages  = Standard
Maximum Concurrent Jobs   = 1
Priority  = 98
Write Bootstrap   = /opt/bacula/bsr/%1-%j.bsr
Allow Duplicate Jobs  = yes
Allow Higher Duplicates   = yes
Cancel Queued Duplicate   = no
Cancel Running Duplicates = no
Spool Attributes  = yes
Selection Type= SQLQuery
Selection Pattern = SELECT DISTINCT Job.Job.Id FROM Job,Pool WHERE 
Pool.Name = 'Disk-Full' AND Pool.PoolId = Job.PoolId AND Job.Type = 'B' 
AND Job.JobStatus IN ('T', 'W') AND Job.jobBytes  0 AND Job.JobId NOT 
IN (SELECT PriorJobId FROM Job WHERE Type IN ('B', 'C') AND 
Job.JobStatus IN ('T', 'W') AND PriorJobId !=0) ORDER by Job.StartTime;
}








--
This SF.net Dev2Dev email is sponsored by:

Show off your parallel programming skills.
Enter the Intel(R) Threading Challenge 2010.
http://p.sf.net/sfu/intel-thread-sfd
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


[Bacula-users] Bacula Developer's Conference

2010-08-30 Thread Kern Sibbald
Hello,

As you can see from the email below, we have been planning to hold a Bacula 
Developer's conference the 27th and 28th of September.  Despite some early 
enthusiasm, there are currently not enough people signed up to hold the 
conference.

However disappointing that is, I am hoping that there will be enough interest 
manifested by the end of this week.  Real interest can be shown by sending a 
registration email to the address noted below.

Best regards,

Kern

=== Email sent 9 August ==
Hello,

We will be holding the first Bacula Developer's conference during the last 
part of September and would enjoy meeting you and having you participate in 
the conference.  Please see below for more information.

For those of you who are planning to give a presentation, please contact me 
with the title and a brief abstract of your talk so that I can add it to the 
bacula.org web site page concerning the conference.

Please see the details below.

Best regards,

Kern


  Bacula Developer's Conference
  27-28 September 2010
Yverdon-les-Bains
   Switzerland

Are you interested in:

- Meeting the Bacula developers in person.
- Learning how we maintain the source code.
- Hearing presentations about Bacula from the developers.
- Learning about and helping define the Bacula roadmap.
- Giving feedback or ideas directly to the developers.
- Learning more about Bacula Systems.
- Hearing from other Bacula developers and users.

If so, you should know that there will be a Bacula Developer's conference in
Yverdon, Switzerland (about 1.5 hour by train from Geneva Airport) on the 27th
and 28th of September.  In addition, Bacula Systems will hold an Open House on
Sunday the 26th beginning at noon where you can meet most of the Bacula
Systems founders and key Bacula developers person.

This conference will be most appropriate for developers and advanced Bacula
users, but it is open and free for everyone interested in Bacula.  If you are
or have been a contributor to the Bacula project, the project may be able to
sponsor part of your trip -- please ask.


Sunday 26 Sept.  
  Bacula Systems Open House.  Visit the Bacula Systems offices and meet the
  Bacula developers and the Bacula Systems founders.  Totally informal and
  optional.


Day 1: Monday 27 September 2010
  Presentations by Bacula developers and anyone else who would like to give a
  formal presentation (30 to 45 minutes)

  Tentative program:
  - Swisscom sharing our experiences with Bacula (SAP backup, ...)
  - DassIT new Bacula conf file GUI editor
  - Bacula in Brazil
  - Linux Bare Metal Recovery
  - Bacula development process
  - Bacula Roadmap
  - Bweb
  - How Bacula Systems supports the project -- Rob Morrison

Day 2: Tuesday 28 September 2010
  Birds of a feather meetings:
  Informal direct conversations with the developers, planning, 
organizing, ...
  - Using git
  - How patches are integrated
  - Regression testing
- CDash regression dashboard
  - Release cycle
  - Roadmap discussion and your input
Brain storming new backup strategies such as deduplication
  - How Bacula plugins work
  - Rpms and how to improve them
...

Naturally, we will furnish plenty of beer and pizza and other goodies.


Your input needed:

We are still in the process of organizing the presentations and the time with
the developers, so if you are planning to attend and you would like to make a
formal presentation, or if you have some particular subjects that you would
like us to address, please let us know as soon as possible.

Registration:

To register for this event, please send an email to: 

   isabe...@baculasystems.com

and include your:

Name, Company, Telephone number, Mobile Telephone number, 
Date of arrival, Date of departure, and any special interests you have
for the conference.

In return, Isabelle will send you an email that gives detailed information
on available hotels, our location, how to get here, and other useful 
information for the conference.

--
This SF.net Dev2Dev email is sponsored by:

Show off your parallel programming skills.
Enter the Intel(R) Threading Challenge 2010.
http://p.sf.net/sfu/intel-thread-sfd
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Run backup job from command line in a non-interactive mode

2010-08-30 Thread Sergio Belkin
2010/8/30 Sebastian Bachmann m...@free-minds.net:
 -BEGIN PGP SIGNED MESSAGE-
 Hash: SHA1

 mh you should add a yes to the end, not comment=yes

 hth

 On 30.08.10 21:32, Sergio Belkin wrote:
 2010/8/30  m...@free-minds.net:
 when=universal-time-specification

 Thanks, I did the following:

 [r...@pepino ~]# echo run job=Backup_centeno_Martes-LTO3
 client=centeno-fd fileset=dumpcenteno  storage=LTO3 when=2010-08-30
 21:00:00 comment=yes | bconsole
 Connecting to Director pepino:9101
 1000 OK: pepino-dir Version: 2.4.3 (10 October 2008)
 Enter a period to cancel a command.
 run job=Backup_centeno_Martes-LTO3 client=centeno-fd
 fileset=dumpcenteno  storage=LTO3 when=2010-08-30 21:00:00 comment=yes
 Using Catalog MyCatalog
 Invalid keyword: 21:00:00


 I've tried quoting date but the error is the same, what am I doing wrong?

 Thanks in advance!!


The error is still there :(



-- 
--
Sergio Belkin http://www.sergiobelkin.com
Watch More TV http://sebelk.blogspot.com
Sergio Belkin -

--
This SF.net Dev2Dev email is sponsored by:

Show off your parallel programming skills.
Enter the Intel(R) Threading Challenge 2010.
http://p.sf.net/sfu/intel-thread-sfd
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Run backup job from command line in a non-interactive mode

2010-08-30 Thread Sergio Belkin
2010/8/30 Sergio Belkin seb...@gmail.com:
 2010/8/30 Sebastian Bachmann m...@free-minds.net:
 -BEGIN PGP SIGNED MESSAGE-
 Hash: SHA1

 mh you should add a yes to the end, not comment=yes

 hth


It's getting better

echo run job=Backup_centeno_Martes-LTO3 client=spectrum-fd
fileset=dumpcenteno  storage=LTO3 when=2010-08-30_21:00:00 yes |
bconsole

It runs the job but tight now, I can set the hourly right.

TIA
-- 
--
Sergio Belkin http://www.sergiobelkin.com
Watch More TV http://sebelk.blogspot.com
Sergio Belkin -

--
This SF.net Dev2Dev email is sponsored by:

Show off your parallel programming skills.
Enter the Intel(R) Threading Challenge 2010.
http://p.sf.net/sfu/intel-thread-sfd
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


[Bacula-users] Job waiting on mount request blocking other jobs

2010-08-30 Thread Greg Golin
Hello,

Recently we went through a following scenario:

An external USB drive to which JobA was supposed to write became unusable. JobA 
requested a mount. Jobs B through Z that were scheduled to run after JobA did 
not run. As a result we lost two days of backups.

How can we make sure that jobs whose destination drives are accessible do not 
get blocked by jobs whose destination drives are inaccessible?

Thanks,
Greg Golin

smime.p7s
Description: S/MIME cryptographic signature
--
This SF.net Dev2Dev email is sponsored by:

Show off your parallel programming skills.
Enter the Intel(R) Threading Challenge 2010.
http://p.sf.net/sfu/intel-thread-sfd___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Bacula VMs

2010-08-30 Thread Dan Langille
On 8/30/2010 1:00 PM, Josh Fisher wrote:
 On 8/30/2010 10:50 AM, Dan Langille wrote:
 This just came up on IRC.  As an aid to newcomers, a ready-made VM may
 be useful. A potential user could download and run the VM, without
 having to go through the install process.

 If you use VMs (eg VMWare, VirtualBox), then I think this would be a
 great way for you to contribute to the project. And help to lower the
 barrier to entry. Sometimes, the hardest step can be installation and
 just getting it running.

 :)


   I assume you mean SD and Dir in a VM.

And an FD.  Everything a novice would need in order to play with Bacula.

  I have been using Bacula in a
 qemu-kvm guest for some time now. Dir and database run beautifully in a
 VM. SD is a bit trickier. SD works well with disk storage. All works as
 expected even with USB disk, but only at USB 1 speeds. That is only due
 to qemu-kvm not having EHCI (USB 2.0) device emulation, however I hear
 that is forthcoming. The problem lies in using SCSI or SAS tape drives.
 Passing particular SCSI or SATA/SAS devices through to the VM is
 accomplished by passing through the PCI / PCIe controller hardware,
 rather than the drive itself. Setting up qemu-kvm to attach anything
 other than disk storage devices to the VM is more complex, and I'm not
 sure it would make it any easier. I've never tried with VMWare or
 VirtualBox, though.

I would expect any VM example to be doing disk-based backups.

To be clear, I think we need snapshots of working  functional VMs which 
a novice can download, install, run, and then play with Bacula.

-- 
Dan Langille - http://langille.org/

--
This SF.net Dev2Dev email is sponsored by:

Show off your parallel programming skills.
Enter the Intel(R) Threading Challenge 2010.
http://p.sf.net/sfu/intel-thread-sfd
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users