Hi,
Breaking the server into multiple file daemons sounds as broken as
breaking the stuff amanda users had to do (break your filesystem into
something that fits a tape).
Saving multiple streams is something that has been proven as a
solution for many years, and where that is still too slow NDMP
source one if that was possible (it's not like I'm a large company
here at all).
-Original Message-
From: Florian Heigl [mailto:florian.he...@gmail.com]
Sent: Wednesday, July 6, 2011 09:20 AM
To: bacula-users@lists.sourceforge.net
Subject: Re: [Bacula-users] Performance options for single
Hello,
On 07/06/2011 04:20 PM, Florian Heigl wrote:
Saving multiple streams is something that has been proven as a
solution for many years, and where that is still too slow NDMP comes
into place. (in case of ZFS NDMP is still at a unusable stage)
100TB is a lot, but I wonder if everyone
]
Sent: Wednesday, July 6, 2011 11:20 AM
To: bacula-users@lists.sourceforge.net
Subject: Re: [Bacula-users] Performance options for single large (100TB) server
backup?
Hello,
On 07/06/2011 04:20 PM, Florian Heigl wrote:
Saving multiple streams is something that has been proven as a
solution for many
and How long can I stand a complete disaster recovery restore
taking).
From: Steve Costarasstev...@chaven.com
Subject: [Bacula-users] Performance options for single large (100TB)
server backup?
To:bacula-users@lists.sourceforge.net
Message-ID:W210986168202161309221804@webmail17
Content-Type
Am 28.06.2011 18:40, schrieb Steve Costaras:
How would the the various parts communicate if you're running multiple
instances on different ports? I would think just by creating multiple
jobs would create multiple socket streams and do the same thing.
I should have gotten another coffee
- File daemon is single threaded so is limiting backup performance. Is there
was a way to start more than one stream at the same time for a single machine
backup? Right now I have all the file systems for a single client in the same
file set.
- Tied in with above, accurate backups cut
On 6/27/2011 8:43 PM, Steve Costaras wrote:
- How to stream a single job to multiple tape drives. Couldn't
figure this out so that only one tape drive is being used.
There are hardware RAIT controllers available from Ultera
(http://www.ultera.com/tapesolutions.htm). A RAIT
How would the the various parts communicate if you're running multiple
instances on different ports? I would think just by creating multiple
jobs would create multiple socket streams and do the same thing.
On 2011-06-28 02:09, Christian Manal wrote:
- File daemon is single threaded
Problem is not really just tape I/O speeds but the ability to get data
to it. I.e. the SD is running at about 50% cpu overhead right now
(single core) so it could possible handle (2) LTO4 drives assuming a new
SD is not spawned off per drive?
I don't really need 'rait' itself as that
Hi Out of curiosity, why do you do such forklift replacements when ZFS
supports replacing individual drives, letting the pool resilver and then
automatically grow to the new size? roy - Original Message -
I have been using Bacula for over a year now and it has been providing
'passable'
disaster recovery restore
taking).
From: Steve Costaras stev...@chaven.com
Subject: [Bacula-users] Performance options for single large (100TB)
server backup?
To: bacula-users@lists.sourceforge.net
Message-ID: W210986168202161309221804@webmail17
Content-Type: text/plain; charset=utf
configurations and test which ones work best for your
design parameters (i.e. questions like How long can I go w/o a full
backup and How long can I stand a complete disaster recovery restore
taking).
From: Steve Costarasstev...@chaven.com
Subject: [Bacula-users] Performance options for single large
I have been using Bacula for over a year now and it has been providing
'passable' service though I think since day one I have been streching it to
it's limits or need a paradigm shift in how I am configuring it.
Basically, I have a single server which has direct atached disk (~128TB / 112
14 matches
Mail list logo