Sorry for being a bit late in replying to this.

As you indicate, distributed processing ought to be more efficient, depending on the capabilities and loads of respective servers. I typically use client side compression, because all the servers I'm backing up have sufficient resources to do it themselves, and that would reduce the load on the server running Amanda. However, it is also the case that far and away the bulk of my data is on the servers that are running Amanda. They have the external arrays with tens of TBs of data. So, local compression in that case is actually on the Amanda server.

Having the bulk of the data and the Amanda backups on the same server has some advantages. It means that the path from the Raid array to the holding SSD to the LTO6 tape is all direct over SAS, and the SAS connections are all separate. That is, the SAS card that the arrays are attached to is different from the SAS card that the LTO6 is attached to, so they don't compete for bandwidth. The SSDs are internal, so they are on a separate connection. Then the trick is to have adequate CPU with enough cores to have a mess of processes running for Amanda without causing a slowdown of the other services that are running on that server.

I could imagine a case in which a dedicated Amanda server with sufficient cores might be doing all the compression for a number of other servers that were too busy with their own services to manage the compression of the backups. I just don't have that case.

The only issues I've run into with recovery were with a weird version of gnutar that had issues with Amanda. The solution was to make sure I was using a compatible gnutar and to test recoveries early in the installation/setup process so that I know the combination I have actually works.



On 11/1/16 6:17 PM, Ochressandro Rettinger wrote:

So, I understand the subject of software vs: hardware compression now (thanks for that very enlightening description of why it’s good to let Amanda handle the compression, Chris) but now I’m wondering about the question of server vs: client compression. I would tend to think that it would be best to use client compression, if possible, because then you can take advantage of the parallel nature of all of those separate processors working on their compression tasks. But from what several people have said, it seems like server side compression is a common choice.

Why would I want to choose one over the other? (Beyond the fact that in my particular case, client side compression isn’t actually working for recovery purposes.)

                Thanks for answering all these questions, guys.  :)

                -Sandro


--
---------------

Chris Hoogendyk

-
   O__  ---- Systems Administrator
  c/ /'_ --- Biology & Geosciences Departments
 (*) \(*) -- 315 Morrill Science Center
~~~~~~~~~~ - University of Massachusetts, Amherst

<[email protected]>

---------------

Erdös 4

Reply via email to