Re: [zfs-discuss] Building an On-Site and Off-Size ZFS server, replication question

2012-10-07 Thread Johannes Totz
On 05/10/2012 15:01, Edward Ned Harvey
(opensolarisisdeadlongliveopensolaris) wrote:
>> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- 
>> boun...@opensolaris.org] On Behalf Of Tiernan OToole
>> 
>> I am in the process of planning a system which will have 2 ZFS 
>> servers, one on site, one off site. The on site server will be
>> used by workstations and servers in house, and most of that will
>> stay in house. There will, however, be data i want backed up
>> somewhere else, which is where the offsite server comes in... This
>> server will be sitting in a Data Center and will have some storage 
>> available to it (the whole server currently has 2 3Tb drives, 
>> though they are not dedicated to the ZFS box, they are on VMware 
>> ESXi). There is then some storage (currently 100Gb, but more can
>> be requested) of SFTP enabled backup which i plan to use for some 
>> snapshots, but more on that later.
>> 
>> Anyway, i want to confirm my plan and make sure i am not missing 
>> anything here...
>> 
>> * build server in house with storage, pools, etc... * have a
>> server in data center with enough storage for its reason, plus the
>> extra for offsite backup * have one pool set as my "offsite"
>> pool... anything in here should be backed up off site also... *
>> possibly have another set as "very offsite" which will also be
>> pushed to the SFTP server, but not sure... * give these pools out
>> via SMB/NFS/iSCSI * every 6 or so hours take a snapshot of the 2 
>> offsite pools. * do a ZFS send to the data center box * nightly,
>> on the very offsite pool, do a ZFS send to the SFTP server * if 
>> anything goes wrong (my server dies, DC server dies, etc), Panic, 
>> download, pray... the usual... :)
>> 
>> Anyway, I want to make sure i am doing this correctly... Is there 
>> anything on that list that sounds stupid or am i doing anything 
>> wrong? am i missing anything?
>> 
>> Also, as a follow up question, but slightly unrelated, when it 
>> comes to the ZFS Send, i could use SSH to do the send, directly to 
>> the machine... Or i could upload the compressed, and possibly 
>> encrypted dump to the server... Which, for resume-ability and 
>> speed, would be suggested? And if i where to go with an upload 
>> option, any suggestions on what i should use?
> 
> It is recommended, whenever possible, you should pipe the "zfs send" 
> directly into a "zfs receive" on the receiving system.  For two
> solid reasons:
> 
> If a single bit is corrupted, the whole stream checksum is wrong and 
> therefore the whole stream is rejected.  So if this occurs, you want 
> to detect it (in the form of one incremental failed) and then
> correct it (in the form of the next incremental succeeding).
> Whereas, if you store your streams on storage, it will go undetected,
> and everything after that point will be broken.
> 
> If you need to do a restore, from a stream stored on storage, then 
> your only choice is to restore the whole stream.  You cannot look 
> inside and just get one file.  But if you had been doing send | 
> receive, then you obviously can look inside the receiving filesystem 
> and extract some individual specifics.
> 
> If the recipient system doesn't support "zfs receive," [...]

On that note, is there a minimal user-mode zfs thing that would allow
receiving a stream into an image file? No need for file/directory access
etc.
I was thinking maybe the zfs-fuse-on-linux project may have suitable bits?


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Migrating 512 byte block zfs root pool to 4k disks

2012-06-15 Thread Johannes Totz
On 15/06/2012 13:22, Sašo Kiselkov wrote:
> On 06/15/2012 02:14 PM, Hans J Albertsson wrote:
>> I've got my root pool on a mirror on 2 512 byte blocksize disks. I
>> want to move the root pool to two 2 TB disks with 4k blocks. The
>> server only has room for two disks. I do have an esata connector,
>> though, and a suitable external cabinet for connecting one extra disk.
>> 
>> How would I go about migrating/expanding the root pool to the
>> larger disks so I can then use the larger disks for booting?
>> I have no extra machine to use.
> 
> Suppose we call the disks like so:
> 
>   A, B: your old 512-block drives
>   X, Y: your new 2TB drives
> 
> The easiest way would be to simply:
> 
> 1) zpool set autoexpand=on rpool
> 2) offline the A drive
> 3) physically replace it with the X drive
> 4) do a "zpool replace" on it and wait for it to resilver

When sector size differs, attaching it is going to fail (at least on fbsd).
You might not get around a send-receive cycle...

> 5) offline the B drive
> 6) physically replace it with the Y drive
> 7) do a "zpool replace" on it and wait for it to resilver
> 
> At this point, you should have a 2TB rpool (thanks to the
> "autoexpand=on" in step 1). Unfortunately, to my knowledge, there is no
> way to convert a bshift=9 pool (512 byte sectors) to a bshift=13 pool
> (4k sectors). Perhaps some great ZFS guru can shed more light on this.
> 
> --
> Saso
> 



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss