Le 01/12/2012 ? 08:33:31-0700, Jan Owoc a ?crit
Sorry, I'm very busy those past few days.
> >> >
> >> > http://tldp.org/HOWTO/LVM-HOWTO/removeadisk.html
> The commands described on that page do not have direct equivalents in
> zfs. There is currently no way to reduce the number of "top-level
> vdevs" in a pool or to change the RAID level.
> >> > I have a zpool with 48 disks with 4 raidz2 (12 disk). Inside those 48
> >> > disk
> >> > I've 36x 3T and 12 x 2T.
> >> > Can I buy new 12x4 To disk put in the server, add in the zpool, ask zpool
> >> > to migrate all data on those 12 old disk on the new and remove those old
> >> > disk ?
> In your specific example this means that you have 4 * RAIDZ2 of 12
> disks each. Zfs doesn't allow you to reduce that there are 4. Zfs
> doesn't allow you to change any of them from RAIDZ2 to any other
> configuration (eg RAIDZ). Zfs doesn't allow you to change the fact
> that you have 12 disks in a vdev.
> If you don't have a full set of new disks on a new system, or enough
> room on backup tapes to do a backup-restore, there are only two ways
> for to add capacity to the pool:
> 1) add a 5th top-level vdev (eg. another set of 12 disks)
That's not a problem.
> 2) replace the disks with larger ones one-by-one, waiting for a
> resilver in between
This is the point I don't see how to do it. I've 48 disk actually from
/dev/da0 -> /dev/da47 (I'm under FreeBSD 9.0) lets say 3To.
I've 4 raidz2 the first from /dev/da0 -> /dev/da11 etc..
So I add physically a new enclosure with new 12 disks for example 4To disk.
I'm going to have new /dev/da48 --> /dev/da59.
Say I want remove /dev/da0 -> /dev/da11. First I pull out the /dev/da0.
The first raidz2 going to be in «degraded state». So I going to tell the
pool the new disk is /dev/da48.
repeat this_process until /dev/da11 replace by /dev/da59.
But at the end how many space I'm going to use on those /dev/da48 -->
/dev/da51. Am I going to have 3To or 4To ? Because each time before
complete ZFS going to use only 3 To how at the end he going to magically
use 4To ?
Second question, when I'm going to pull out the first enclosure meaning the
old /dev/da0 --> /dev/da11 and reboot the server the kernel going to give
new number of those disk meaning
old /dev/da12 --> /dev/da0
old /dev/da13 --> /dev/da1
old /dev/da59 --> /dev/da47
how zfs going to manage that ?
> > When I would like to change the disk, I also would like change the disk
> > enclosure, I don't want to use the old one.
> You didn't give much detail about the enclosure (how it's connected,
> how many disk bays it has, how it's used etc.), but are you able to
> power off the system and transfer the all the disks at once?
Server : Dell PowerEdge 610
4 x enclosure : MD1200 with 12 disk of 3To
Connection : SAS
SAS Card : LSI
enclosures are chained :
server --> MD1200.1 --> MD1200.2 --> MD1200.3 --> MD1200.4
> > And what happen if I have 24, 36 disks to change ? It's take mounth to do
> > that.
> Those are the current limitations of zfs. Yes, with 12x2TB of data to
> copy it could take about a month.
> If you are feeling particularly risky and have backups elsewhere, you
> could swap two drives at once, but then you lose all your data if one
> of the remaining 10 drives in the vdev failed.
Thanks for the help
DIO bâtiment 15
Observatoire de Paris
5 Place Jules Janssen
92195 Meudon Cedex
Téléphone : 01 45 07 76 26/06 86 69 95 71
Heure local/Local time:
jeu 6 déc 2012 09:20:55 CET
zfs-discuss mailing list