Thanks much for your reply, Olaf.
Below for the answers.
On 11/23/2015 4:35 PM, Faaland, Olaf P. wrote:
Hello Bob,
We did something similar - our MDS's used zpools based on spinning
disks in JBODs and we switched to SSDs without bringing the filesystem
down, using ZFS to replicate data. It worked great for us.
How are your pools organized (ie what does zpool status show)? There
might be options that are more or less risky, or take more or less
time, depending on how zfs is using the disks.
This is a typical pool, which consists of disks on Dell MD1000 shelves,
each a single-disk raid-0 to make the JBOD group, then assembled via pci
address into the zpools of 10 disks each.
# zpool status ost-001
pool: ost-001
state: ONLINE
scan: none requested
config:
NAME STATE READ WRITE CKSUM
ost-001 ONLINE 0 0 0
raidz2-0 ONLINE 0 0 0
pci-0000:08:00.0-scsi-0:2:0:0 ONLINE 0 0 0
pci-0000:08:00.0-scsi-0:2:1:0 ONLINE 0 0 0
pci-0000:08:00.0-scsi-0:2:2:0 ONLINE 0 0 0
pci-0000:08:00.0-scsi-0:2:3:0 ONLINE 0 0 0
pci-0000:08:00.0-scsi-0:2:4:0 ONLINE 0 0 0
pci-0000:08:00.0-scsi-0:2:15:0 ONLINE 0 0 0
pci-0000:08:00.0-scsi-0:2:16:0 ONLINE 0 0 0
pci-0000:08:00.0-scsi-0:2:17:0 ONLINE 0 0 0
pci-0000:08:00.0-scsi-0:2:18:0 ONLINE 0 0 0
pci-0000:08:00.0-scsi-0:2:19:0 ONLINE 0 0 0
errors: No known data errors
Filesystem Size Used Avail Use% Mounted on
ost-001/ost0024 5.4T 3.4T 2.0T 64% /mnt/ost-001
Also, how often are disks failing and how long does a replacement take
to resilver, with your current disks?
These are old systems, and the underlying WD 750GB disks are going at an
average of 1 or 2 per week, with some 270 disks running this way. Some
shelves have newer larger disks, and those disks are giving us no
issues. We have a number of bigger, newer disk spares that we wanted to
swap in, giving us more of the 750GB units as spares.
Resilvering typically takes 6hrs or so these days.
bob
-Olaf
------------------------------------------------------------------------
*From:* Bob Ball [[email protected]]
*Sent:* Monday, November 23, 2015 12:22 PM
*To:* Faaland, Olaf P.; Morrone, Chris
*Cc:* Bob Ball
*Subject:* Expanding a zfsonlinux OST pool
Hi,
We have some zfsonlinux pools in use with Lustre 2.7 that use some
older disks, and we are rapidly running out of spares for those. What
we would _like_ to do, if possible, is replace all of those 750GB
disks in an OST, one at a time with re-silver between, with 1TB disks,
then expand the OST when the last is complete to utilize the larger
space and the more reliable disks.
Is this going to work? One of us here found the following:
According to the Oracle docs, a pool can autoexpand if you set it to
do so. I think the default must be off because the one I checked is
off (but that does indicate support of the feature in the linux
release also).
http://docs.oracle.com/cd/E19253-01/819-5461/githb/index.html
[root@umdist02 ~]# zpool get autoexpand ost-006
NAME PROPERTY VALUE SOURCE
ost-006 autoexpand off default
We are using zfsonlinux version 0.6.4.2. Can we follow the procedures outlined
in the oracle doc using zfsonlinux?
I guess my initial question assumed the expansion would not happen until the
last disk is added and re-silvered, but the document indicates this is not
really necessary?
Thanks,
bob
_______________________________________________
lustre-discuss mailing list
[email protected]
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org