Chris Cosby wrote: > About the best I can see: > > zpool create dirtypool raidz 250a 250b 320a raidz 320b 400a 400b raidz > 500a 500b 750a > > And you have to do them in that order. The zpool will create using the > smallest device. This gets you about 2140GB (500 + 640 + 1000) of > space. Your desired method is only 2880GB (720 * 4) and is WAY harder > to setup and maintain, especially if you get into the SDS configuration. > > I, for one, welcome our convoluted configuration overlords. I'd also > like to see what the zpool looks like if it works. This is, obviously, > untested. > I don't think I'd be that comfortable doing it, but I suppose you could just add each drive as a separate vDev, and set copies=2, but even that would get you about 1825GB (If my math is right the disks add up to 3650GB)
-Kyle > chris > > > On Fri, Aug 22, 2008 at 11:03 AM, Nils Goroll <[EMAIL PROTECTED] > <mailto:[EMAIL PROTECTED]>> wrote: > > Hi, > > John wrote: > > I'm setting up a ZFS fileserver using a bunch of spare drives. > I'd like some redundancy and to maximize disk usage, so my plan > was to use raid-z. The problem is that the drives are considerably > mismatched and I haven't found documentation (though I don't see > why it shouldn't be possible) to stripe smaller drives together to > match bigger ones. The drives are: 1x750, 2x500, 2x400, 2x320, > 2x250. Is it possible to accomplish the following with those drives: > > > > raid-z > > 750 > > 500+250=750 > > 500+250=750 > > 400+320=720 > > 400+720=720 > > > Though I've never used this in production, it seems possible to > layer ZFS on good old SDS (aka SVM, disksuite). > > At least I managed to create a trivial pool on > what-10-mins-ago-was-my-swap-slice: > > haggis:/var/tmp# metadb -f -a -c 3 /dev/dsk/c5t0d0s7 > haggis:/var/tmp# metainit d10 1 1 /dev/dsk/c5t0d0s1 > d10: Concat/Stripe is setup > haggis:/var/tmp# zpool create test /dev/md/dsk/d10 > haggis:/var/tmp# zpool status test > pool: test > state: ONLINE > scrub: none requested > config: > > NAME STATE READ WRITE CKSUM > test ONLINE 0 0 0 > /dev/md/dsk/d10 ONLINE 0 0 0 > > So it looks like you could do the follwing: > > * Put a small slice (10-20m should suffice, by convention it's > slice 7 on the first cylinders) on each of your disks and make > them the metadb, if you are not using SDS already > metadb -f -a -c 3 <all your slices_7> > > make slice 0 the remainder of each disk > > * for your 500/250G drives, create a concat (stripe not possible) > for each pair. for clarity, I'd recommend to include the 750G disk > as well (syntax from memory, apologies if I'm wrong with details): > > metainit d11 1 1 <700G disk>s0 > metainit d12 2 1 <500G disk>s0 1 <250G disk>s0 > metainit d13 2 1 <500G disk>s0 1 <250G disk>s0 > metainit d14 2 1 <400G disk>s0 1 <320G disk>s0 > metainit d15 2 1 <400G disk>s0 1 <320G disk>s0 > > * create a raidz pool on your metadevices > > zpool create <name> raidz /dev/md/dsk/d11 /dev/md/dsk/d12 > /dev/md/dsk/d13 /dev/md/dsk/d14 /dev/md/dsk/d15 > > Again: I have never tried this, so please don't blame me if this > doesn't work. > > Nils > > > This message posted from opensolaris.org <http://opensolaris.org> > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org <mailto:zfs-discuss@opensolaris.org> > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > > > > -- > chris -at- microcozm -dot- net > === Si Hoc Legere Scis Nimium Eruditionis Habes > ------------------------------------------------------------------------ > > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss