Re: [zfs-discuss] raidz recovery

2010-12-18 Thread Tuomas Leikola
On Wed, Dec 15, 2010 at 3:29 PM, Gareth de Vaux z...@lordcow.org wrote:
 On Mon 2010-12-13 (16:41), Marion Hakanson wrote:
 After you clear the errors, do another scrub before trying anything
 else.  Once you get a complete scrub with no new errors (and no checksum
 errors), you should be confident that the damaged drive has been fully
 re-integrated into the pool.

 Ok I did a scrub after zero'ing, and the array came back clean, apparently, 
 but
 same final result - the array faults as soon as I 'offline' a different vdev.
 The zero'ing is just a pretend-the-errors-aren't-there directive, and the 
 scrub
 seems to be listening to that. What I need in this situation is a way to
 prompt ad6 to resilver from scratch.


I think scrub doesn't replace all superblocks or other stuff not in
the active dataset but rather some drive labels.

have you tried zpool replace? like remove ad6, fill with zeroes,
replace, command zpool replace tank ad6. That should simulate drive
failure and replacement with a new disk.

-- 
- Tuomas
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Mixing different disk sizes in a pool?

2010-12-18 Thread Ian D
Another question:  all those disks are on Dell MD1000 JBODs (11 of them) and we 
have 12 SAS ports on three LSI 9200-16e HBAs.  Is there any point connecting 
each JBOD on a separate port or is it ok cascading them in groups of three?  Is 
there a bandwidth limit we'll be hitting doing that?

Thanks
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] A few questions

2010-12-18 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Alexander Lesle
 
 at Dezember, 17 2010, 17:48 Lanky Doodle wrote in [1]:
 
  By single drive mirrors, I assume, in a 14 disk setup, you mean 7
  sets of 2 disk mirrors - I am thinking of traditional RAID1 here.
 
  Or do you mean 1 massive mirror with all 14 disks?
 
 Edward means a set of two-way-mirrors.

Correct.
mirror disk0 disk1 mirror disk2 disk3 mirror disk4 disk5 ...
You would normally call this a stripe of mirrors.  Even though the ZFS
concept of striping is more advanced than traditional raid striping...  We
still call this a ZFS stripe for lack of any other term.  A ZFS stripe has
all the good characteristics of raid concatenation and striping, without any
of the bad characteristics.  It can utilize bandwidth on multiple disks when
it wants to, or use a single device when it wants to for small blocks.  It
can dynamically add randomly sized devices, and it can be done
one-at-a-time.  Gaining everything good of traditional raid stripe or
concatenation, without any of the negatives of traditional raid stripe and
concatenation.


 At Sol11 Express Oracle announced that at TestInstall you can set
 RootPool to mirror during installation. At the moment I try it out
 in a VM but I didnt find this option. :-(

Actually, even in solaris 10, I habitually install the root filesystem onto
a ZFS mirror.  You just select 2 disks, and it's automatically a mirror.


 zpool create lankyserver mirror vdev1 vdev2 mirror vdev3 vdev4
 
 When you need more space you can expand a bundle of two disks to your
 lankyserver. Each pair with the same capacity is effective.
 
 zpool add lankyserver mirror vdev5 vdev6 mirror vdev7 vdev8  ...

Correct.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] A few questions

2010-12-18 Thread Edward Ned Harvey
 From: Bob Friesenhahn [mailto:bfrie...@simple.dallas.tx.us]
 Sent: Friday, December 17, 2010 9:16 PM
 
 While I agree that smaller vdevs are more reliable, your statement
 about the failure being more likely be in the same vdev if you have
 only 2 vdev's to be a rather useless statement.  The probability of
 vdev failure does not have anything to do with the number of vdevs.
 However, the probability of vdev failure increases tremendously if
 there is only one vdev and there is a second disk failure.

I'm not sure you got what I meant.  I'll rephrase and see if it's more
clear:

Correct, the number of vdev's doesn't affect the probability of a failure in
a specific vdev, but the number of disks in a vdev does.  Lanky said he was
considering 2x7disk raidz, versus 3x5disk raidz.  So when I said he's more
likely to have a 2nd disk fail in the same vdev if he only has 2 vdev's ...
That was meant to be taken in context, not as a generalization about pools
in general.

Consider a single disk.  Let P be the probability of the disk failing,
within 1 day.

If you have 5 disks in a raidz vdev, and one fails, there are 4 remaining.
If resilver will last 8 days, then the probability of a 2nd disk failing is
4*8*P = 32P

If you have 7 disks in a raidz vdev, and one fails, there are 6 remaining.
If a resilver will last 12 days, then the probability of a 2nd disk failing
is 6*12*P = 72P



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] copy complete zpool via zfs send/recv

2010-12-18 Thread Stephan Budach

Am 18.12.10 05:44, schrieb Edward Ned Harvey:

From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
boun...@opensolaris.org] On Behalf Of Stephan Budach

Now, I want to use zfs send -R t...@movetank | zfs recv targetTank/...
which would place all zfs fs one level down below targetTank.
Overwriting targetTank is not an option, since the zfs fs musn't exist

prior to

zfs recv.

You lost me in that one.

If there is a zfs filesystem at the recipient...  and you don't want to
receive into a subdirectory of it ... then you have to overwrite it.  And I
don't see any reason not to overwrite it.

You say the destination filesystem mustn't exist prior to receive...  But
did you know of the -F option?  It forces the overwrite, if you want to
receive and overwrite an existing filesystem.

What am I missing?  Or did I just answer it?


Ehh... well... you answered it... sort of. ;)
I think I simply didn't dare to overwrite the root zfs on the 
destination zpool with -F, but of course you're right, that this is the 
way to go.


Thanks,
budy
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] a single nfs file system shared out twice with different permissions

2010-12-18 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Geoff Nordli
 
 I am trying to configure a system where I have two different NFS shares
 which point to the same directory.  The idea is if you come in via one
path,
 you will have read-only access and can't delete any files, if you come in
 the 2nd path, then you will have read/write access.

I think you can do this client-side.

mkdir /foo1
mkdir /foo2
mount nfsserver:/exports/bar /foo1
mount -o ro nfsserver:/exports/bar /foo2

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] copy complete zpool via zfs send/recv

2010-12-18 Thread Edward Ned Harvey
 From: Stephan Budach [mailto:stephan.bud...@jvm.de]
 
 Ehh. well. you answered it. sort of. ;)
 I think I simply didn't dare to overwrite the root zfs on the destination
zpool
 with -F, but of course you're right, that this is the way to go.

What are you calling the root zfs on the destination?
You're not trying to overwrite / are you?  That would ... admittedly ... not
be so straightforward.  But I don't think it's impossible.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] copy complete zpool via zfs send/recv

2010-12-18 Thread Stephan Budach

Am 18.12.10 15:14, schrieb Edward Ned Harvey:

From: Stephan Budach [mailto:stephan.bud...@jvm.de]

Ehh. well. you answered it. sort of. ;)
I think I simply didn't dare to overwrite the root zfs on the destination

zpool

with -F, but of course you're right, that this is the way to go.

What are you calling the root zfs on the destination?
You're not trying to overwrite / are you?  That would ... admittedly ... not
be so straightforward.  But I don't think it's impossible.

The root zfs, to me, is the fs that gets created once you create the 
zpool. So, if I create the zpool tank, I also get the zfs fs tank, no?


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] copy complete zpool via zfs send/recv

2010-12-18 Thread eXeC001er
2010/12/18 Stephan Budach stephan.bud...@jvm.de

  Am 18.12.10 15:14, schrieb Edward Ned Harvey:

  From: Stephan Budach [mailto:stephan.bud...@jvm.de stephan.bud...@jvm.de]

 Ehh. well. you answered it. sort of. ;)
 I think I simply didn't dare to overwrite the root zfs on the destination

  zpool

  with -F, but of course you're right, that this is the way to go.

  What are you calling the root zfs on the destination?
 You're not trying to overwrite / are you?  That would ... admittedly ... not
 be so straightforward.  But I don't think it's impossible.


  The root zfs, to me, is the fs that gets created once you create the
 zpool. So, if I create the zpool tank, I also get the zfs fs tank, no?

 Yes, but zfs receive can put received data only to another pool. You cannot
zfs receive to RAW  disk


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] raidz recovery

2010-12-18 Thread Gareth de Vaux
On Sat 2010-12-18 (14:55), Tuomas Leikola wrote:
 have you tried zpool replace? like remove ad6, fill with zeroes,
 replace, command zpool replace tank ad6. That should simulate drive
 failure and replacement with a new disk.

'replace' requires a different disk to replace with.

How do you remove ad6?
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] A few questions

2010-12-18 Thread Lanky Doodle
On the subject of where to install ZFS, I was planning to use either Compact 
Flash or USB drive (both of which would be mounted internally); using up 2 of 
the drive bays for a mirrored install is possibly a waste of physical space, 
considering it's a) a home media server and b) the config can be backed up to a 
protected ZFS pool - if the CF or USB drive failed I would just replace and 
restore the config.

Can you have an equivalent of a global hot spare in ZFS. If I did go down the 
mirror route (mirror disk0 disk1 mirror disk2 disk3 mirror disk4 disk5 etc) all 
the way up to 14 disks that would leave the 15th disk spare.

Now this is getting really complex, but can you have server failover in ZFS, 
much like DFS-R in Windows - you point clients to a clustered ZFS namespace so 
if a complete server failed nothing is interrupted.

I am still undecided as to mirror vs RAID Z. I am going to be ripping 
uncompressed Blu-Rays so space is vital. I use RAID DP in NetApp kit at work 
and I'm guessing RAID Z2 is the equivalent? I have 5TB space at the moment so 
going to the expense of mirroring for only 2TB extra doesn't seem much of a pay 
off.

Maybe a compromise of 2x 7-disk RAID Z1 with global hotspare is the way to go?

Put it this way, I currently use Windows Home Server, which has no true disk 
failure protection, so any of ZFS's redundancy schemes is going to be a step 
up; is there an equivalent system in ZFS where if 1 disk fails you only lose 
that disks data, like unRAID?

Thanks everyone for your input so far :)
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Mixing different disk sizes in a pool?

2010-12-18 Thread Tim Cook
On Sat, Dec 18, 2010 at 7:26 AM, Ian D rewar...@hotmail.com wrote:

 Another question:  all those disks are on Dell MD1000 JBODs (11 of them)
 and we have 12 SAS ports on three LSI 9200-16e HBAs.  Is there any point
 connecting each JBOD on a separate port or is it ok cascading them in groups
 of three?  Is there a bandwidth limit we'll be hitting doing that?

 Thanks


It's fine to cascade them.  SAS is all point-to-point.  I strongly doubt
you'll hit a bandwidth constraint on the backend, especially if you have the
shelves multipathed, but if that's a concern you will get more peak
bandwidth putting them on separate ports.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] a single nfs file system shared out twice with different permissions

2010-12-18 Thread Geoff Nordli


-Original Message-
From: Edward Ned Harvey
[mailto:opensolarisisdeadlongliveopensola...@nedharvey.com]
Sent: Saturday, December 18, 2010 6:13 AM
To: 'Geoff Nordli'; zfs-discuss@opensolaris.org
Subject: RE: [zfs-discuss] a single nfs file system shared out twice with
different
permissions

 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Geoff Nordli

 I am trying to configure a system where I have two different NFS
 shares which point to the same directory.  The idea is if you come in
 via one
path,
 you will have read-only access and can't delete any files, if you come
 in the 2nd path, then you will have read/write access.

I think you can do this client-side.

mkdir /foo1
mkdir /foo2
mount nfsserver:/exports/bar /foo1
mount -o ro nfsserver:/exports/bar /foo2

Thanks Edward.

The client side solution works great. 

Happy holidays!!

Geoff 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Mixing different disk sizes in a pool?

2010-12-18 Thread Ian Collins

 On 12/18/10 10:49 AM, Ian D wrote:

I have 159x 15K RPM SAS drives I want to build a ZFS appliance with.

75x 145G
60x 300G
24x 600G

The box has 4 CPUs, 256G of RAM, 14x 100G SLC SSDs for the cache and a mirrored 
pair of 4G DDRDrive X1s for the SLOG.

My plan is to mirror all these drives and keep some hot spares.

My question is:  should I create three pools (one for each size of drives) and 
share the cache and slog among them or should I create a single pool with them 
all?

The answer really depends on what you want to do with pool(s).  You'll 
have to provide more information.


--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS ... open source moving forward?

2010-12-18 Thread Frank Cusack

On 12/16/10 10:24 AM -0500 Linder, Doug wrote:

Tim Cook wrote:


Claiming you'd start paying for Solaris if they gave you ZFS for free
in Linux is absolutely ridiculous.


*Start* paying?  You clearly have NO idea what it costs to run Solaris in
a production environment with support.


In my experience, it's less than RedHat.  Also TCO is less since Solaris
offers more to begin with.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS ... open source moving forward?

2010-12-18 Thread Frank Cusack

On 12/16/10 9:11 AM -0500 Linder, Doug wrote:

The only thing I'll add is that I, as I said, I really don't care at all
about licenses.


Then you have no room to complain or even suggest a specific license!


 When it comes to licenses, to me (and, I suspect, the
vast majority of other OSS users), GPL is synonymous with open
source.  Is that correct?  No.  Am I aware that plenty of other licenses
exist?  Yes.  Is the issue important?  Sure.


Agreed.


 Do I have time or interest
to worry about niggly little details?  No.


Well the problem with licenses is that they are decidedly NOT niggly
little details.  You should consider re-evaluating what you have time
or interest for, if you care about the things you say (such as maximum
and flexible use of the products you are using).


 All I want is to be able to
use the best technology in the ways that are most useful to me without
artificial restrictions.  Anything that advances that, I'm for.


CDDL is close to that, much closer than GPL.

-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS ... open source moving forward?

2010-12-18 Thread Frank Cusack

On 12/16/10 11:32 AM +0100 Joerg Schilling wrote:

 Note that while there existist
numerous papers  from lawyers that consistently explain which parts of
the GPLv2 are violating  US law and thus are void,


Can you elaborate?
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS ... open source moving forward?

2010-12-18 Thread Joerg Schilling
Frank Cusack frank+lists/z...@linetwo.net wrote:

 On 12/16/10 11:32 AM +0100 Joerg Schilling wrote:
   Note that while there existist
  numerous papers  from lawyers that consistently explain which parts of
  the GPLv2 are violating  US law and thus are void,

 Can you elaborate?

See: http://www.osscc.net/en/gpl.html for a list.

e.g. the papers from Lawrence Rosen, Tom Gordon  and Lothar Determan.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Mixing different disk sizes in a pool?

2010-12-18 Thread Ian D
 The answer really depends on what you want to do with
 pool(s).  You'll 
 have to provide more information.

Get the maximum of very random IOPS I get can out of those drives for database 
usage.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Mixing different disk sizes in a pool?

2010-12-18 Thread Tim Cook
On Sat, Dec 18, 2010 at 4:24 PM, Ian D rewar...@hotmail.com wrote:

  The answer really depends on what you want to do with
  pool(s).  You'll
  have to provide more information.

 Get the maximum of very random IOPS I get can out of those drives for
 database usage.
 --


Random IOPS won't max out the SAS link.  You'll be fine stacking them.  But
again, if you have the ports available, and already have the cables, it
won't hurt anything to use them.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] A few questions

2010-12-18 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Lanky Doodle
 
 On the subject of where to install ZFS, I was planning to use either
Compact
 Flash or USB drive (both of which would be mounted internally); using up 2
of
 the drive bays for a mirrored install is possibly a waste of physical
space,
 considering it's a) a home media server and b) the config can be backed up
to
 a protected ZFS pool - if the CF or USB drive failed I would just replace
and
 restore the config.

All of the above is correct.  One thing you should keep in mind however:  If
your unmirrored rpool (usb fob) fails...  Although yes you can restore
assuming you have been sufficiently backing it up ... You will suffer an
ungraceful halt.  Maybe you can live with that.


 Can you have an equivalent of a global hot spare in ZFS. If I did go down
the
 mirror route (mirror disk0 disk1 mirror disk2 disk3 mirror disk4 disk5
etc) all
 the way up to 14 disks that would leave the 15th disk spare.

Check the zpool man page for spare, but I know you can have spares
assigned to a vdev, and I'm pretty sure you can assign any given spare to
multiples, effectively making it a global hotspare.  So yes is the answer.


 Now this is getting really complex, but can you have server failover in
ZFS,
 much like DFS-R in Windows - you point clients to a clustered ZFS
namespace
 so if a complete server failed nothing is interrupted.

If that's somehow possible, it's something I don't know.  I don't believe
you can do that with ZFS.


 I am still undecided as to mirror vs RAID Z. I am going to be ripping
 uncompressed Blu-Rays so space is vital. 

For both read and write, raidz works extremely well for sequential
operations.  It sounds like you're probably going to be doing mostly
sequential operations, so raidz should perform very well for you.  A lot of
people will avoid raidzN because it doesn't perform very well for random
reads, so they opt for mirrors instead.  But in your case, no so much.

In your case, the only reason I can think to avoid raidz would be if you're
worrying about resilver times.  That's a valid concern, but you can linearly
choose any number of disks you want ... You could make raidz using 3-disks
each...  It's just a compromise between the mirror and the larger raidz
vdev.


 I use RAID DP in NetApp kit at work
 and I'm guessing RAID Z2 is the equivalent? 

Yup, raid-dp and raidz2 are conceptually pretty much the same.


 Put it this way, I currently use Windows Home Server, which has no true
disk
 failure protection, so any of ZFS's redundancy schemes is going to be a
step
 up; is there an equivalent system in ZFS where if 1 disk fails you only
lose that
 disks data, like unRAID?

No.  Not unless you make that many separate volumes.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS ... open source moving forward?

2010-12-18 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Frank Cusack
 
  Claiming you'd start paying for Solaris if they gave you ZFS for free
  in Linux is absolutely ridiculous.
 
  *Start* paying?  You clearly have NO idea what it costs to run Solaris
in
  a production environment with support.
 
 In my experience, it's less than RedHat.  Also TCO is less since Solaris
 offers more to begin with.

Guys...   The discussion of whether or not ZFS is open source moving forward
has long since been concluded.  Of course, please feel free to discuss
anything you like, but maybe you want to start a new thread to argue about
whether solaris is better than redhat, or GPL is legally significant and
blah blah blah, and so forth?

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss