Re: [zfs-discuss] ZFS ARC DNLC Limitation

2007-09-25 Thread Roch - PAE

Hi Jason, This should have helped.

6542676 ARC needs to track meta-data memory overhead

Some of the lines to arc.c:

1551  1.36  if (arc_meta_used = arc_meta_limit) {
1552/*
1553 * We are exceeding our meta-data cache limit.
1554 * Purge some DNLC entries to release holds on 
meta-data.
1555 */
1556dnlc_reduce_cache((void 
*)(uintptr_t)arc_reduce_dnlc_percent);
1557}

-r



Jason J. W. Williams writes:
  Hello All,
  
  Awhile back (Feb '07) when we noticed ZFS was hogging all the memory
  on the system, y'all were kind enough to help us use the arc_max
  tunable to attempt to limit that usage to a hard value. Unfortunately,
  at the time a sticky problem was that the hard limit did not include
  DNLC entries generated by ZFS.
  
  I've been watching the list since then and trying to watch the Nevada
  commits. I haven't noticed that anything has been committed back so
  that arc_max truly enforces the max amount of memory ZFS is allowed to
  consume (including DNLC entries). Has this been corrected and I just
  missed it? Thank you in advance for you any help.
  
  Best Regards,
  Jason
  ___
  zfs-discuss mailing list
  zfs-discuss@opensolaris.org
  http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS speed degraded in S10U4 ?

2007-09-25 Thread Andy Lubel


On 9/25/07 3:37 AM, Sergiy Kolodka [EMAIL PROTECTED]
wrote:

 Hi Guys,
 
 I'm playing with Blade 6300 to check performance of compressed ZFS with Oracle
 database.
 After some really simple tests I noticed that default (well, not really
 default, some patches applied, but definitely noone bother to tweak disk
 subsystem or something else) installation of S10U3 is actually faster than
 S10U4, and a lot faster. Actually it's even faster on compressed ZFS with
 S10U3 than on uncompressed with S10U4.
 
 My configuration - default Update 3 LiveUpgraded to Update 4 with ZFS
 filesystem on dedicated disk, and I'm working with same files which are on
 same physical cylinders, so it's not likely a problem with HDD itself.
 
Did you do a 'zpool upgrade -a'?

 I'm doing as simple as just $time dd if=file.dbf of=/dev/null in few parallel
 tasks. On Update3 it's somewhere close to 11m32s and on Update 4 it's around
 12m6s. And it's both reading from compressed or uncompressed ZFS, numbers a
 little bit higher with compressed, couple of seconds more, which impressive by
 itself, but difference is the same, and strangest part is that reading file
 from compressed ZFS on U3 is faster than reading uncompressed with U4.
 
 I'm really surprised by this results, anyone else noticed that ?
  
I'm running a 'motley group of disks' on an e450 acting as our jumpstart
server and server build times are noticeably quicker since u4.

  
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

-Andy

-- 


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] enterprise scale redundant Solaris 10/ZFS server providing NFSv4/CIFS

2007-09-25 Thread James F. Hranicky
Paul B. Henson wrote:

 But all quotas were set in a single flat text file. Anytime you added a new
 quota, you needed to turn off quotas, then turn them back on, and quota
 enforcement was disabled while it recalculated space utilization.

I believe in later versions of the OS 'quota resize' did this without
the massive recalculation.

Jim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS array NVRAM cache?

2007-09-25 Thread Vincent Fox
Where is ZFS with regards to the NVRAM cache present on arrays?

I have a pile of 3310 with 512 megs cache, and even some 3510FC with 1-gig 
cache.  It seems silly that it's going to waste.  These are dual-controller 
units so I have no worry about loss of cache information.

It looks like OpenSolaris has a way to force arguably correct behavior, but 
Solaris 10u3/4 do not.  I see some threads from early this year about it, and 
nothing since.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS array NVRAM cache?

2007-09-25 Thread Jens Elkner
On Tue, Sep 25, 2007 at 10:14:57AM -0700, Vincent Fox wrote:
 Where is ZFS with regards to the NVRAM cache present on arrays?
 
 I have a pile of 3310 with 512 megs cache, and even some 3510FC with 1-gig 
 cache.  It seems silly that it's going to waste.  These are dual-controller 
 units so I have no worry about loss of cache information.
 
 It looks like OpenSolaris has a way to force arguably correct behavior, but 
 Solaris 10u3/4 do not.  I see some threads from early this year about it, and 
 nothing since.

Made some simple tests wrt. cont. seq. writes/reads for a 3510 (singl.
controller), single Host (v490) with 2 FC-HBAs - so, yes - I'm running
now ZFS single disk over HW Raid10 (10disks) ...
Haven't had the time, to test all combinations or mixed load cases, 
however, in case you wanna check, what I got:
http://iws.cs.uni-magdeburg.de/~elkner/3510.txt

Have fun,
jel.
-- 
Otto-von-Guericke University http://www.cs.uni-magdeburg.de/
Department of Computer Science   Geb. 29 R 027, Universitaetsplatz 2
39106 Magdeburg, Germany Tel: +49 391 67 12768
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS array NVRAM cache?

2007-09-25 Thread Bill Sommerfeld
On Tue, 2007-09-25 at 10:14 -0700, Vincent Fox wrote:
 Where is ZFS with regards to the NVRAM cache present on arrays?
 
 I have a pile of 3310 with 512 megs cache, and even some 3510FC with
 1-gig cache.  It seems silly that it's going to waste.  These are
 dual-controller units so I have no worry about loss of cache
 information.

I've done a few experiments with using small LUNs from a surplus 3510
raid unit for a separate intent log while putting the main body of the
pool in directly connected 3510 JBOD arrays.  Seems to work well; writes
to the intent log device show a much lower asvc_t (average service time)
value in iostat than writes to the main pool disks, and NFS performance
in a few completely unscientific and uncontrolled tests that are vaguely
representative of our workload seems to be better.

The behavior I'm seeing is consistent with the 3510 raid controller
ignoring the synchronize cache command.

I haven't put this into production just yet.

- Bill


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] enterprise scale redundant Solaris 10/ZFS server providing NFSv4/CIFS

2007-09-25 Thread Peter Tribble
On 9/24/07, Paul B. Henson [EMAIL PROTECTED] wrote:
 On Sat, 22 Sep 2007, Peter Tribble wrote:

  filesystem per user on the server, just to see how it would work. While
  managing 20,00 filesystems with the automounter was trivial, the attempt
  to manage 20,000 zfs filesystems wasn't entirely successful. In fact,
  based on that experience I simply wouldn't go down the road of one user
  per filesystem.

 Really? Could you provide further detail about what problems you
 experienced? Our current filesystem based on DFS effectively utilizes a
 separate filesystem per user (although in DFS terminology they are called
 filesets), and we've never had a problem managing them.

This was some time ago (a very long time ago, actually). There are two
fundamental problems:

1. Each zfs filesystem consumes kernel memory. Significant amounts, 64K
is what we worked out at the time. For normal numbers of filesystems that's
not a problem; multiply it by tens of thousands and you start to hit serious
resource usage.

2. The zfs utilities didn't scale well as the number of filesystems increased.

I just kept on issuing zfs create until the machine had had enough. It got
through the first 10,000 without too much difficulty (as I recall that
took several
hours), but soon got bogged down after that, to the point where it took a day
to do anything. At which point (at about 15000 filesystems on a 1G system)
it ran out of kernel memory and died. At this point it wouldn't even boot.

I know that some work has gone into improving the performance of the
utilities, and things like in-kernel sharetab (we never even tried to
share all those filesystems) are there to improve scalability. Perhaps
I should find a spare machine and try repeating the experiment.

-- 
-Peter Tribble
http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] device alias

2007-09-25 Thread Gregory Shaw
Hi.  I'd like to request a feature be added to zfs.  Currently, on  
SAN attached disk, zpool shows up with a big WWN for the disk.   If  
ZFS (or the zpool command, in particular) had a text field for  
arbitrary information, it would be possible to add something that  
would indicate what LUN on what array the disk in question might be.   
This would make troubleshooting and general understanding of the  
actual storage layout much simpler, as you'd know something about any  
disks that are encountering problems.

Something like:

zpool status
   pool: local
state: ONLINE
scrub: scrub completed with 0 errors on Sun Sep 23 04:16:33 2007
config:

 NAMESTATE READ WRITE CKSUM  NOTE
 local   ONLINE   0 0 0
   raidz1ONLINE   0 0 0
 c2t0d0  ONLINE   0 0 0  Internal SATA on  
left side
 c2t2d0  ONLINE   0 0 0  Internal SATA on  
right side
 c2t3d0  ONLINE   0 0 0  External SATA disk 1  
in box on top
 c2t4d0  ONLINE   0 0 0  External SATA disk 2  
in box on top
 spares
   c2t5d0AVAIL   External SATA disk 3  
in box on top

errors: No known data errors


The above would be very useful should a disk fail to identify what  
device is what.

Thanks!

-
Gregory Shaw, IT Architect
IT CTO Group, Sun Microsystems Inc.
Phone: (303)-272-8817 (x78817)
500 Eldorado Blvd, UBRM02-157 [EMAIL PROTECTED] (work)
Broomfield, CO 80021   [EMAIL PROTECTED] (home)
When Microsoft writes an application for Linux, I've won. - Linus  
Torvalds




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread James C. McPherson
Gregory Shaw wrote:
 Hi.  I'd like to request a feature be added to zfs.  Currently, on  
 SAN attached disk, zpool shows up with a big WWN for the disk.   If  
 ZFS (or the zpool command, in particular) had a text field for  
 arbitrary information, it would be possible to add something that  
 would indicate what LUN on what array the disk in question might be.   
 This would make troubleshooting and general understanding of the  
 actual storage layout much simpler, as you'd know something about any  
 disks that are encountering problems.
 
 Something like:
 
 zpool status
pool: local
 state: ONLINE
 scrub: scrub completed with 0 errors on Sun Sep 23 04:16:33 2007
 config:
 
  NAMESTATE READ WRITE CKSUM  NOTE
  local   ONLINE   0 0 0
raidz1ONLINE   0 0 0
  c2t0d0  ONLINE   0 0 0  Internal SATA on  
 left side
  c2t2d0  ONLINE   0 0 0  Internal SATA on  
 right side
  c2t3d0  ONLINE   0 0 0  External SATA disk 1  
 in box on top
  c2t4d0  ONLINE   0 0 0  External SATA disk 2  
 in box on top
  spares
c2t5d0AVAIL   External SATA disk 3  
 in box on top
 
 errors: No known data errors
 
 
 The above would be very useful should a disk fail to identify what  
 device is what.

How would you gather that information?
How would you ensure that it stayed accurate in
a hotplug world?



James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread Tim Spriggs
James C. McPherson wrote:
 Gregory Shaw wrote:
   
 Hi.  I'd like to request a feature be added to zfs.  Currently, on  
 SAN attached disk, zpool shows up with a big WWN for the disk.   If  
 ZFS (or the zpool command, in particular) had a text field for  
 arbitrary information, it would be possible to add something that  
 would indicate what LUN on what array the disk in question might be.   
 This would make troubleshooting and general understanding of the  
 actual storage layout much simpler, as you'd know something about any  
 disks that are encountering problems.

 Something like:

 zpool status
pool: local
 state: ONLINE
 scrub: scrub completed with 0 errors on Sun Sep 23 04:16:33 2007
 config:

  NAMESTATE READ WRITE CKSUM  NOTE
  local   ONLINE   0 0 0
raidz1ONLINE   0 0 0
  c2t0d0  ONLINE   0 0 0  Internal SATA on  
 left side
  c2t2d0  ONLINE   0 0 0  Internal SATA on  
 right side
  c2t3d0  ONLINE   0 0 0  External SATA disk 1  
 in box on top
  c2t4d0  ONLINE   0 0 0  External SATA disk 2  
 in box on top
  spares
c2t5d0AVAIL   External SATA disk 3  
 in box on top

 errors: No known data errors


 The above would be very useful should a disk fail to identify what  
 device is what.
 

 How would you gather that information?
 How would you ensure that it stayed accurate in
 a hotplug world?
   

If it is stored on the device itself it would keep the description with 
the same device.

In the case of iSCSI, it would be nice to keep lun info instead of 
having to correlate the drive id to the iqn to the lun, especially when 
working with luns in one place and drive ids in another.

-Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread James C. McPherson
Tim Spriggs wrote:
 James C. McPherson wrote:
 Gregory Shaw wrote:
...
 The above would be very useful should a disk fail to identify what  
 device is what.
 How would you gather that information?
 How would you ensure that it stayed accurate in
 a hotplug world?
 If it is stored on the device itself it would keep the description with 
 the same device.


What about when you replace a disk in an array? What
process would go and write the appropriate information?

How about a bootstrap for this? What would accomplish it?

What if you moved your entire system and re-cabled your
arrays slightly different? We have the beauty of devids
to make sure that the pools still work, but the physical
information Greg is talking about would almost certainly
be incorrect.


 In the case of iSCSI, it would be nice to keep lun info instead of 
 having to correlate the drive id to the iqn to the lun, especially when 
 working with luns in one place and drive ids in another.

When I was playing around with iscsi a few months back, I
used 1gb files on zfs which I then exported because
I could. What sort of lun info do you want in that sort of
use case? Position would be quite useless in some respects.



James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] enterprise scale redundant Solaris 10/ZFS server providing NFSv4/CIFS

2007-09-25 Thread Paul B. Henson
On Mon, 24 Sep 2007, Dale Ghent wrote:

 Not to sway you away from ZFS/NFS considerations, but I'd like to add
 that people who in the past used DFS typically went on to replace it with
 AFS. Have you considered it?

You're right, AFS is the first choice coming to mind when replacing DFS. We
actually implemented an OpenAFS prototype last year and have been running
it for internal use only since then.

Unfortunately, like almost everything we've looked at, AFS is a step
backwards from DFS. As the precursor to DFS, AFS has enough similarities to
DFS to make the features it lacks almost more painful.

No per file access control lists is a serious bummer. Integration with
Kerberos 5 rather than the internal kaserver is still at a bit of a duct
tape level, and only support DES. Having to maintain an additional
repository of user/group information (pts) is a bit of a pain, while there
are long-term goals to replace that with some type of LDAP integration I
don't see that anytime soon.

One of the most annoying things is that AFS requires integration at the
kernel level, yet is not maintained by the same people that maintain the
kernel. Frequently a Linux kernel upgrade will break AFS, and the
developers need to scramble to release a patch or update to resolve it.
While we are not currently using AFS under Solaris, based on mailing list
traffic similar issues arise. One of the benefits of NFSv4 is that it is a
core part of the operating system, unlikely to be lightly broken during
updates.


-- 
Paul B. Henson  |  (909) 979-6361  |  http://www.csupomona.edu/~henson/
Operating Systems and Network Analyst  |  [EMAIL PROTECTED]
California State Polytechnic University  |  Pomona CA 91768
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread Bill Sommerfeld
On Wed, 2007-09-26 at 08:26 +1000, James C. McPherson wrote:
 How would you gather that information?

the tools to use would be dependant on the actual storage device in use.
luxadm for A5x00 and V8x0 internal storage, sccli for 3xxx, etc., etc., 

 How would you ensure that it stayed accurate in
 a hotplug world?

See above.   

I'm told that with many jbod arrays, SES has the information.

- Bill


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] enterprise scale redundant Solaris 10/ZFS server providing NFSv4/CIFS

2007-09-25 Thread Paul B. Henson
On Tue, 25 Sep 2007, Peter Tribble wrote:

 This was some time ago (a very long time ago, actually). There are two
 fundamental problems:

 1. Each zfs filesystem consumes kernel memory. Significant amounts, 64K
 is what we worked out at the time. For normal numbers of filesystems that's
 not a problem; multiply it by tens of thousands and you start to hit serious
 resource usage.

Every server we've bought for about the last year came with 4 GB of memory,
the servers we would deploy for this would have at least 8 if not 16GB.
Given the downtrend in memory prices, hopefully memory would not be an
issue.


 2. The zfs utilities didn't scale well as the number of filesystems
 increased.
[...]
 share all those filesystems) are there to improve scalability. Perhaps
 I should find a spare machine and try repeating the experiment.

There have supposedly been lots of improvements in scalability, based on my
review of mailing list archives. If you do find the time to experiment
again, I'd appreciate hearing what you find...


-- 
Paul B. Henson  |  (909) 979-6361  |  http://www.csupomona.edu/~henson/
Operating Systems and Network Analyst  |  [EMAIL PROTECTED]
California State Polytechnic University  |  Pomona CA 91768
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] io:::start and zfs filenames?

2007-09-25 Thread Neelakanth Nadgir
io:::start probe does not seem to get zfs filenames in
args[2]-fi_pathname. Any ideas how to get this info?
-neel

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread James C. McPherson
Bill Sommerfeld wrote:
 On Wed, 2007-09-26 at 08:26 +1000, James C. McPherson wrote:
 How would you gather that information?
 
 the tools to use would be dependant on the actual storage device in use.
 luxadm for A5x00 and V8x0 internal storage, sccli for 3xxx, etc., etc., 

No consistent interface to use, then, unless another tool
or cruft gets added to ZFS to make this happen. That would
seem to defeat one of the major wins of ZFS - storage
neutrality.


 How would you ensure that it stayed accurate in
 a hotplug world?
 
 See above.   
 
 I'm told that with many jbod arrays, SES has the information.

True, but that's still many, but not all.


James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread Greg Shaw
It would be a manual process.  As with any arbitrary name, it's a useful 
tag, not much more.

James C. McPherson wrote:
 Gregory Shaw wrote:
   
 Hi.  I'd like to request a feature be added to zfs.  Currently, on  
 SAN attached disk, zpool shows up with a big WWN for the disk.   If  
 ZFS (or the zpool command, in particular) had a text field for  
 arbitrary information, it would be possible to add something that  
 would indicate what LUN on what array the disk in question might be.   
 This would make troubleshooting and general understanding of the  
 actual storage layout much simpler, as you'd know something about any  
 disks that are encountering problems.

 Something like:

 zpool status
pool: local
 state: ONLINE
 scrub: scrub completed with 0 errors on Sun Sep 23 04:16:33 2007
 config:

  NAMESTATE READ WRITE CKSUM  NOTE
  local   ONLINE   0 0 0
raidz1ONLINE   0 0 0
  c2t0d0  ONLINE   0 0 0  Internal SATA on  
 left side
  c2t2d0  ONLINE   0 0 0  Internal SATA on  
 right side
  c2t3d0  ONLINE   0 0 0  External SATA disk 1  
 in box on top
  c2t4d0  ONLINE   0 0 0  External SATA disk 2  
 in box on top
  spares
c2t5d0AVAIL   External SATA disk 3  
 in box on top

 errors: No known data errors


 The above would be very useful should a disk fail to identify what  
 device is what.
 

 How would you gather that information?
 How would you ensure that it stayed accurate in
 a hotplug world?



 James C. McPherson
 --
 Senior Kernel Software Engineer, Solaris
 Sun Microsystems
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
   
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread Greg Shaw


James C. McPherson wrote:
 Bill Sommerfeld wrote:
   
 On Wed, 2007-09-26 at 08:26 +1000, James C. McPherson wrote:
 
 How would you gather that information?
   
 the tools to use would be dependant on the actual storage device in use.
 luxadm for A5x00 and V8x0 internal storage, sccli for 3xxx, etc., etc., 
 

 No consistent interface to use, then, unless another tool
 or cruft gets added to ZFS to make this happen. That would
 seem to defeat one of the major wins of ZFS - storage
 neutrality.


   
I'd be happy with an arbitrary field that could be assigned via a 
command.  Intelligence could be added later if appropriate, but at this 
point, figuring out what-is-where on a big list of disk IDs on a SAN 
device is very difficult.

So, aiming low, a text field that could be assigned.   In the future, 
perhaps something that would associate a serial number of something 
similar to that name?
 How would you ensure that it stayed accurate in
 a hotplug world?
   
 See above.   

 I'm told that with many jbod arrays, SES has the information.
 

 True, but that's still many, but not all.


 James C. McPherson
 --
 Senior Kernel Software Engineer, Solaris
 Sun Microsystems
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
   
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread James C. McPherson
Greg Shaw wrote:
 James C. McPherson wrote:
 Bill Sommerfeld wrote:
  
 On Wed, 2007-09-26 at 08:26 +1000, James C. McPherson wrote:

 How would you gather that information?
   
 the tools to use would be dependant on the actual storage device in use.
 luxadm for A5x00 and V8x0 internal storage, sccli for 3xxx, etc., 
 etc., 

 No consistent interface to use, then, unless another tool
 or cruft gets added to ZFS to make this happen. That would
 seem to defeat one of the major wins of ZFS - storage
 neutrality.


   
 I'd be happy with an arbitrary field that could be assigned via a 
 command.  Intelligence could be added later if appropriate, but at this 
 point, figuring out what-is-where on a big list of disk IDs on a SAN 
 device is very difficult.
 
 So, aiming low, a text field that could be assigned.   In the future, 
 perhaps something that would associate a serial number of something 
 similar to that name?

That sounds like an ok RFE to me.

For some of the arrays (eg HDS) that we come
into contact with, it's possible to decode the
device guid into something meaningful to a
human, but that's generally closed information.


James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread Richard Elling
Dale Ghent wrote:
 On Sep 25, 2007, at 7:48 PM, Richard Elling wrote:
 
 The problem with this is that wrong information is much worse than no
 information, there is no way to automatically validate the  
 information,
 and therefore people are involved.  If people were reliable, then even
 a text file would work.  If it can't be automatic and reliable, then
 it isn't worth doing.
 
 I dunno if I think we have to think this far into it.
 
 Consider the Clearview project and their implementation of vanity  
 names for network interfaces. Conceivably, such a feature is useful  
 to admins as they can set the name to be a particular vlan number, or  
 the switch/blade/port where the other end of the ethernet line is  
 terminated. Or their ex-girlfriend's name if it's a particular  
 troublesome interface. The point is, it allows arbitrary naming of  
 something so that the admin(s) can associate with it better as an  
 object. Most importantly, there's a distinction here. Solaris  
 provides the facility. It's up to the admin to maintain it. That's as  
 far as it should go.

Actually, you can use the existing name space for this.  By default,
ZFS uses /dev/dsk.  But everything in /dev is a symlink.  So you could
setup your own space, say /dev/myknowndisks and use more descriptive
names.  You might need to hack on the startup service to not look at
/dev, but that shouldn't be too hard.  In other words, if the answer
is let the sysadmin do it then it can be considered solved.  The
stretch goal is to make some sort of reasonable name service.  At
this point I'll note the the FC folks envisioned something like that,
but never implemented it.

# ramdiskadm -a BrownDiskWithWhiteHat 150m /dev/ramdisk/BrownDiskWithWhiteDot
# zpool create zwimming /dev/ramdisk/BrownDiskWithWhiteDot
# zpool status zwimming
   pool: zwimming
  state: ONLINE
  scrub: none requested
config:

 NAME  STATE READ WRITE CKSUM
 zwimming  ONLINE   0 0 0
   /dev/ramdisk/BrownDiskWithWhiteDot  ONLINE   0 0 0

errors: No known data errors
# ls -l /dev/ramdisk/BrownDiskWithWhiteHat
lrwxrwxrwx   1 root root  55 Sep 25 17:59 
/dev/ramdisk/BrownDiskWithWhiteDot - ../../devices/pseudo/[EMAIL 
PROTECTED]:BrownDiskWithWhiteDot
# zpool export zwimming
# mkdir /dev/whee
# cd /dev/whee
# ln -s ../../devices/pseudo/[EMAIL PROTECTED]:BrownDiskWithWhiteHat 
YellowDiskUnderPinkBox
# zpool import -d /dev/whee zwimming
# zpool status zwimming
   pool: zwimming
  state: ONLINE
  scrub: none requested
config:

 NAMESTATE READ WRITE CKSUM
 zwimmingONLINE   0 0 0
   /dev/whee/YellowDiskUnderPinkBox  ONLINE   0 0 0

errors: No known data errors


  -- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread Gregory Shaw


On Sep 25, 2007, at 7:09 PM, Richard Elling wrote:


Dale Ghent wrote:

On Sep 25, 2007, at 7:48 PM, Richard Elling wrote:
The problem with this is that wrong information is much worse  
than no
information, there is no way to automatically validate the   
information,
and therefore people are involved.  If people were reliable, then  
even

a text file would work.  If it can't be automatic and reliable, then
it isn't worth doing.

I dunno if I think we have to think this far into it.
Consider the Clearview project and their implementation of vanity   
names for network interfaces. Conceivably, such a feature is  
useful  to admins as they can set the name to be a particular vlan  
number, or  the switch/blade/port where the other end of the  
ethernet line is  terminated. Or their ex-girlfriend's name if  
it's a particular  troublesome interface. The point is, it allows  
arbitrary naming of  something so that the admin(s) can associate  
with it better as an  object. Most importantly, there's a  
distinction here. Solaris  provides the facility. It's up to the  
admin to maintain it. That's as  far as it should go.


Actually, you can use the existing name space for this.  By default,
ZFS uses /dev/dsk.  But everything in /dev is a symlink.  So you could
setup your own space, say /dev/myknowndisks and use more descriptive
names.  You might need to hack on the startup service to not look at
/dev, but that shouldn't be too hard.  In other words, if the answer
is let the sysadmin do it then it can be considered solved.  The
stretch goal is to make some sort of reasonable name service.  At
this point I'll note the the FC folks envisioned something like that,
but never implemented it.

# ramdiskadm -a BrownDiskWithWhiteHat 150m /dev/ramdisk/ 
BrownDiskWithWhiteDot

# zpool create zwimming /dev/ramdisk/BrownDiskWithWhiteDot
# zpool status zwimming
  pool: zwimming
 state: ONLINE
 scrub: none requested
config:

NAME  STATE READ WRITE  
CKSUM
zwimming  ONLINE   0  
0 0
  /dev/ramdisk/BrownDiskWithWhiteDot  ONLINE   0  
0 0


errors: No known data errors
# ls -l /dev/ramdisk/BrownDiskWithWhiteHat
lrwxrwxrwx   1 root root  55 Sep 25 17:59 /dev/ramdisk/ 
BrownDiskWithWhiteDot - ../../devices/pseudo/ 
[EMAIL PROTECTED]:BrownDiskWithWhiteDot

# zpool export zwimming
# mkdir /dev/whee
# cd /dev/whee
# ln -s ../../devices/pseudo/[EMAIL PROTECTED]:BrownDiskWithWhiteHat  
YellowDiskUnderPinkBox

# zpool import -d /dev/whee zwimming
# zpool status zwimming
  pool: zwimming
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
zwimmingONLINE   0 0 0
  /dev/whee/YellowDiskUnderPinkBox  ONLINE   0 0 0

errors: No known data errors


 -- richard


But nobody would actually do this.  If the process can't be condensed  
into a single step (e.g. a single command), people won't bother.


Besides, who would take the chance that a 'boot -r' would keep their  
elaborate symbolic link tree intact?   I wouldn't.


I've learned that you can't count on anything in /dev remaining over  
'boot -r', patches, driver updates, san events, etc.


-
Gregory Shaw, IT Architect
IT CTO Group, Sun Microsystems Inc.
Phone: (303)-272-8817 (x78817)
500 Eldorado Blvd, UBRM02-157 [EMAIL PROTECTED] (work)
Broomfield, CO 80021   [EMAIL PROTECTED] (home)
When Microsoft writes an application for Linux, I've won. - Linus  
Torvalds





___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread Jason King
On 9/25/07, Gregory Shaw [EMAIL PROTECTED] wrote:



 On Sep 25, 2007, at 7:09 PM, Richard Elling wrote:

 Dale Ghent wrote:
 On Sep 25, 2007, at 7:48 PM, Richard Elling wrote:
 The problem with this is that wrong information is much worse than no
 information, there is no way to automatically validate the  information,
 and therefore people are involved.  If people were reliable, then even
 a text file would work.  If it can't be automatic and reliable, then
 it isn't worth doing.
 I dunno if I think we have to think this far into it.
 Consider the Clearview project and their implementation of vanity  names for
 network interfaces. Conceivably, such a feature is useful  to admins as they
 can set the name to be a particular vlan number, or  the switch/blade/port
 where the other end of the ethernet line is  terminated. Or their
 ex-girlfriend's name if it's a particular  troublesome interface. The point
 is, it allows arbitrary naming of  something so that the admin(s) can
 associate with it better as an  object. Most importantly, there's a
 distinction here. Solaris  provides the facility. It's up to the admin to
 maintain it. That's as  far as it should go.

 Actually, you can use the existing name space for this.  By default,
 ZFS uses /dev/dsk.  But everything in /dev is a symlink.  So you could
 setup your own space, say /dev/myknowndisks and use more descriptive
 names.  You might need to hack on the startup service to not look at
 /dev, but that shouldn't be too hard.  In other words, if the answer
 is let the sysadmin do it then it can be considered solved.  The
 stretch goal is to make some sort of reasonable name service.  At
 this point I'll note the the FC folks envisioned something like that,
 but never implemented it.

 # ramdiskadm -a BrownDiskWithWhiteHat 150m
 /dev/ramdisk/BrownDiskWithWhiteDot
 # zpool create zwimming /dev/ramdisk/BrownDiskWithWhiteDot
 # zpool status zwimming
   pool: zwimming
  state: ONLINE
  scrub: none requested
 config:

 NAME  STATE READ WRITE CKSUM
 zwimming  ONLINE   0 0 0
   /dev/ramdisk/BrownDiskWithWhiteDot  ONLINE
 0 0 0

 errors: No known data errors
 # ls -l /dev/ramdisk/BrownDiskWithWhiteHat
 lrwxrwxrwx   1 root root  55 Sep 25 17:59
 /dev/ramdisk/BrownDiskWithWhiteDot -
 ../../devices/pseudo/[EMAIL PROTECTED]:BrownDiskWithWhiteDot
 # zpool export zwimming
 # mkdir /dev/whee
 # cd /dev/whee
 # ln -s
 ../../devices/pseudo/[EMAIL PROTECTED]:BrownDiskWithWhiteHat
 YellowDiskUnderPinkBox
 # zpool import -d /dev/whee zwimming
 # zpool status zwimming
   pool: zwimming
  state: ONLINE
  scrub: none requested
 config:

 NAMESTATE READ WRITE CKSUM
 zwimmingONLINE   0 0 0
   /dev/whee/YellowDiskUnderPinkBox  ONLINE   0
0 0

 errors: No known data errors


  -- richard

 But nobody would actually do this.  If the process can't be condensed into a
 single step (e.g. a single command), people won't bother.

 Besides, who would take the chance that a 'boot -r' would keep their
 elaborate symbolic link tree intact?   I wouldn't.

 I've learned that you can't count on anything in /dev remaining over 'boot
 -r', patches, driver updates, san events, etc.


 -
 Gregory Shaw, IT Architect
 IT CTO Group, Sun Microsystems Inc.
 Phone: (303)-272-8817 (x78817)
 500 Eldorado Blvd, UBRM02-157 [EMAIL PROTECTED] (work)
 Broomfield, CO 80021   [EMAIL PROTECTED] (home)
 When Microsoft writes an application for Linux, I've won. - Linus Torvalds





 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



I've actually contemplated requesting such a feature for /dev
(creating vanity aliases that are presistant) as it would also be
useful for things like databases that use raw disk (i.e. Oracle ASM).
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS array NVRAM cache?

2007-09-25 Thread Albert Chin
On Tue, Sep 25, 2007 at 06:01:00PM -0700, Vincent Fox wrote:
 I don't understand.  How do you
 
 setup one LUN that has all of the NVRAM on the array dedicated to it
 
 I'm pretty familiar with 3510 and 3310. Forgive me for being a bit
 thick here, but can you be more specific for the n00b?

If you're using CAM, disable NVRAM on all of your LUNs. Then, create
another LUN equivalent to the size of your NVRAM. Assign the ZIL to
this LUN. You'll then have an NVRAM-backed ZIL.

I posted a question along these lines to storage-discuss:
  http://mail.opensolaris.org/pipermail/storage-discuss/2007-July/003080.html

You'll need to determine the performance impact of removing NVRAM from
your data LUNs. Don't blindly do it.

-- 
albert chin ([EMAIL PROTECTED])
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] device alias

2007-09-25 Thread Mark Ashley
Please don't do this as a rule, it makes for horrendous support issues 
and breaks a lot of health check tools.


 Actually, you can use the existing name space for this.  By default,
 ZFS uses /dev/dsk.  But everything in /dev is a symlink.  So you could
 setup your own space, say /dev/myknowndisks and use more descriptive
 names.  You might need to hack on the startup service to not look at
 /dev, but that shouldn't be too hard.  In other words, if the answer
 is let the sysadmin do it then it can be considered solved.  The
 stretch goal is to make some sort of reasonable name service.  At
 this point I'll note the the FC folks envisioned something like that,
 but never implemented it.

 # ramdiskadm -a BrownDiskWithWhiteHat 150m
 /dev/ramdisk/BrownDiskWithWhiteDot
 # zpool create zwimming /dev/ramdisk/BrownDiskWithWhiteDot
 # zpool status zwimming
   pool: zwimming
  state: ONLINE
  scrub: none requested
 config:

 NAME  STATE READ WRITE CKSUM
 zwimming  ONLINE   0 0 0
   /dev/ramdisk/BrownDiskWithWhiteDot  ONLINE
 0 0 0

 errors: No known data errors
 # ls -l /dev/ramdisk/BrownDiskWithWhiteHat
 lrwxrwxrwx   1 root root  55 Sep 25 17:59
 /dev/ramdisk/BrownDiskWithWhiteDot -
 ../../devices/pseudo/[EMAIL PROTECTED]:BrownDiskWithWhiteDot
 # zpool export zwimming
 # mkdir /dev/whee
 # cd /dev/whee
 # ln -s
 ../../devices/pseudo/[EMAIL PROTECTED]:BrownDiskWithWhiteHat
 YellowDiskUnderPinkBox
 # zpool import -d /dev/whee zwimming
 # zpool status zwimming
   pool: zwimming
  state: ONLINE
  scrub: none requested
 config:

 NAMESTATE READ WRITE CKSUM
 zwimmingONLINE   0 0 0
   /dev/whee/YellowDiskUnderPinkBox  ONLINE   0
0 0

 errors: No known data errors


  -- richard

 But nobody would actually do this.  If the process can't be condensed into a
 single step (e.g. a single command), people won't bother.

 Besides, who would take the chance that a 'boot -r' would keep their
 elaborate symbolic link tree intact?   I wouldn't.

 I've learned that you can't count on anything in /dev remaining over 'boot
 -r', patches, driver updates, san events, etc.
 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] RAID-Z in ZFS

2007-09-25 Thread sontas
Anybody can tell me about RAID-Z architecture , cuz i tried to understand it by 
seach in google but it doesn't clear. I don't know why it can beat raid-5. I 
know it can prove about RAID-5 Write Hole cuz it has copy-on-write feature for 
data intigrity. But i don't understand about write full stripe for RAID-Z cus 
raid-5 when it write  , it must create parity for recover. Does anyone can 
clear it for me?

Thank in advance

Ko
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss