Re: [zfs-discuss] OpenStorage GUI
-Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Chris Greer Sent: Wednesday, November 12, 2008 3:20 PM To: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] OpenStorage GUI Do you have any info on this upgrade path? I can't seem to find anything about this... I would also like to throw in my $0.02 worth that I would like to see the software offered to existing sun X4540 (or upgraded X4500) customers. Chris G. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- I too would like to see how this happens, checked with some Sun people and they didn't know of a way to upgrade a 4500 other than trading it in. Im assuming the Motherboard/CPU/Memory get swapped out, and from the chasis layout, looks fairly involved. We don't want to upgrade something that we just bought so we can take advantage of this software which appears to finally complete the Sun NAS picture with zfs! -Andy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] OpenStorage GUI
The word Module makes it sound really easy :) Has anyone ever swapped this module out, and if so - was it painful? Since our 4500's went from the pallet to the offsite datacenter I never did really get a chance to look closely at it. I found a picture of one and it looks like you could take out the whole guts in one tray (from the bottom rear?). -Andy -Original Message- From: Chris Greer [mailto:[EMAIL PROTECTED] Sent: Wednesday, November 12, 2008 3:57 PM To: Andy Lubel; zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] OpenStorage GUI I was hoping for a swap out of the system board module. Chris G. - Original Message - From: Andy Lubel [EMAIL PROTECTED] To: Chris Greer; zfs-discuss@opensolaris.org zfs-discuss@opensolaris.org Sent: Wed Nov 12 14:38:03 2008 Subject: RE: [zfs-discuss] OpenStorage GUI -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Chris Greer Sent: Wednesday, November 12, 2008 3:20 PM To: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] OpenStorage GUI Do you have any info on this upgrade path? I can't seem to find anything about this... I would also like to throw in my $0.02 worth that I would like to see the software offered to existing sun X4540 (or upgraded X4500) customers. Chris G. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- I too would like to see how this happens, checked with some Sun people and they didn't know of a way to upgrade a 4500 other than trading it in. Im assuming the Motherboard/CPU/Memory get swapped out, and from the chasis layout, looks fairly involved. We don't want to upgrade something that we just bought so we can take advantage of this software which appears to finally complete the Sun NAS picture with zfs! -Andy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] OpenStorage GUI
Afaik, the drives are pretty much the same, its the chipset that changed, which also meant a change of cpu and memory. -Andy From: Tim [mailto:[EMAIL PROTECTED] Sent: Wednesday, November 12, 2008 7:24 PM To: Andy Lubel Cc: Chris Greer; zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] OpenStorage GUI On Wed, Nov 12, 2008 at 2:38 PM, Andy Lubel [EMAIL PROTECTED] wrote: I too would like to see how this happens, checked with some Sun people and they didn't know of a way to upgrade a 4500 other than trading it in. Im assuming the Motherboard/CPU/Memory get swapped out, and from the chasis layout, looks fairly involved. We don't want to upgrade something that we just bought so we can take advantage of this software which appears to finally complete the Sun NAS picture with zfs! -Andy Couldn't you just swap out the hard drives? --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Sun Storage 7000
LOL, I guess Sun forgot that they had xvm! I wonder if you could use a converter (vmware converter) to make it work on vbox etc? I would also like to see this available as an upgrade to our 4500's.. Webconsole/zfs just stinks because it only paints a tiny fraction of the overall need for a web driven GUI. Anyone know if something like that is in the works? It looks like a nice appliance for file shares in a corp network. -Andy -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Tom Buskey Sent: Monday, November 10, 2008 3:40 PM To: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Sun Storage 7000 What, no VirtualBox image? This VMware image won't run on VMware Workstation 5.5 either :-( -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Ideal Setup: RAID-5, Areca, etc!
We have been using some 1068-1078 based cards (both raid:AOC-USAS-H4IR and jbod:LSISAS3801E) with b87-b90 and in s10u5 without issue for some time. Both the downloaded LSI driver and the bundled one have worked fine for us for around 6 months of moderate usage. The LSI jbod card is similar to the Sun SAS HBA card ;) -Andy -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of James C. McPherson Sent: Saturday, July 26, 2008 8:18 AM To: Miles Nordin Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Ideal Setup: RAID-5, Areca, etc! Miles Nordin wrote: bh == Brandon High [EMAIL PROTECTED] writes: bh a system built around the Marvell or LSI chipsets according to The Blogosphere, source of all reliable information, there's some issue with LSI, too. The driver is not available in stable Solaris nor OpenSolaris, or there are two drivers, or something. the guy is so upset, I can't figure out wtf he's trying to say: http://www.osnews.com/thread?317113 The driver for LSI's MegaRAID SAS card is mega_sas which was integrated into snv_88. It's planned for backporting to a Solaris 10 update. And I can't figure out what his beef is either. James C. McPherson -- Senior Kernel Software Engineer, Solaris Sun Microsystems http://blogs.sun.com/jmcp http://www.jmcp.homeunix.com/blog ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] SSD reliability, wear levelling, warranty period
On Jun 11, 2008, at 11:35 AM, Bob Friesenhahn wrote: On Wed, 11 Jun 2008, Al Hopper wrote: disk drives. But - based on personal observation - there is a lot of hype surrounding SSD reliability. Obviously the *promise* of this technology is higher performance and *reliability* with lower power requirements due to no (mechanical) moving parts. But... if you look broadly at the current SSD product offerings, you see: a) lower than expected performance - particularly in regard to write IOPS (I/O Ops per Second) and b) warranty periods that are typically 1 year - with the (currently rare) exception of products that are offered with a 5 year warranty. Other than the fact that SSDs eventually wear out from use, SSDs are no different from any other electronic device in that the number of individual parts, and the individual reliability of those parts, results in an overall reliability factor for the subsystem comprised of those parts. SSDs are jam-packed with parts. In fact, if you were to look inside an SSD and then look at how typical computers are implemented these days, you will see that one SSD has a whole lot more complex parts than the rest of the computer. SSDs will naturally become more reliable as their parts count is reduced due to higher integration and product maturity. Large SSD storage capacity requires more parts so large storage devices have less relability than smaller devices comprised of similar parts. SSDs are good for laptop reliability since hard drives tend to fail with high shock levels and laptops are often severely abused. Yeah I was going to add the fact that they dont spin at 7k+ rpm and have no 'moving' parts. I do agree that there is a lot of circuitry involved and eventually they will reduce that just like they did with mainboards. Remember how packed they used to be? Either way, I'm really interested in the vendor and technology Sun will choose for providing these SSD's in systems or as an add on card/ drive. -Andy Bob == Bob Friesenhahn [EMAIL PROTECTED], http://www.simplesystems.org/users/bfriesen/ GraphicsMagick Maintainer,http://www.GraphicsMagick.org/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs/nfs issue editing existing files
On Jun 6, 2008, at 11:22 AM, Andy Lubel wrote: That was it! hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R GETATTR3 OK hpux-is-old.com - nearline.host NFS C SETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R SETATTR3 Update synch mismatch hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R GETATTR3 OK hpux-is-old.com - nearline.host NFS C SETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R SETATTR3 Update synch mismatch hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R GETATTR3 OK hpux-is-old.com - nearline.host NFS C SETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R SETATTR3 Update synch mismatch hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 It is too bad our silly hardware only allows us to go to 11.23. That's OK though, in a couple months we will be dumping this server with new x4600's. Thanks for the help, -Andy On Jun 5, 2008, at 6:19 PM, Robert Thurlow wrote: Andy Lubel wrote: I've got a real doozie.. We recently implemented a b89 as zfs/ nfs/ cifs server. The NFS client is HP-UX (11.23). What's happening is when our dba edits a file on the nfs mount with vi, it will not save. I removed vi from the mix by doing 'touch /nfs/file1' then 'echo abc/nfs/file1' and it just sat there while the nfs servers cpu went up to 50% (one full core). Hi Andy, This sounds familiar: you may be hitting something I diagnosed last year. Run snoop and see if it loops like this: 10920 0.00013 141.240.193.235 - 141.240.193.27 NFS C GETATTR3 FH=6614 10921 0.7 141.240.193.27 - 141.240.193.235 NFS R GETATTR3 OK 10922 0.00017 141.240.193.235 - 141.240.193.27 NFS C SETATTR3 FH=6614 10923 0.7 141.240.193.27 - 141.240.193.235 NFS R SETATTR3 Update synch mismatch 10924 0.00017 141.240.193.235 - 141.240.193.27 NFS C GETATTR3 FH=6614 10925 0.00023 141.240.193.27 - 141.240.193.235 NFS R GETATTR3 OK 10926 0.00026 141.240.193.235 - 141.240.193.27 NFS C SETATTR3 FH=6614 10927 0.9 141.240.193.27 - 141.240.193.235 NFS R SETATTR3 Update synch mismatch If you see this, you've hit what we filed as Sun bugid 6538387, HP-UX automount NFS client hangs for ZFS filesystems. It's an HP-UX bug, fixed in HP-UX 11.31. The synopsis is that HP-UX gets bitten by the nanosecond resolution on ZFS. Part of the CREATE handshake is for the server to send the create time as a 'guard' against almost-simultaneous creates - the client has to send it back in the SETATTR to complete the file creation. HP-UX has only microsecond resolution in their VFS, and so the 'guard' value is not sent accurately and the server rejects it, lather rinse and repeat. The spec, RFC 1813, talks about this in section 3.3.2. You can use NFSv2 in the short term until you get that update. If you see something different, by all means send us a snoop. Update: We tried nfs v2 and the speed was terrible but the gettattr/setattr issue was gone. So what I'm looking at doing now is to create a raw volume, format it with ufs, mount it locallly, then share it over nfs. Luckily we will only have to do it this way for a few months, I don't like the extra layer and the block device isn't as fast as we hoped (I get about 400MB/s on the zfs filesystem and 180MB/s using the ufs-formatted local disk.. I just sure hope I'm not breaking any rules by implementing this workaround that will come back to haunt me later. -Andy Rob T ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs/nfs issue editing existing files
On Jun 9, 2008, at 12:28 PM, Andy Lubel wrote: On Jun 6, 2008, at 11:22 AM, Andy Lubel wrote: That was it! hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R GETATTR3 OK hpux-is-old.com - nearline.host NFS C SETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R SETATTR3 Update synch mismatch hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R GETATTR3 OK hpux-is-old.com - nearline.host NFS C SETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R SETATTR3 Update synch mismatch hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R GETATTR3 OK hpux-is-old.com - nearline.host NFS C SETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R SETATTR3 Update synch mismatch hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 It is too bad our silly hardware only allows us to go to 11.23. That's OK though, in a couple months we will be dumping this server with new x4600's. Thanks for the help, -Andy On Jun 5, 2008, at 6:19 PM, Robert Thurlow wrote: Andy Lubel wrote: I've got a real doozie.. We recently implemented a b89 as zfs/ nfs/ cifs server. The NFS client is HP-UX (11.23). What's happening is when our dba edits a file on the nfs mount with vi, it will not save. I removed vi from the mix by doing 'touch /nfs/file1' then 'echo abc/nfs/file1' and it just sat there while the nfs servers cpu went up to 50% (one full core). Hi Andy, This sounds familiar: you may be hitting something I diagnosed last year. Run snoop and see if it loops like this: 10920 0.00013 141.240.193.235 - 141.240.193.27 NFS C GETATTR3 FH=6614 10921 0.7 141.240.193.27 - 141.240.193.235 NFS R GETATTR3 OK 10922 0.00017 141.240.193.235 - 141.240.193.27 NFS C SETATTR3 FH=6614 10923 0.7 141.240.193.27 - 141.240.193.235 NFS R SETATTR3 Update synch mismatch 10924 0.00017 141.240.193.235 - 141.240.193.27 NFS C GETATTR3 FH=6614 10925 0.00023 141.240.193.27 - 141.240.193.235 NFS R GETATTR3 OK 10926 0.00026 141.240.193.235 - 141.240.193.27 NFS C SETATTR3 FH=6614 10927 0.9 141.240.193.27 - 141.240.193.235 NFS R SETATTR3 Update synch mismatch If you see this, you've hit what we filed as Sun bugid 6538387, HP-UX automount NFS client hangs for ZFS filesystems. It's an HP-UX bug, fixed in HP-UX 11.31. The synopsis is that HP-UX gets bitten by the nanosecond resolution on ZFS. Part of the CREATE handshake is for the server to send the create time as a 'guard' against almost-simultaneous creates - the client has to send it back in the SETATTR to complete the file creation. HP-UX has only microsecond resolution in their VFS, and so the 'guard' value is not sent accurately and the server rejects it, lather rinse and repeat. The spec, RFC 1813, talks about this in section 3.3.2. You can use NFSv2 in the short term until you get that update. If you see something different, by all means send us a snoop. Update: We tried nfs v2 and the speed was terrible but the gettattr/setattr issue was gone. So what I'm looking at doing now is to create a raw volume, format it with ufs, mount it locallly, then share it over nfs. Luckily we will only have to do it this way for a few months, I don't like the extra layer and the block device isn't as fast as we hoped (I get about 400MB/s on the zfs filesystem and 180MB/s using the ufs-formatted local disk.. I just sure hope I'm not breaking any rules by implementing this workaround that will come back to haunt me later. -Andy Tried this today and although things appear to function correctly, the performance seems to be steadily degrading. Am I getting burnt by double-caching? If so, what is the best way to workaround for my sad situation? I tried directio for the ufs volume and it made it even worse.. The only next thing I know to do is destroy one of my zfs pools and go back to SVM until we can get some newer nfs clients writing to this nearline. It pains me deeply!! TIA, -Andy Rob T ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs/nfs issue editing existing files
That was it! hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R GETATTR3 OK hpux-is-old.com - nearline.host NFS C SETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R SETATTR3 Update synch mismatch hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R GETATTR3 OK hpux-is-old.com - nearline.host NFS C SETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R SETATTR3 Update synch mismatch hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R GETATTR3 OK hpux-is-old.com - nearline.host NFS C SETATTR3 FH=F6B3 nearline.host - hpux-is-old.com NFS R SETATTR3 Update synch mismatch hpux-is-old.com - nearline.host NFS C GETATTR3 FH=F6B3 It is too bad our silly hardware only allows us to go to 11.23. That's OK though, in a couple months we will be dumping this server with new x4600's. Thanks for the help, -Andy On Jun 5, 2008, at 6:19 PM, Robert Thurlow wrote: Andy Lubel wrote: I've got a real doozie.. We recently implemented a b89 as zfs/ nfs/ cifs server. The NFS client is HP-UX (11.23). What's happening is when our dba edits a file on the nfs mount with vi, it will not save. I removed vi from the mix by doing 'touch /nfs/file1' then 'echo abc/nfs/file1' and it just sat there while the nfs servers cpu went up to 50% (one full core). Hi Andy, This sounds familiar: you may be hitting something I diagnosed last year. Run snoop and see if it loops like this: 10920 0.00013 141.240.193.235 - 141.240.193.27 NFS C GETATTR3 FH=6614 10921 0.7 141.240.193.27 - 141.240.193.235 NFS R GETATTR3 OK 10922 0.00017 141.240.193.235 - 141.240.193.27 NFS C SETATTR3 FH=6614 10923 0.7 141.240.193.27 - 141.240.193.235 NFS R SETATTR3 Update synch mismatch 10924 0.00017 141.240.193.235 - 141.240.193.27 NFS C GETATTR3 FH=6614 10925 0.00023 141.240.193.27 - 141.240.193.235 NFS R GETATTR3 OK 10926 0.00026 141.240.193.235 - 141.240.193.27 NFS C SETATTR3 FH=6614 10927 0.9 141.240.193.27 - 141.240.193.235 NFS R SETATTR3 Update synch mismatch If you see this, you've hit what we filed as Sun bugid 6538387, HP-UX automount NFS client hangs for ZFS filesystems. It's an HP-UX bug, fixed in HP-UX 11.31. The synopsis is that HP-UX gets bitten by the nanosecond resolution on ZFS. Part of the CREATE handshake is for the server to send the create time as a 'guard' against almost-simultaneous creates - the client has to send it back in the SETATTR to complete the file creation. HP-UX has only microsecond resolution in their VFS, and so the 'guard' value is not sent accurately and the server rejects it, lather rinse and repeat. The spec, RFC 1813, talks about this in section 3.3.2. You can use NFSv2 in the short term until you get that update. If you see something different, by all means send us a snoop. Rob T ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zfs/nfs issue editing existing files
Hello, I've got a real doozie.. We recently implemented a b89 as zfs/nfs/ cifs server. The NFS client is HP-UX (11.23). What's happening is when our dba edits a file on the nfs mount with vi, it will not save. I removed vi from the mix by doing 'touch /nfs/file1' then 'echo abc /nfs/file1' and it just sat there while the nfs servers cpu went up to 50% (one full core). This nfsstat is most troubling (I zeroed it and only tried to echo data into a file so this is the numbers for about 2 minutes before I CTRL-C'ed the echo command). Version 3: (11242416 calls) nullgetattr setattr lookup access readlink 0 0%5600958 49% 5600895 49% 19 0% 9 0%0 0% readwrite create mkdir symlink mknod 0 0%40494 0%5 0%0 0%0 0%0 0% remove rmdir rename linkreaddir readdirplus 3 0%0 0%0 0%0 0%0 0%7 0% fsstat fsinfo pathconfcommit 12 0% 0 0%0 0%14 0% Thats a lot of getattr and setattr! Does anyone have any advice on where I should start to figure out what is going on? truss, dtrace, snoop.. so many choices! Thanks, -Andy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] SMC Webconsole 3.1 and ZFS Administration 1.0 - stacktraces in snv_b89
On May 29, 2008, at 9:52 AM, Jim Klimov wrote: I've installed SXDE (snv_89) and found that the web console only listens on https://localhost:6789/ now, and the module for ZFS admin doesn't work. It works for out of the box without any special mojo. In order to get the webconsole to listen on something other than localhost did you do this? # svccfg -s svc:/system/webconsole setprop options/tcp_listen = true # svcadm disable svc:/system/webconsole # svcadm enable svc:/system/webconsole -Andy When I open the link, the left frame lists a stacktrace (below) and the right frame is plain empty. Any suggestions? I tried substituting different SUNWzfsgr and SUNWzfsgu packages from older Solarises (x86/sparc, snv_77/84/89, sol10u3/u4), and directly substituting the zfs.jar file, but these actions resulted in either the same error or crash-and-restart of SMC Webserver. I didn't yet try installing an older SUNWmco* packages (a 10u4 system with SMC 3.0.2 works ok), I'm not sure it's a good idea ;) The system has JDK 1.6.0_06 per default, maybe that's the culprit? I tried setting it to JDL 1.5.0_15 and web-module zfs refused to start and register itself... === Application Error com.iplanet.jato.NavigationException: Exception encountered during forward Root cause = [java.lang.IllegalArgumentException: No enum const class com.sun.zfs.common.model.AclInheritProperty $AclInherit.restricted] Notes for application developers: * To prevent users from seeing this error message, override the onUncaughtException() method in the module servlet and take action specific to the application * To see a stack trace from this error, see the source for this page Generated Thu May 29 17:39:50 MSD 2008 === In fact, the traces in the logs are quite long (several screenfulls) and nearly the same; this one starts as: === com.iplanet.jato.NavigationException: Exception encountered during forward Root cause = [java.lang.IllegalArgumentException: No enum const class com.sun.zfs.common.model.AclInheritProperty $AclInherit.restricted] at com.iplanet.jato.view.ViewBeanBase.forward(ViewBeanBase.java:380) at com.iplanet.jato.view.ViewBeanBase.forwardTo(ViewBeanBase.java:261) at com .iplanet .jato .ApplicationServletBase.dispatchRequest(ApplicationServletBase.java: 981) at com .iplanet .jato .ApplicationServletBase.processRequest(ApplicationServletBase.java: 615) at com .iplanet .jato.ApplicationServletBase.doGet(ApplicationServletBase.java:459) at javax.servlet.http.HttpServlet.service(HttpServlet.java:690) ... === This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Disabling ZFS ACL
Did you try mounting with nfs version 3? mount -o vers=3 On May 28, 2008, at 10:38 AM, kevin kramer wrote: that is my thread and I'm still having issues even after applying that patch. It just came up again this week. [locahost] uname -a Linux dv-121-25.centtech.com 2.6.18-53.1.14.el5 #1 SMP Wed Mar 5 11:37:38 EST 2008 x86_64 x86_64 x86_64 GNU/Linux [localhost] cat /etc/issue CentOS release 5 (Final) Kernel \r on an \m [localhost: /n/scr20] touch test [localhost: /n/scr20] mv test /n/scr01/test/ ** this is a UFS mount on FreeBSD mv: preserving permissions for `/n/scr01/test/test': Operation not supported mv: preserving ACL for `/n/scr01/test/test': Operation not supported mv: preserving permissions for `/n/scr01/test/test': Operation not supported If I move it to the local /tmp, I get no errors. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS in S10U6 vs openSolaris 05/08
On May 27, 2008, at 1:44 PM, Rob Logan wrote: There is something more to consider with SSDs uses as a cache device. why use SATA as the interface? perhaps http://www.tgdaily.com/content/view/34065/135/ would be better? (no experience) We are pretty happy with RAMSAN SSD's (ours is RAM based, not flash). -Andy cards will start at 80 GB and will scale to 320 and 640 GB next year. By the end of 2008, Fusion io also hopes to roll out a 1.2 TB card. 160 parallel pipelines that can read data at 800 megabytes per second and write at 600 MB/sec 4K blocks and then streaming eight simultaneous 1 GB reads and writes. In that test, the ioDrive clocked in at 100,000 operations per second... beat $30 dollars a GB, ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Per-user home filesystems and OS-X Leopard anomaly
On May 21, 2008, at 11:15 AM, Bob Friesenhahn wrote: I encountered an issue that people using OS-X systems as NFS clients need to be aware of. While not strictly a ZFS issue, it may be encounted most often by ZFS users since ZFS makes it easy to support and export per-user filesystems. The problem I encountered was when using ZFS to create exported per-user filesystems and the OS-X automounter to perform the necessary mount magic. OS-X creates hidden .DS_Store directories in every directory which is accessed (http://en.wikipedia.org/wiki/.DS_Store). OS-X decided that it wanted to create the path /home/.DS_Store and it would not take `no' for an answer. First it would try to create /home/.DS_Store and then it would try an alternate name. Since the automounter was used, there would be an automount request for /home/.DS_Store, which does not exist on the server so the mount request would fail. Since OS-X does not take 'no' for an answer, there would be subsequent thousands of back to back mount requests. The end result was that 'mountd' was one of the top three resource consumers on my system, there would be bursts of high network traffic (1500 packets/second), and the affected OS-X system would operate more strangely than normal. The simple solution was to simply create a /home/.DS_Store directory on the server so that the mount request would succeed. Did you try this? http://support.apple.com/kb/HT1629 -Andy Bob == Bob Friesenhahn [EMAIL PROTECTED], http://www.simplesystems.org/users/bfriesen/ GraphicsMagick Maintainer,http://www.GraphicsMagick.org/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS and Sun Disk arrays - Opinions?
On May 16, 2008, at 10:04 AM, Robert Milkowski wrote: Hello James, 2) Does anyone have experiance with the 2540? JCM Kinda. I worked on adding MPxIO support to the mpt driver so JCM we could support the SAS version of this unit - the ST2530. JCM What sort of experience are you after? I'ver never used one JCM of these boxes in production - only ever for benchmarking and JCM bugfixing :-) I think Robert Milkowski might have one or two JCM of them, however. Yeah, I do have several of them (both 2530 and 2540). we did a try and buy of the 2510,2530 and 2540. 2530 (SAS) - cables tend to pop-out sometimes when you are around servers... then MPxIO does not work properly if you just hot-unplug and hot-replug the sas cable... there is still 2TB LUN size limit IIRC... other than that generally it is a good value Yeah the sff-8088 connectors are a bit rigid and clumsy, but the performance was better than everything we tested in the 2500 series. 2540 (FC) - 2TB LUN size limit IIRC, other than that it is a good value array Echo. We like the 2540 as well, and will be buying lots of them shortly. -- Best regards, Robert Milkowskimailto:[EMAIL PROTECTED] http://milek.blogspot.com -Andy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Image with DD from ZFS partition
On May 14, 2008, at 10:39 AM, Chris Siebenmann wrote: | Think what you are looking for would be a combination of a snapshot | and zfs send/receive, that would give you an archive that you can use | to recreate your zfs filesystems on your zpool at will at later time. Talking of using zfs send/recieve for backups and archives: the Solaris 10U4 zfs manpage contains some blood-curdling warnings about there being no cross-version compatability promises for the output of 'zfs send'. Can this be ignored in practice, or is it a real issue? It's real! You cant send and receive between versions of ZFS. (Speaking as a sysadmin, I certainly hope that it is a misplaced warning. Even ignoring backups and archives, imagine the fun if you cannot use 'zfs send | zfs receive' to move a ZFS filesystem from an old but reliable server running a stable old Solaris to your new, just installed server running the latest version of Solaris.) If you use external storage array attached via FC,iscsi,SAS etc, you can just do a 'zpool export', disconnect the storage from the old server, attach it to the new server then run 'zpool import' - and then do a 'zpool upgrade'. Unfortunately this doesn't help the thumpers so much :( - cks ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -Andy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS cli for REMOTE Administration
Paul B. Henson wrote: On Thu, 8 May 2008, Mark Shellenbaum wrote: we already have the ability to allow users to create/destroy snapshots over NFS. Look at the ZFS delegated administration model. If all you want is snapshot creation/destruction then you will need to grant snapshot,mount,destroy permissions. then on the NFS client mount go into .zfs/snapshot and do mkdir snapname. Providing the user has the appropriate permission the snapshot will be created. rmdir can be used to remove the snapshot. Now that is just uber-cool. Can you do that through the in kernel CIFS server too? Yes, it works over CIFS too. -Mark Great stuff! I confirmed that it does work, but its strange that I don't see the snapshot in 'zfs list' on the zfs box. Is that a bug or a feature? Im using XP - another thing is that if you right click in the .zfs/snapshot directory and do new - folder you will be stuck with a snapshot called New Folder. I couldn't rename it and the only way to delete it was to log into the machine and do a lil 'rm -Rf'. good news is that it is snapshotting :) I have a simple backup script where I use robocopy and then at the end I want to do a 'mkdir .zfs/snapshot/xxx', but I would eventually want to delete the oldest snapshot, similar to the zsnap.pl script floating around. Cant wait to try this on NFS, the whole reason we objected to snapshots in the first place in our org was because our admins didn't want to be involved with the users for the routine of working with snapshots. -Andy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS cli for REMOTE Administration
I can now see the snapshots in the CLI if i do a mkdir from the windows command line as opposed to getting stuck with a New Folder, which seems to confuse the CLI (maybe snapshots cant have spaces in the name? I was wondering, if there was an easy way for us to script zfs home filesystem creation upon connection to an AD joined cifs server? samba had some cool stuff with preexec and I just wonder if something like that is available for the kernel mode cifs driver. -Andy -Original Message- From: [EMAIL PROTECTED] on behalf of Andy Lubel Sent: Sun 5/11/2008 2:24 AM To: Mark Shellenbaum; Paul B. Henson Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] ZFS cli for REMOTE Administration Paul B. Henson wrote: On Thu, 8 May 2008, Mark Shellenbaum wrote: we already have the ability to allow users to create/destroy snapshots over NFS. Look at the ZFS delegated administration model. If all you want is snapshot creation/destruction then you will need to grant snapshot,mount,destroy permissions. then on the NFS client mount go into .zfs/snapshot and do mkdir snapname. Providing the user has the appropriate permission the snapshot will be created. rmdir can be used to remove the snapshot. Now that is just uber-cool. Can you do that through the in kernel CIFS server too? Yes, it works over CIFS too. -Mark Great stuff! I confirmed that it does work, but its strange that I don't see the snapshot in 'zfs list' on the zfs box. Is that a bug or a feature? Im using XP - another thing is that if you right click in the .zfs/snapshot directory and do new - folder you will be stuck with a snapshot called New Folder. I couldn't rename it and the only way to delete it was to log into the machine and do a lil 'rm -Rf'. good news is that it is snapshotting :) I have a simple backup script where I use robocopy and then at the end I want to do a 'mkdir .zfs/snapshot/xxx', but I would eventually want to delete the oldest snapshot, similar to the zsnap.pl script floating around. Cant wait to try this on NFS, the whole reason we objected to snapshots in the first place in our org was because our admins didn't want to be involved with the users for the routine of working with snapshots. -Andy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS partition info makes my system not boot
On Mar 11, 2008, at 4:58 PM, Bart Smaalders wrote: Frank Bottone wrote: I'm using the latest build of opensolaris express available from opensolaris.org. I had no problems with the install (its an AMD64 x2 3800+, 1gb physical ram, 1 ide drive for the os and 4*250GB sata drives attached to the motherboard - nforce based chipset). I create a zfs pool on the 4 sata drives as a raidZ and the pool works fine. If I reboot with any of the 4 drives connected the system hangs right after all the drives are detected on the post screen. I need to put them in a different system and zero them with dd in order to be able to reconnect them to my server and still have the system boot properly. Any ideas on how I can get around this? It seems like the onboard system itself is getting confused by the metadata ZFS is adding to the drive. The system already has the latest available bios from the manufacturer - I'm not using any hardware raid of any sort. This is likely the BIOS getting confused by the EFI label on the disks. Since there's no newer BIOS available there are two ways around this problem: 1) put a normal label on the disk and give zfs slice 2, or 2) don't have the BIOS do auto-detect on those drives. Many BIOSs let you select None for the disk type; this will allow the system to boot. Solaris has no problem finding the drives even w/o the BIOSs help... See if a BIOS update is available as well? - Bart -- Bart SmaaldersSolaris Kernel Performance [EMAIL PROTECTED] http://blogs.sun.com/barts You will contribute more with mercurial than with thunderbird. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Hardware RAID vs. ZFS RAID
With my (COTS) LSI 1068 and 1078 based controllers I get consistently better performance when I export all disks as jbod (MegaCli - CfgEachDskRaid0). Is that really 'all disks as JBOD'? or is it 'each disk as a single drive RAID0'? single disk raid0: ./MegaCli -CfgEachDskRaid0 Direct -a0 It may not sound different on the surface, but I asked in another thread and others confirmed, that if your RAID card has a battery backed cache giving ZFS many single drive RAID0's is much better than JBOD (using the 'nocacheflush' option may even improve it more.) My understanding is that it's kind of like the best of both worlds. You get the higher number of spindles and vdevs for ZFS to manage, ZFS gets to do the redundancy, and the the HW RAID Cache gives virtually instant acknowledgement of writes, so that ZFS can be on it's way. So I think many RAID0's is not always the same as JBOD. That's not to say that even True JBOD doesn't still have an advantage over HW RAID. I don't know that for sure. I have tried mixing hardware and zfs raid but it just doesn't make sense to use from a performance or redundancy standpoint why we would add those layers of complexity. In this case I'm building nearline so there isn't even a battery attached and I have disabled any caching on the controller. I have a SUN SAS HBA on the way which would be what I would use ultimately for my JBOD attachment. But I think there is a use for HW RAID in ZFS configs which wasn't always the theory I've heard. I have really learned not to do it this way with raidz and raidz2: #zpool create pool2 raidz c3t8d0 c3t9d0 c3t10d0 c3t11d0 c3t12d0 c3t13d0 c3t14d0 c3t15d0 Why? I know creating raidz's with more than 9-12 devices, but that doesn't cross that threshold. Is there a reason you'd split 8 disks up into 2 groups of 4? What experience led you to this? (Just so I don't have to repeat it. ;) ) I don't know why but with most setups I have tested (8 and 16 drive configs) dividing raid5 into 4 disks per vdev and 5 for a raidz2 perform better. Take a look at my simple dd test (filebench results as soon as I can figure out how to get it working proper with SOL10). = 8 SATA 500gb disk system with LSI 1068 (megaRAID ELP) - no BBU - bash-3.00# zpool history History for 'pool0-raidz': 2008-02-11.16:38:13 zpool create pool0-raidz raidz c2t0d0 c2t1d0 c2t2d0 c2t3d0 c2t4d0 c2t5d0 c2t6d0 c2t7d0 bash-3.00# zfs list NAME USED AVAIL REFER MOUNTPOINT pool0-raidz 117K 3.10T 42.6K /pool0-raidz bash-3.00# time dd if=/dev/zero of=/pool0-raidz/w-test.lo0 bs=8192 count=131072;time sync 131072+0 records in 131072+0 records out real0m1.768s user0m0.080s sys 0m1.688s real0m3.495s user0m0.001s sys 0m0.013s bash-3.00# time dd if=/pool0-raidz/w-test.lo0 of=/pool0-raidz/rw-test.lo0 bs=8192; time sync 131072+0 records in 131072+0 records out real0m6.994s user0m0.097s sys 0m2.827s real0m1.043s user0m0.001s sys 0m0.013s bash-3.00# time dd if=/dev/zero of=/pool0-raidz/w-test.lo1 bs=8192 count=655360;time sync 655360+0 records in 655360+0 records out real0m24.064s user0m0.402s sys 0m8.974s real0m1.629s user0m0.001s sys 0m0.013s bash-3.00# time dd if=/pool0-raidz/w-test.lo1 of=/pool0-raidz/rw-test.lo1 bs=8192; time sync 655360+0 records in 655360+0 records out real0m40.542s user0m0.476s sys 0m16.077s real0m0.617s user0m0.001s sys 0m0.013s bash-3.00# time dd if=/pool0-raidz/w-test.lo0 of=/dev/null bs=8192; time sync 131072+0 records in 131072+0 records out real0m3.443s user0m0.084s sys 0m1.327s real0m0.013s user0m0.001s sys 0m0.013s bash-3.00# time dd if=/pool0-raidz/w-test.lo1 of=/dev/null bs=8192; time sync 655360+0 records in 655360+0 records out real0m15.972s user0m0.413s sys 0m6.589s real0m0.013s user0m0.001s sys 0m0.012s --- bash-3.00# zpool history History for 'pool0-raidz': 2008-02-11.17:02:16 zpool create pool0-raidz raidz c2t0d0 c2t1d0 c2t2d0 c2t3d0 2008-02-11.17:02:51 zpool add pool0-raidz raidz c2t4d0 c2t5d0 c2t6d0 c2t7d0 bash-3.00# zfs list NAME USED AVAIL REFER MOUNTPOINT pool0-raidz 110K 2.67T 36.7K /pool0-raidz bash-3.00# time dd if=/dev/zero of=/pool0-raidz/w-test.lo0 bs=8192 count=131072;time sync 131072+0 records in 131072+0 records out real0m1.835s user0m0.079s sys 0m1.687s real0m2.521s user0m0.001s sys 0m0.013s bash-3.00# time dd if=/pool0-raidz/w-test.lo0 of=/pool0-raidz/rw-test.lo0 bs=8192; time sync 131072+0 records in 131072+0 records out real0m2.376s user0m0.084s sys 0m2.291s real0m2.578s user0m0.001s sys 0m0.013s bash-3.00# time dd if=/dev/zero of=/pool0-raidz/w-test.lo1 bs=8192 count=655360;time sync 655360+0 records in 655360+0 records out real0m19.531s user0m0.404s sys 0m8.731s real0m2.255s user0m0.001s sys
Re: [zfs-discuss] Hardware RAID vs. ZFS RAID
With my (COTS) LSI 1068 and 1078 based controllers I get consistently better performance when I export all disks as jbod (MegaCli - CfgEachDskRaid0). I even went through all the loops and hoops with 6120's, 6130's and even some SGI storage and the result was always the same; better performance exporting single disk than even the ZFS profiles within CAM. --- 'pool0': #zpool create pool0 mirror c2t0d0 c2t1d0 #zpool add pool0 mirror c2t2d0 c2t3d0 #zpool add pool0 mirror c2t4d0 c2t5d0 #zpool add pool0 mirror c2t6d0 c2t7d0 'pool2': #zpool create pool2 raidz c3t8d0 c3t9d0 c3t10d0 c3t11d0 #zpool add pool2 raidz c3t12d0 c3t13d0 c3t14d0 c3t15d0 I have really learned not to do it this way with raidz and raidz2: #zpool create pool2 raidz c3t8d0 c3t9d0 c3t10d0 c3t11d0 c3t12d0 c3t13d0 c3t14d0 c3t15d0 So when is thumper going to have an all SAS option? :) -Andy On Feb 7, 2008, at 2:28 PM, Joel Miller wrote: Much of the complexity in hardware RAID is in the fault detection, isolation, and management. The fun part is trying to architect a fault-tolerant system when the suppliers of the components can not come close to enumerating most of the possible failure modes. What happens when a drive's performance slows down because it is having to go through internal retries more than others? What layer gets to declare a drive dead? What happens when you start declaring the drives dead one by one because of they all seemed to stop responding but the problem is not really the drives? Hardware RAID systems attempt to deal with problems that are not always straight forward...Hopefully we will eventually get similar functionality in Solaris... Understand that I am a proponent of ZFS, but everything has it's use. -Joel This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Home Motherboard
Arcea, nice! Any word on whether 3ware has come around yet? I've been bugging them for months to do something to get a driver made for solaris. -Andy From: [EMAIL PROTECTED] on behalf of James C. McPherson Sent: Thu 11/22/2007 5:06 PM To: mike Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Home Motherboard mike wrote: I actually have a related motherboard, chassis, dual power-supplies and 12x400 gig drives already up on ebay too. If I recall Areca cards are supported in OpenSolaris... At the moment you can download the Areca arcmsr driver from areca.com.tw, but I'm in the process of integrating it into OpenSolaris http://bugs.opensolaris.org/view_bug.do?bug_id=6614012 6614012 add Areca SAS/SATA RAID adapter driver James C. McPherson -- Senior Kernel Software Engineer, Solaris Sun Microsystems http://blogs.sun.com/jmcp http://www.jmcp.homeunix.com/blog ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Yager on ZFS
On 11/15/07 9:05 AM, Robert Milkowski [EMAIL PROTECTED] wrote: Hello can, Thursday, November 15, 2007, 2:54:21 AM, you wrote: cyg The major difference between ZFS and WAFL in this regard is that cyg ZFS batch-writes-back its data to disk without first aggregating cyg it in NVRAM (a subsidiary difference is that ZFS maintains a cyg small-update log which WAFL's use of NVRAM makes unnecessary). cyg Decoupling the implementation from NVRAM makes ZFS usable on cyg arbitrary rather than specialized platforms, and that without cyg doubt constitutes a significant advantage by increasing the cyg available options (in both platform and price) for those cyg installations that require the kind of protection (and ease of cyg management) that both WAFL and ZFS offer and that don't require cyg the level of performance that WAFL provides and ZFS often may not cyg (the latter hasn't gotten much air time here, and while it can be cyg discussed to some degree in the abstract a better approach would cyg be to have some impartial benchmarks to look at, because the cyg on-disk block layouts do differ significantly and sometimes cyg subtly even if the underlying approaches don't). Well, ZFS allows you to put its ZIL on a separate device which could be NVRAM. Like RAMSAN SSD http://www.superssd.com/products/ramsan-300/ It is the only FC attached, Battery-backed SSD that I know of, and we have dreams of clusterfication. Otherwise we would use one of those PCI-Express based NVRAM cards that are on the horizon. My initial results for lots of small files was very pleasing. I dream of a JBOD with lots of disks + something like this built into 3u. Too bad Sun's forthcoming JBODS probably wont have anything similar to this... -Andy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Force SATA1 on AOC-SAT2-MV8
Jumpering drives by removing the cover? Do you mean opening the chassis because they aren't removable from the outside? Your cable is longer than 1 meter inside of a chasis?? I think sataI is 2 meters and sataII is 1 meter. As far as a system setting for demoting these to sataI I don't know, but I don't think its possible.. Don't hold me to that however, I only say that because THE way I demote them to sataI is by removing a jumper actually :) HTH, Andy On 11/2/07 12:29 PM, Eric Haycraft [EMAIL PROTECTED] wrote: I have a supermicro AOC-SAT2-MV8 and am having some issues getting drives to work. From what I can tell, my cables are to long to use with SATA2. I got some drives to work by jumpering them down to sata1, but other drives I can't jumper without opening the case and voiding the drive warranty. Does anyone know if there is a system setting to drop it back to SATA1? I use zfs on a raid2 if makes a difference. This is on release of OpenSolaris 74. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Mac OS X 10.5.0 Leopard ships with a readonly ZFS
Yeah im pumped about this new release today.. such harmony in my storage to be had. now if OSX only had a native iscsi target/initiator! -Andy Lubel -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Peter Woodman Sent: Friday, October 26, 2007 8:14 AM To: Kugutsumen Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Mac OS X 10.5.0 Leopard ships with a readonly ZFS it would seem that the reason that it's been pulled is that it's installed by default in the release version (9A581) - just tested it here, and willikers, it works! On 10/26/07, Kugutsumen [EMAIL PROTECTED] wrote: # zfs list ZFS Readonly implemntation is loaded! To download the full ZFS read/write kext with all functionality enabled, please go to http://developer.apple.com no datasets available Unfortunately, I can't find it on ADC yet and it seems that it was removed by Apple: Another turn in the Apple-ZFS saga. Apple has made available a developer preview of ZFS for Mac OS X with read/write capability. The preview is available to all ADC members. From the readme file: ZFS is a new filesystem from Sun Microsystems which has been ported by Apple to Mac OS X. The initial (10.5.0) release of Leopard will restrict ZFS to read-only, so no ZFS pools or filesystems can be modified or created. This Developer Preview will enable full read/write capability, which includes the creation/destruction of ZFS pools and filesystems. Update: Will it ever end? The release has been pulled from ADC by Apple. I can't wait to reformat all my external 2.5 drives with zfs. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS speed degraded in S10U4 ?
On 9/25/07 3:37 AM, Sergiy Kolodka [EMAIL PROTECTED] wrote: Hi Guys, I'm playing with Blade 6300 to check performance of compressed ZFS with Oracle database. After some really simple tests I noticed that default (well, not really default, some patches applied, but definitely noone bother to tweak disk subsystem or something else) installation of S10U3 is actually faster than S10U4, and a lot faster. Actually it's even faster on compressed ZFS with S10U3 than on uncompressed with S10U4. My configuration - default Update 3 LiveUpgraded to Update 4 with ZFS filesystem on dedicated disk, and I'm working with same files which are on same physical cylinders, so it's not likely a problem with HDD itself. Did you do a 'zpool upgrade -a'? I'm doing as simple as just $time dd if=file.dbf of=/dev/null in few parallel tasks. On Update3 it's somewhere close to 11m32s and on Update 4 it's around 12m6s. And it's both reading from compressed or uncompressed ZFS, numbers a little bit higher with compressed, couple of seconds more, which impressive by itself, but difference is the same, and strangest part is that reading file from compressed ZFS on U3 is faster than reading uncompressed with U4. I'm really surprised by this results, anyone else noticed that ? I'm running a 'motley group of disks' on an e450 acting as our jumpstart server and server build times are noticeably quicker since u4. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -Andy -- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] enterprise scale redundant Solaris 10/ZFS server providing NFSv4/CIFS
On 9/20/07 7:31 PM, Paul B. Henson [EMAIL PROTECTED] wrote: On Thu, 20 Sep 2007, Tim Spriggs wrote: It's an IBM re-branded NetApp which can which we are using for NFS and iSCSI. Yeah its fun to see IBM compete with its OEM provider Netapp. Ah, I see. Is it comparable storage though? Does it use SATA drives similar to the x4500, or more expensive/higher performance FC drives? Is it one of the models that allows connecting dual clustered heads and failing over the storage between them? I agree the x4500 is a sweet looking box, but when making price comparisons sometimes it's more than just the raw storage... I wish I could just drop in a couple of x4500's and not have to worry about the complexity of clustering sigh... zfs send/receive. Netapp is great, we have about 6 varieties in production here. But what I pay in maintenance and up front cost on just 2 filers, I can buy a x4500 a year, and have a 3 year warranty each time I buy. It just depends on the company you work for. I haven't played too much with anything but netapp and storagetek.. But once I got started on zfs I just knew it was the future; and I think netapp realizes that too. And if apple does what I think it will, it will only get better :) Fast, Cheap, Easy - you only get 2. Zfs may change that. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] enterprise scale redundant Solaris 10/ZFS server providing NFSv4/CIFS
a wide variety of JBOD and RAID array solutions. For ZFS, I recommend a configuration which allows ZFS to repair corrupted data. That would also be my preference, but if I were forced to use hardware RAID, the additional loss of storage for ZFS redundancy would be painful. Would anyone happen to have any good recommendations for an enterprise scale storage subsystem suitable for ZFS deployment? If I recall correctly, the SE we spoke with recommended the StorageTek 6140 in a hardware raid configuration, and evidently mistakenly claimed that Cluster would not work with JBOD. I really have to disagree, we have 6120 and 6130's and if I had the option to actually plan out some storage I would have just bought a thumper. You could probably buy 2 for the cost of that 6140. Thanks... -Andy Lubel -- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Zfs log device (zil) ever coming to Sol10?
I think we are very close to using zfs in our production environment.. Now that I have snv_72 installed and my pools set up with NVRAM log devices things are hauling butt. I've been digging to find out whether this capability would be put into Solaris 10, does anyone know? If not, then I guess we can probably be OK using SXCE (as Joyent did). Thanks, Andy Lubel ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Zfs log device (zil) ever coming to Sol10?
On 9/18/07 1:02 PM, Bryan Cantrill [EMAIL PROTECTED] wrote: Hey Andy, On Tue, Sep 18, 2007 at 12:59:02PM -0400, Andy Lubel wrote: I think we are very close to using zfs in our production environment.. Now that I have snv_72 installed and my pools set up with NVRAM log devices things are hauling butt. Interesting! Are you using a MicroMemory device, or is this some other NVRAM concoction? RAMSAN :) http://www.superssd.com/products/ramsan-400/ I've been digging to find out whether this capability would be put into Solaris 10, does anyone know? I would say it's probably unlikely, but I'll let Neil and the ZFS team speak for that. Do you mind if I ask what you're using ZFS for? - Bryan Today's answer: We want to use it for nearline backups (via nfs), eventually we would like to use zvols+iscsi to serve up for Oracle databases. My future answer: What cant we use ZFS for? If anyone wants to see my iozones just let me know. -Andy Lubel ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Zfs log device (zil) ever coming to Sol10?
On 9/18/07 2:26 PM, Neil Perrin [EMAIL PROTECTED] wrote: Andy Lubel wrote: On 9/18/07 1:02 PM, Bryan Cantrill [EMAIL PROTECTED] wrote: Hey Andy, On Tue, Sep 18, 2007 at 12:59:02PM -0400, Andy Lubel wrote: I think we are very close to using zfs in our production environment.. Now that I have snv_72 installed and my pools set up with NVRAM log devices things are hauling butt. Interesting! Are you using a MicroMemory device, or is this some other NVRAM concoction? RAMSAN :) http://www.superssd.com/products/ramsan-400/ May I ask roughly what you paid for it. I think perhaps we ought to get one in-house and check it out as well. Thanks: Neil. ~80k for the 128gb model. But we didn't pay anything for it, it was a customer return that the vendor wouldn't take back. Being that they have a fancy (we love) sun logo on the homepage I'm willing to bet that they would send you a demo unit. Let me know if I can help at all with that. -Andy Lubel -- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Zfs with storedge 6130
On 9/4/07 4:34 PM, Richard Elling [EMAIL PROTECTED] wrote: Hi Andy, my comments below... note that I didn't see zfs-discuss@opensolaris.org in the CC for the original... Andy Lubel wrote: Hi All, I have been asked to implement a zfs based solution using storedge 6130 and im chasing my own tail trying to decide how best to architect this. The storage space is going to be used for database dumps/backups (nearline storage). What is killing me is that I must mix hardware raid and zfs.. Why should that be killing you? ZFS works fine with RAID arrays. What kills me is the fact that I have a choice and it was hard to decide on which one was going to be at the top of the totem pole. From now on I only want JBOD! Works even better when I export each disk in my array as a single raid0 x14 then create the zpool :) #zpool create -f vol0 c2t1d12 c2t1d11 c2t1d10 c2t1d9 c2t1d8 c2t1d7 c2t1d6 c2t1d5 c2t1d4 c2t1d3 c2t1d2 c2t1d1 c2t1d0 spare c2t1d13 The storedge shelf has 14 FC 72gb disks attached to a solaris snv_68. I was thinking that since I cant export all the disks un-raided out to the solaris system that I would instead: (on the 6130) Create 3 raid5 volumes of 200gb each using the Sun_ZFS pool (128k segment size, read ahead enabled 4 disk). (On the snv_68) Create a raid0 using zfs of the 3 volumes from the 6130, using the same 128k stripe size. OK It seemed to me that if I was going to go for redundancy with a mixture of zfs and hardware raid that I would put the redundancy into the hardware raid and use striping at the zfs level, is that methodology the best way to think of it? The way to think about this is that ZFS can only correct errors when it has redundancy. By default, for dynamic stripes, only metadata is redundant. You can set the copies parameter to add redundancy on a per-file system basis, so you could set a different policy for data you really care about. Makes perfect sense. Since this is a nearline backup solution, I think we will be OK with a dynamic stripe. Once I get approved for thumper im definitely going to go raidz2. Since we are a huge Sun partner.. It should be easier than its been :( The only requirement ive gotten so far is that it can be written to and read from at a minimum of 72mb/s locally and 1gb/35sec via nfs. I suspect I would need at least 600gb of storage. I hope you have a test case for this. It is difficult for us to predict that sort of thing because there are a large number of variables. But in general, to get high bandwidth, you need large I/Os. That implies the application is responsible for it's use of the system, since the application is the source of I/Os. Its all going to be accessed via NFS and eventually iscsi, as soon as we figure out how to backup iscsi targets from the SAN itself. Anyone have any recommendations? The last time tried to create one 13 disk raid5 with zfs filesystem the performance was terrible via nfs.. But when I shared an nfs filesystem via a raidz or mirror things were much better.. So im nervous about doing this with only one volume in the zfs pool. 13 disk RAID-5 will suck. Try to stick with fewer devices in the set. See also http://mail.opensolaris.org/pipermail/zfs-discuss/2006-December/024194.html http://blogs.digitar.com/jjww/?itemid=44 I cant find a santricity download that will work with a 6130, but that's ok.. I just created 14 volumes per shelf :) hardware raid is so yesterday. That data is somewhat dated, as we now have the ability to put the ZIL on a different log device (Nevada b70 or later). This will be more obvious if the workload creates a lot of small files, less of a performance problem for large files. -- richard Got my hands on a Ram-San SSD 64gb and I'm using that for the zil.. Its crazy fast now. -Andy -- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] MS Exchange storage on ZFS?
On 9/6/07 2:51 PM, Joe S [EMAIL PROTECTED] wrote: Has anyone here attempted to store their MS Exchange data store on a ZFS pool? If so, could you please tell me about your setup? A friend is looking for a NAS solution, and may be interested in a ZFS box instead of a netapp or something like that. I don't see why it wouldn't using zvols and iscsi. We use iscsi in our rather large exchange implementation - not backed by zfs but I don't see why it couldn't be. PS no NAS solution will work for exchange will it? You have to use DAS/SAN or iscsi afaik. -Andy Thanks. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: ZFS Apple WWDC Keynote Absence
Yeah this is pretty sad, we had such plans for actually using our apple (PPC) hardware in our datacenter for something other than AFP and web serving. It also shows how limited apples vision seems to be. For 2 CEO's not to be on the same page demonstrates that there is something else going on rather than just we chose not to put a future ready file system into our next OS. And how its being dismissed by apple is quite upsetting. I wonder when we will see Johnny-cat and Steve-o in the same room talking about it. On 6/12/07 8:23 AM, Sunstar Dude [EMAIL PROTECTED] wrote: Yea, What is the deal with this? I am so bummed :( What the heck was Sun's CEO talking about the other day? And why the heck did Apple not include at least non-default ZFS support in Leopard? If no ZFS in Leapard, then what is all the Apple-induced-hype about? A trapezoidal Dock table? A transparent menu bar? Can anyone explain the absence of ZFS in Leopard??? I signed up for this forum just to post this. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss Andy Lubel -- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] No zfs_nocacheflush in Solaris 10?
Im using: zfs set:zil_disable 1 On my se6130 with zfs, accessed by NFS and writing performance almost doubled. Since you have BBC, why not just set that? -Andy On 5/24/07 4:16 PM, Albert Chin [EMAIL PROTECTED] wrote: On Thu, May 24, 2007 at 11:55:58AM -0700, Grant Kelly wrote: I'm running SunOS Release 5.10 Version Generic_118855-36 64-bit and in [b]/etc/system[/b] I put: [b]set zfs:zfs_nocacheflush = 1[/b] And after rebooting, I get the message: [b]sorry, variable 'zfs_nocacheflush' is not defined in the 'zfs' module[/b] So is this variable not available in the Solaris kernel? I think zfs:zfs_nocacheflush is only available in Nevada. I'm getting really poor write performance with ZFS on a RAID5 volume (5 disks) from a storagetek 6140 array. I've searched the web and these forums and it seems that this zfs_nocacheflush option is the solution, but I'm open to others as well. What type of poor performance? Is it because of ZFS? You can test this by creating a RAID-5 volume on the 6140, creating a UFS file system on it, and then comparing performance with what you get against ZFS. It would also be worthwhile doing something like the following to determine the max throughput the H/W RAID is giving you: # time dd of=raw disk if=/dev/zero bs=1048576 count=1000 For a 2Gbps 6140 with 300GB/10K drives, we get ~46MB/s on a single-drive RAID-0 array, ~83MB/s on a 4-disk RAID-0 array w/128k stripe, and ~69MB/s on a seven-disk RAID-5 array w/128k strip. -- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Motley group of discs? (doing it right, or right now)
I think it will be in the next.next (10.6) OSX, we just need to get apple to stop playing with their silly cell phone (that I cant help but want, damn them!). I have similar situation at home, but what I do is use Solaris 10 on a cheapish x86 box with 6 400gb IDE/SATA disks, I then make them into ISCSI targets and use that free GlobalSAN initiator ([EMAIL PROTECTED]). I once was like you, had 5 USB/Firewire drives hanging off everything and eventually I just got fed up with the mess of cables and wall warts. Perhaps my method of putting redundant and fast storage isn't as easy to achieve to everyone else. If you want more details about my setup, just email me directly, I don't mind :) -Andy On 5/7/07 4:48 PM, [EMAIL PROTECTED] [EMAIL PROTECTED] wrote: Lee, Yes, the hot spare (disk4) should kick if another disk in the pool fails and yes, the data is moved to disk4. You are correct: 160 GB (the smallest disk) * 3 + raidz parity info Here's the size of raidz pool comprised of 3 136-GB disks: # zpool list NAMESIZEUSED AVAILCAP HEALTH ALTROOT pool408G 98K408G 0% ONLINE - # zfs list NAME USED AVAIL REFER MOUNTPOINT pool 89.9K 267G 32.6K /pool The pool is 408GB in size but usable space in the pool is 267GB. If you added the 600GB disk to the pool, then you'll still lose out on the extra capacity because of the smaller disks, which is why I suggested using it as a spare. Regarding this: If I didn't need a hot spare, but instead could live with running out and buying a new drive to add on as soon as one fails, what configuration would I use then? I don't have any add'l ideas but I still recommend going with a spare. Cindy Lee Fyock wrote: Cindy, Thanks so much for the response -- this is the first one that I consider an actual answer. :-) I'm still unclear on exactly what I end up with. I apologize in advance for my ignorance -- the ZFS admin guide assumes knowledge that I don't yet have. I assume that disk4 is a hot spare, so if one of the other disks die, it'll kick into active use. Is data immediately replicated from the other surviving disks to disk4? What usable capacity do I end up with? 160 GB (the smallest disk) * 3? Or less, because raidz has parity overhead? Or more, because that overhead can be stored on the larger disks? If I didn't need a hot spare, but instead could live with running out and buying a new drive to add on as soon as one fails, what configuration would I use then? Thanks! Lee On May 7, 2007, at 2:44 PM, [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] wrote: Hi Lee, You can decide whether you want to use ZFS for a root file system now. You can find this info here: http://opensolaris.org/os/community/zfs/boot/ Consider this setup for your other disks, which are: 250, 200 and 160 GB drives, and an external USB 2.0 600 GB drive 250GB = disk1 200GB = disk2 160GB = disk3 600GB = disk4 (spare) I include a spare in this setup because you want to be protected from a disk failure. Since the replacement disk must be equal to or larger than the disk to replace, I think this is best (safest) solution. zpool create pool raidz disk1 disk2 disk3 spare disk4 This setup provides less capacity but better safety, which is probably important for older disks. Because of the spare disk requirement (must be equal to or larger in size), I don't see a better arrangement. I hope someone else can provide one. Your questions remind me that I need to provide add'l information about the current ZFS spare feature... Thanks, Cindy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
RE: [zfs-discuss] XServe Raid Complex Storage Considerations
They do need to start on the next filesystem and it seems very ideal for Apple. If they didn't then apple will be making a huge mistake because whatever FS's exist now, zfs has already pretty much trump'd it on almost every level except for maturity. I'm expecting ZFS and ISCSI(initiator and target) in Leopard. After all, OS X borrows from FreeBSD.. FreeBSD 7 has zfs ;) Andy -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Luke Scharf Sent: Wednesday, April 25, 2007 3:00 PM To: Toby Thain Cc: [EMAIL PROTECTED] Subject: Re: [zfs-discuss] XServe Raid Complex Storage Considerations Toby Thain wrote: On 25-Apr-07, at 12:17 PM, cedric briner wrote: hello the list, After reading the _excellent_ ZFS Best Practices Guide, I've seen in the section: ZFS and Complex Storage Consideration that we should configure the storage system to ignore command which will flush the memory into the disk. So does some of you knows how to tell Xserve Raid to ignore ``fsync'' requests ? After the announce that zfs will be included in Tiger, Much as I would like to see it, I am not aware of any such announcement from Apple, only rumours. FWIW, I heard the rumor from their sales guys. The wouldn't say whether it ZFS would be available in 10.5 or 10.6, and they wouldn't say whether it ZFS-boot would be available when ZFS is introduced -- but they did confirm that it's being worked on. -Luke ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
RE: [zfs-discuss] Re: ZFS+NFS on storedge 6120 (sun t4)
What I'm saying is ZFS doesn't play nice with NFS in all the scenarios I could think of: -Single second disk in a v210 (sun72g) write cache on and off = ~1/3 the performance of UFS when writing files using dd over an NFS mount using the same disk. -2 raid 5 volumes composing of 6 spindles each taking ~53 seconds to write 1gb over a NFS mounted zfs stripe,raidz or mirror of a storedge 6120 array with bbc, zil_disable'd and write cache off/on. In some testing dd would even seem to 'hang'. When any volslice is formatted UFS with the same NFS client - its ~17 seconds! We are likely going to just try iscsi instead, the behavior is non-existent. At some point though we would like to use ZFS based NFS mounts for things.. the current difference in performance just scares us! -Andy -Original Message- From: [EMAIL PROTECTED] on behalf of Roch - PAE Sent: Mon 4/23/2007 5:32 AM To: Leon Koll Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Re: ZFS+NFS on storedge 6120 (sun t4) Leon Koll writes: Welcome to the club, Andy... I tried several times to attract the attention of the community to the dramatic performance degradation (about 3 times) of NFZ/ZFS vs. ZFS/UFS combination - without any result : a href=http://www.opensolaris.org/jive/thread.jspa?messageID=98592;[1]/a , a href=http://www.opensolaris.org/jive/thread.jspa?threadID=24015;[2]/a. Just look at two graphs in my a href=http://napobo3.blogspot.com/2006/08/spec-sfs-bencmark-of-zfsufsvxfs.html;posting dated August, 2006/a to see how bad the situation was and, unfortunately, this situation wasn't changed much recently: http://photos1.blogger.com/blogger/7591/428/1600/sfs.1.png I don't think the storage array is a source of the problems you reported. It's somewhere else... Why do you say this ? My reading is that almost all NFS/ZFS complaints are either complaining about NFS performance vs direct attach, comparing UFS vs ZFS on disk with write cache enabled, or complaining about ZFS running on storage with NVRAM. Your complain is the one exception, SFS being worst with ZFS backend vs say UFS or VxFS. My points being: So NFS cannot match direct attach for some loads. It's a fact that we can't get around . Enabling the write cache gives is not a valid way to run NFS over UFS. ZFS on NVRAM storage, we need to make sure the storage does not flush the cache in response to ZFS requests. Then SFS over ZFS is being investigated by others within Sun. I believe we have stuff in the pipe to make ZFS match or exceed UFS on small server level loads. So I think your complaint is being heard. I personally find it always incredibly hard to do performance engineering around SFS. So my perspective is that improving the SFS numbers will more likely come from finding ZFS/NFS performance deficiencies on simpler benchmarks. -r [i]-- leon[/i] This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
RE: [zfs-discuss] ZFS+NFS on storedge 6120 (sun t4)
yeah i saw that post about the other arrays but none for this EOL'd hunk of metal. i have some 6130's but hopefully by the time they are implemented we will have retired this nfs stuff and stepped into zvol iscsi targets. thanks anyways.. back to the drawing board on how to resolve this! -Andy -Original Message- From: [EMAIL PROTECTED] on behalf of Torrey McMahon Sent: Fri 4/20/2007 6:00 PM To: Marion Hakanson Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] ZFS+NFS on storedge 6120 (sun t4) Marion Hakanson wrote: [EMAIL PROTECTED] said: We have been combing the message boards and it looks like there was a lot of talk about this interaction of zfs+nfs back in november and before but since i have not seen much. It seems the only fix up to that date was to disable zil, is that still the case? Did anyone ever get closure on this? There's a way to tell your 6120 to ignore ZFS cache flushes, until ZFS learns to do that itself. See: http://mail.opensolaris.org/pipermail/zfs-discuss/2006-December/024194.html The 6120 isn't the same as a 6130/61340/6540. The instructions referenced above won't work on a T3/T3+/6120/6320 ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
RE: [zfs-discuss] ZFS+NFS on storedge 6120 (sun t4)
Im not sure about the workload but I did configure the volumes with the block size in mind.. didnt seem to do much. it could be due to the fact im basically HW raid then zfs raid and i just dont know the equation to define a smarter blocksize. seems like if i have 2 arrays with 64kb striped together that 128k would be ideal for my zfs datasets, but again.. my logic isnt infinite when it comes to this fun stuff ;) The 6120 has 2 volumes each with 64k stripe size blocks. i then raidz'ed the 2 volumes and tried both 64k and 128k. i do get a bit of a performance gain on rewrite at 128k. These are dd tests by the way: *this one is locally, and works just great. bash-3.00# date ; uname -a Thu Apr 19 21:11:22 EDT 2007 SunOS yuryaku 5.10 Generic_125100-04 sun4u sparc SUNW,Sun-Fire-V210 ^---^ bash-3.00# df -k Filesystemkbytesused avail capacity Mounted on ... se6120 697761792 26 666303904 1%/pool/se6120 se6120/rfs-v10 31457280 9710895 2174638431%/pool/se6120/rfs-v10 bash-3.00# time dd if=/dev/zero of=/pool/se6120/rfs-v10/rw-test-1.loo bs=8192 count=131072 131072+0 records in 131072+0 records out real0m13.783s real0m14.136s user0m0.331s sys 0m9.947s *this one is from a HP-UX 11i system mounted to the v210 listed above: onyx:/rfs# date ; uname -a Thu Apr 19 21:15:02 EDT 2007 HP-UX onyx B.11.11 U 9000/800 1196424606 unlimited-user license ^^ onyx:/rfs# bdf Filesystem kbytesused avail %used Mounted on ... yuryaku.sol:/pool/se6120/rfs-v10 31457280 9710896 21746384 31% /rfs/v10 onyx:/rfs# time dd if=/dev/zero of=/rfs/v10/rw-test-2.loo bs=8192 count=131072 131072+0 records in 131072+0 records out real1m2.25s real0m29.02s real0m50.49s user0m0.30s sys 0m8.16s *my 6120 tidbits of interest: 6120 Release 3.2.6 Mon Feb 5 02:26:22 MST 2007 (xxx.xxx.xxx.xxx) Copyright (C) 1997-2006 Sun Microsystems, Inc. All Rights Reserved. daikakuji:/:1vol mode volume mounted cachemirror v1 yes writebehind off v2 yes writebehind off daikakuji:/:5vol list volumecapacity raid data standby v1 340.851 GB5 u1d01-06 u1d07 v2 340.851 GB5 u1d08-13 u1d14 daikakuji:/:6sys list controller : 2.5 blocksize : 64k cache : auto mirror : auto mp_support : none naca : off rd_ahead : off recon_rate : med sys memsize: 256 MBytes cache memsize : 1024 MBytes fc_topology: auto fc_speed : 2Gb disk_scrubber : on ondg : befit Am i missing something? As far as the RW test, i will tinker some more and paste the results soonish. Thanks in advance, Andy Lubel -Original Message- From: Bill Moore [mailto:[EMAIL PROTECTED] Sent: Fri 4/20/2007 5:13 PM To: Andy Lubel Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] ZFS+NFS on storedge 6120 (sun t4) When you say rewrites, can you give more detail? For example, are you rewriting in 8K chunks, random sizes, etc? The reason I ask is because ZFS will, by default, use 128K blocks for large files. If you then rewrite a small chunk at a time, ZFS is forced to read 128K, modify the small chunk you're changing, and then write 128K. Obviously, this has adverse effects on performance. :) If your typical workload has a preferred block size that it uses, you might try setting the recordsize property in ZFS to match - that should help. If you're completely rewriting the file, then I can't imagine why it would be slow. The only thing I can think of is the forced sync that NFS does on a file closed. But if you set zil_disable in /etc/system and reboot, you shouldn't see poor performance in that case. Other folks have had good success with NFS/ZFS performance (while other have not). If it's possible, could you characterize your workload in a bit more detail? --Bill On Fri, Apr 20, 2007 at 04:07:44PM -0400, Andy Lubel wrote: We are having a really tough time accepting the performance with ZFS and NFS interaction. I have tried so many different ways trying to make it work (even zfs set:zil_disable 1) and I'm still no where near the performance of using a standard NFS mounted UFS filesystem - insanely slow; especially on file rewrites. We have been combing the message boards and it looks like there was a lot of talk about this interaction of zfs+nfs back in november and before but since i have not seen much. It seems the only fix up to that date was to disable zil, is that still the case? Did anyone ever get closure on this? We are running solaris 10 (SPARC) .latest patched 11/06 release connecting directly via FC to a 6120 with 2 raid 5 volumes over a bge interface (gigabit). tried raidz, mirror and stripe