Re: [zfs-discuss] how to destroy a pool by id?
On Sun, 21 Jun 2009 13:56:52 +1000 (EST) Andre van Eyssen an...@purplecow.org wrote: On Sat, 20 Jun 2009, Cindy Swearingen wrote: I wish we had a zpool destroy option like this: # zpool destroy -really_dead tank2 Cindy, The moment we implemented such a thing, there would be a rash of requests saying: a) I just destroyed my pool with -really_dead - how can I get my data back??! b) I was able to recover my data from -really_dead - can we have -ultra-nuke please? Following your logic there shouldn't have existed a rm -f * option too. -- Dick Hoogendijk -- PGP/GnuPG key: 01D2433D + http://nagual.nl/ | nevada / OpenSolaris 2009.06 release + All that's really worth doing is what we do for others (Lewis Carrol) ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Mobo SATA migration to AOC-SAT2-MV8 SATA card
There are probably too many questions in my last post, so I will post the questions as separate forum threads. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Best controller card for 8 SATA drives ?
Hi, I'm trying to find out which controller card people here recommend that can drive 8 SATA hard drives and that would work with my Asus M2N-SLI Deluxe motherboard, which has following expansion slots: 2 x PCI Express x16 slot at x16, x8 speed (PCIe) The main requirements I have are: - drive 8 SATA drives - rock solid reliability with x86 OpenSolaris 2009.06 or SXCE - easy to identify failed drives and replace them (hot swap is not necessary but a bonus if supported) - I must be able to move disks with data from one controller to another of different brands (and back!), only doing zpool export and import, which implies the HBA must be able to run in JBOD-mode without storing or modify anything on the disks. And preferably, the drives must show up with the format command. - should support staggered spinup of drives preferably From limited research I can see that at least the following 3 main possibilities exist: 1. Supermicro AOC-SAT2-MV8 (PCI-X interface) (pure SATA) (~$100) 2. Supermicro AOC-USAS-L8i / AOC-USASLP-L8i (PCIe interface) (miniSAS to SATA cables) (~$100) 3. LSI SAS 3081E-R or other LSI 'MegaRAID' cards (PCIe interface) (miniSAS to SATA cables) (~$200+) 1. AOC-SAT2-MV8 : Again, from reading a bit, I can see that although the M2N-SLI Deluxe motherboard does not have a PCI-X slot, apparently it could take the AOC-SAT2-MV8 card in one of the PCIe slots, although the card would only run in 32-bit mode, instead of 64-bit mode, and would therefore run slower. 2. AOC-USAS-L8i : The AOC-USAS-L8i card looks possible too, again running in the PCIe slot, but the old threads I saw on this seem to talk about some device numbering issue which could make determining the right failed drive to pull out, a difficult task -- see here for more details: http://www.opensolaris.org/jive/thread.jspa?messageID=271751 http://www.opensolaris.org/jive/thread.jspa?threadID=46982tstart=90 3. LSI SAS 3081E-R or other LSI 'MegaRAID' cards : http://www.lsi.com/storage_home/products_home/host_bus_adapters/sas_hbas/lsisas3081er/index.html?remote=1locale http://www.newegg.com/Product/Product.aspx?Item=N82E16816118100 This forum thread from DEC 2007 doesn't sound too good regarding drive numbering (for identifying failed drives etc), but the thread is 18 months old, and perhaps the issues may have been resolved now? Also I noticed an extra '-R' in the model number I found, but this might be an omission of the original forum poster -- see here: http://www.opensolaris.org/jive/thread.jspa?threadID=46982tstart=90 I saw Ben Rockwood saying good things about the LSI MegaRAID cards, although the model he references supports only 4 internal and 4 external drives so is not what I want -- see here: http://opensolaris.org/jive/message.jspa?messageID=368445#368445 Perhaps there are better LSI MegaRAID cards that people know of and can recommend? Preferably not too expensive though, as it's for a home system :) If anyone can throw some light on these topics, I would be pleased to hear from you. Thanks a lot. Simon http://breden.org.uk/2008/03/02/a-home-fileserver-using-zfs/ -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
On Sun, 21 Jun 2009 06:35:50 PDT Simon Breden no-re...@opensolaris.org wrote: If anyone can throw some light on these topics, I would be pleased to hear from you. Thanks a lot. I follow this thread with much interest. Curious to see what'll come out of it. -- Dick Hoogendijk -- PGP/GnuPG key: 01D2433D + http://nagual.nl/ | nevada / OpenSolaris 2009.06 release + All that's really worth doing is what we do for others (Lewis Carrol) ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
After checking some more sources, it seems that if I used the AOC-SAT2-MV8 with this motherboard, I would need to run it on the standard PCI slot. Here is the full listing of the motherboard's expansion slots: 2 x PCI Express x16 slot at x16, x8 speed 2 x PCI Express x1 3 x PCI 2.2 --- this one -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Things I Like About ZFS
On Sat, Jun 20, 2009 at 3:30 AM, Dave Ringkorno-re...@opensolaris.org wrote: I'll start: - The commands are easy to remember -- all two of them. Which is easier, SVM or ZFS, to mirror your disks? I've been using SVM for years and still have to break out the manual to use metadb, metainit, metastat, metattach, metadetach, etc. I hardly ever have to break out the ZFS manual. I can actually remember the commands and options to do things. Don't even start me on VxVM. Hehe. The simplicity is interesting. I'm actually starting to get confused by those two commands, and start to wish it went down to one. - Boasting to the unconverted. We still have a lot of VxVM and SVM on Solaris, and LVM on AIX, in the office. The other admins are always having issues with storage migrations, full filesystems, Live Upgrade, corrupted root filesystems, etc. I love being able to offer solutions to their immediate problems, and follow it up with, You know, if your box was on ZFS this wouldn't be an issue. Out of inodes. Huh? The very concept is so antiquated. The great success of ZFS, to me, is the fact that it rapidly became essentially invisible. It just does its job and you soon forget that it's there (until you have to deal with one of the alternatives, which throws it into sharp relief). -- -Peter Tribble http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] x4500 resilvering spare taking forever?
UPDATE: It's now back down to 0.9% complete. Does anyone have a clue as to whats happening here or where I can look for problems? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] x4500 resilvering spare taking forever?
Are you taking snapshots periodically? If so, you're using a build old enough to restart resilver/scrub whenever a snapshot is taken. Actually yes, I take snapshots once an hour of various things. I'll try disabling them for the time being and see how far along it gets. Thanks! -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] x4500 resilvering spare taking forever?
Also, b57 is about 2 years old and misses the improvements in performance, especially in scrub performance. Yep, I know. I'll upgrade them at some point down the road, but they've been serving our needs nicely so far. Thanks! -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
I use the AOC-SAT2-MV8 in a ordinary PCI slot. The PCI slot maxes at 150MB/sec or so. That is the fastest you will get. That card works very good with Solaris/OpenSolaris. Detects automatically, etc. Ive heard though that it does not work with hot swapping discs - avoid this. However, In a PCI-X it will max at 1GB/sec. I have been thinkin about buying a server mobo (they have PCI-X) to get 1GB/sec. Or should I buy a PCIe card instead? I dont know. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] x4500 resilvering spare taking forever?
Joe Kearney wrote: Also, b57 is about 2 years old and misses the improvements in performance, especially in scrub performance. Yep, I know. I'll upgrade them at some point down the road, but they've been serving our needs nicely so far. Yep, it also suffers from the bug that restarts resilvers when you take a snapshot. This was fixed in b94. http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6343667 -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
just a side-question: I folthis thread with much interest. what are these * for ? why is followed turned into fol* on this board? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
On Sun, 21 Jun 2009 14:07:49 PDT roland no-re...@opensolaris.org wrote: just a side-question: I folthis thread with much interest. what are these * for ? why is followed turned into fol* on this board? The text of my original message was: On Sun, 21 Jun 2009 06:35:50 PDT Simon Breden no-re...@opensolaris.org wrote: If anyone can throw some light on these topics, I would be pleased to hear from you. Thanks a lot. I follow this thread with much interest. Curious to see what'll come out of it. Does the change occur again? -- Dick Hoogendijk -- PGP/GnuPG key: 01D2433D + http://nagual.nl/ | nevada / OpenSolaris 2009.06 release + All that's really worth doing is what we do for others (Lewis Carrol) ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
Hey Kebabber, long time no hear! :) It's great to hear that you've had good experiences with the card. It's a great pity to have throughput drop from a potential 1GB/s to 150MB/s, but as most of my use of the NAS is across the network, and not local intra-NAS transfers, this should not be a problem. Of course, with a single GbE connection speeds are normally limited to around 50MB/s or so anyway... Tell me, have you had any drives fail and had to figure out how to identify the failed drive and replace it resilver using the AOC-SAT2-MV8, or have you tried any experiments to test resilvering ? I'm just curious as to how easy it is to do this with this controller card. Like yourself, I was toying with the idea of upgrading and buying a shiny new mobo with dual 64-bit PCI-X slots and socket LGA1366 for Xeon 5500 series (Nehalem) processors -- the SuperMicro X8SAX here: http://www.supermicro.com/products/motherboard/Xeon3000/X58/X8SAX.cfm Then I added up the price of all the components and decided to try to make do with the existing kit and just do an upgrade. So I narrowed down possible SATA controllers to the above choices and I'm interested in people's experiences of using these controllers to help me decide. Seems like the cheapest and simplest choice will be the AOC-SAT2-MV8, and I just take a hit on the reduced speed -- but that won't be a big problem. However, as I have 2 x PCIe x16 slots available, if the AOC-USAS-L8i is reliable and doesn't have issues now with identifying drive ids, and supports JBOD mode, then it looks like the better choice. It is uses the more modern PCI Express (PCIe) interface, rather than the ageing PCI-X interface, fine as I'm sure it is. Simon -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
On Mon 22/06/09 02:07 , roland no-re...@opensolaris.org sent: just a side-question: I folthis thread with much interest. what are these * for ? why is followed turned into fol* on this board? It isn't a board, it's a mail list. All the forum does is bugger up the formatiing and threading of emails! -- Ian ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Things I Like About ZFS
OK, my turn: - combining file system + volume manager + RAID + pool + scrub + resilvering + snapshots + rollback + end-to-end integrity + 256-but block checksums + on-the-fly healing of blocks with checksum errors on read - one liners that are mostly remembered, and simple to guess if forgotten - with RAID-Z2, superb protection + hot spares - easy sharing via CIFS and NFS - iSCSI as a block device target - open source software RAID no proprietary hardware RAID card required - free - did I forget something? :) -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] x4500 resilvering spare taking forever?
Joe Kearney wrote: UPDATE: It's now back down to 0.9% complete. Does anyone have a clue as to whats happening here or where I can look for problems? There was also a bug which restarted resilvers each time you issue a zpool status command as a privilaged user. Make sure to check the progress by issuing zpool status commands as a non-privilaged user. -- Andrew ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zfs find list of properties per dataset
How can I list out all the properties available for a filesystem? Banging away in man zfs seems to be going nowhere pretty fast. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs find list of properties per dataset
On Sun, 21 Jun 2009 18:28:56 -0500 Harry Putnam rea...@newsguy.com wrote: How can I list out all the properties available for a filesystem? Banging away in man zfs seems to be going nowhere pretty fast. Is zfs get all datasetname not helping? $ zfs get all rpool NAME PROPERTYVALUE SOURCE rpool typefilesystem - rpool creationWed May 6 21:51 2009 - rpool used85.9G - rpool available 115G- rpool referenced 81.5K - rpool compressratio 1.00x - rpool mounted yes - rpool quota nonedefault rpool reservation nonedefault rpool recordsize 128Kdefault rpool mountpoint /rpool default rpool sharenfsoff default rpool checksumon default rpool compression off default rpool atime on default rpool devices on default rpool execon default rpool setuid on default rpool readonlyoff default rpool zoned off default rpool snapdir hidden default rpool aclmode groupmask default rpool aclinherit restricted default rpool canmounton default rpool shareiscsi off default rpool xattr on default rpool copies 1 default rpool version 3 - rpool utf8onlyoff - rpool normalization none- rpool casesensitivity sensitive - rpool vscan off default rpool nbmand off default rpool sharesmboff default rpool refquotanonedefault rpool refreservation nonedefault rpool primarycacheall default rpool secondarycache all default rpool usedbysnapshots 0 - rpool usedbydataset 81.5K - rpool usedbychildren 85.9G - rpool usedbyrefreservation0 - rpool org.opensolaris.caiman:install ready local James C. McPherson -- Senior Kernel Software Engineer, Solaris Sun Microsystems http://blogs.sun.com/jmcp http://www.jmcp.homeunix.com/blog Kernel Conference Australia - http://au.sun.com/sunnews/events/2009/kernel ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs find list of properties per dataset
James C. McPherson james.mcpher...@sun.com writes: [...] Is zfs get all datasetname not helping? [...] Gack... quite the reverse ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs find list of properties per dataset
On Sun, 21 Jun 2009 18:47:30 -0500 Harry Putnam rea...@newsguy.com wrote: James C. McPherson james.mcpher...@sun.com writes: [...] Is zfs get all datasetname not helping? [...] Gack... quite the reverse your original question was somewhat vague - could you clarify what it is you need to find out? James -- Senior Kernel Software Engineer, Solaris Sun Microsystems http://blogs.sun.com/jmcp http://www.jmcp.homeunix.com/blog Kernel Conference Australia - http://au.sun.com/sunnews/events/2009/kernel ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs find list of properties per dataset
James C. McPherson james.mcpher...@sun.com writes: your original question was somewhat vague - could you clarify what it is you need to find out? I wanted to see a list of all the properties available to zfs on a filesystem. Your answer was very helpful... thanks My quick reply may not have conveyed that I was happy with your answer and that it helps a lot. The actual exact thing I was looking into was where snapdir was hidden and where visible. But I didn't know the name of the option (snapdir). Once I saw the full list... of course it was obvious. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] two pools on boot disk?
On Sat, Jun 20, 2009 at 7:02 PM, Michael Sullivanmichael.p.sulli...@mac.com wrote: One really interesting bit is how easily it is to make the disk in a pool bigger by doing a zpool replace on the device. It couldn't have been any easier with ZFS. It's interesting how you achieved that, although it'd be much easier if the installer supports that from the GUI instead of having to use zpool replace as a workaround. I believe using export-import as described in http://www.solarisinternals.com/wiki/index.php/ZFS_Troubleshooting_Guide#ZFS_Root_Pool_Recovery should also work. -- Fajar ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
I'll chime in as a happy owner of the LSI SAS 3081E-R PCI-E board. It works just fine. You need to get lsiutil from the LSI web site to fully access all the functionality, and they cleverly hide the download link only under their FC HBAs on their support site, even though it works for everything. As for identifying disks, you can just use lsiutil: root:gandalf 0 # lsiutil -p 1 42 LSI Logic MPT Configuration Utility, Version 1.62, January 14, 2009 1 MPT Port found Port Name Chip Vendor/Type/RevMPT Rev Firmware Rev IOC 1. mpt0 LSI Logic SAS1068E B3 105 011a 0 mpt0 is /dev/cfg/c6 B___T___L Type Operating System Device Name 0 0 0 Disk /dev/rdsk/c6t0d0s2 0 1 0 Disk /dev/rdsk/c6t1d0s2 0 2 0 Disk /dev/rdsk/c6t2d0s2 0 3 0 Disk /dev/rdsk/c6t3d0s2 0 4 0 Disk /dev/rdsk/c6t4d0s2 0 5 0 Disk /dev/rdsk/c6t5d0s2 0 6 0 Disk /dev/rdsk/c6t6d0s2 0 7 0 Disk /dev/rdsk/c6t7d0s2 ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zfs IO scheduler
Hello, Trying to understand the ZFS IO scheduler, because of the async nature it is not very apparent, can someone give a short explanation for each of these stack traces and for their frequency this is the command dd if=/dev/zero of=/test/test1/trash count=1 bs=1024k;sync no other IO is happening to the test pool. OS is on a zfs pool (rpool) I don't see any zio_vdev_io_start in any of the function stacks, any idea why? dtrace -n 'io:::start { @a[stack()] = count(); }' dtrace: description 'io:::start ' matched 6 probes genunix`bdev_strategy+0x44 zfs`vdev_disk_io_start+0x2a8 zfs`zio_execute+0x74 genunix`taskq_thread+0x1a4 unix`thread_start+0x4 20 genunix`bdev_strategy+0x44 zfs`vdev_disk_io_start+0x2a8 zfs`zio_execute+0x74 zfs`vdev_queue_io_done+0x84 zfs`vdev_disk_io_done+0x4 zfs`zio_execute+0x74 genunix`taskq_thread+0x1a4 unix`thread_start+0x4 31 genunix`bdev_strategy+0x44 zfs`vdev_disk_io_start+0x2a8 zfs`zio_execute+0x74 zfs`vdev_mirror_io_start+0x1b4 zfs`zio_execute+0x74 zfs`vdev_mirror_io_start+0x1b4 zfs`zio_execute+0x74 genunix`taskq_thread+0x1a4 unix`thread_start+0x4 34 genunix`bdev_strategy+0x44 zfs`vdev_disk_io_start+0x2a8 zfs`zio_execute+0x74 zfs`vdev_mirror_io_start+0x1b4 zfs`zio_execute+0x74 genunix`taskq_thread+0x1a4 unix`thread_start+0x4 45 genunix`bdev_strategy+0x44 zfs`vdev_disk_io_start+0x2a8 zfs`zio_execute+0x74 zfs`vdev_queue_io_done+0x9c zfs`vdev_disk_io_done+0x4 zfs`zio_execute+0x74 genunix`taskq_thread+0x1a4 unix`thread_start+0x4 53 -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
On Sun, 21 Jun 2009 19:01:31 -0700 Carson Gaspar car...@taltos.org wrote: I'll chime in as a happy owner of the LSI SAS 3081E-R PCI-E board. It works just fine. You need to get lsiutil from the LSI web site to fully access all the functionality, and they cleverly hide the download link only under their FC HBAs on their support site, even though it works for everything. As a member of the team which works on mpt(7d), I'm disappointed that\ you believe you need to use lsiutil to fully access all the functionality of the board. What gaps have you found in mpt(7d) and the standard OpenSolaris tools that lsiutil fixes for you? What is the full functionality that you believe is missing? As for identifying disks, you can just use lsiutil: ... or use cfgadm(1m) which has had this ability for many years. root:gandalf 0 # lsiutil -p 1 42 LSI Logic MPT Configuration Utility, Version 1.62, January 14, 2009 1 MPT Port found Port Name Chip Vendor/Type/RevMPT Rev Firmware Rev IOC 1. mpt0 LSI Logic SAS1068E B3 105 011a 0 mpt0 is /dev/cfg/c6 B___T___L Type Operating System Device Name 0 0 0 Disk /dev/rdsk/c6t0d0s2 0 1 0 Disk /dev/rdsk/c6t1d0s2 0 2 0 Disk /dev/rdsk/c6t2d0s2 0 3 0 Disk /dev/rdsk/c6t3d0s2 0 4 0 Disk /dev/rdsk/c6t4d0s2 0 5 0 Disk /dev/rdsk/c6t5d0s2 0 6 0 Disk /dev/rdsk/c6t6d0s2 0 7 0 Disk /dev/rdsk/c6t7d0s2 You can get that information from use of cfgadm(1m). James C. McPherson -- Senior Kernel Software Engineer, Solaris Sun Microsystems http://blogs.sun.com/jmcp http://www.jmcp.homeunix.com/blog Kernel Conference Australia - http://au.sun.com/sunnews/events/2009/kernel ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
On Sun, 21 Jun 2009, Carson Gaspar wrote: I'll chime in as a happy owner of the LSI SAS 3081E-R PCI-E board. It works just fine. You need to get lsiutil from the LSI web site to fully access all the functionality, and they cleverly hide the download link only under their FC HBAs on their support site, even though it works for everything. I'll add another vote for the LSI products. I have a four port PCI-X card in my V880, and the performance is good and the product is well behaved. The only caveats: 1. Make sure you upgrade the firmware ASAP 2. You may need to use lsiutil to fiddle the target mappings Andre. -- Andre van Eyssen. mail: an...@purplecow.org jabber: an...@interact.purplecow.org purplecow.org: UNIX for the masses http://www2.purplecow.org purplecow.org: PCOWpix http://pix.purplecow.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
I only have a 32bit PCI bus in the Intel Atom 330 board, so I have no choice than to be slower, but I can confirm that the Supermicro dac-sata-mv8 (SATA-1) card works just fine, and does display in cfgadm. (Hot-swapping is possible). I have been told aoc-sat2-mv8 does as well (SATA-II) but I have not personally tried it. Lund -- Jorgen Lundman | lund...@lundman.net Unix Administrator | +81 (0)3 -5456-2687 ext 1017 (work) Shibuya-ku, Tokyo| +81 (0)90-5578-8500 (cell) Japan| +81 (0)3 -3375-1767 (home) ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] two pools on boot disk?
Fajar, Yes, you could probably do send/receive from one pool to another, but that would be somewhat more time consuming and you'd have to make sure everything was right in your GRUB menu.lst as well as boot blocks, not to mention the potential for namespace collisions when dealing with a root pool. But this is missing my point. The thing I found more interesting was that a pool could be increased in space by doing a zpool replace with a larger disk. This means if say, you have a pool of 100GB disks and you want to increase the size, you can replace them with bigger disks effectively growing the pool. Not sure how this works out with configurations other than in RAID 0 and RAID 1, but I thought it was a pretty nice feature knowing I can put bigger disks in really easily. I also agree the installer should have an expert mode for configuring disks. The all-or-nothing approach is easy for people who have never been exposed to Solaris or OpenSolaris, but leaves people out in the cold if you wish to have different configuration for your disks. The Automated Installer, is supposed to give this sort of flexibility, but I haven't tried it out yet. Regards, Mike --- Michael Sullivan michael.p.sulli...@me.com http://www.kamiogi.net/ Japan Mobile: +81-80-3202-2599 US Phone: +1-561-283-2034 On 22 Jun 2009, at 11:00 , Fajar A. Nugraha wrote: On Sat, Jun 20, 2009 at 7:02 PM, Michael Sullivanmichael.p.sulli...@mac.com wrote: One really interesting bit is how easily it is to make the disk in a pool bigger by doing a zpool replace on the device. It couldn't have been any easier with ZFS. It's interesting how you achieved that, although it'd be much easier if the installer supports that from the GUI instead of having to use zpool replace as a workaround. I believe using export-import as described in http://www.solarisinternals.com/wiki/index.php/ZFS_Troubleshooting_Guide#ZFS_Root_Pool_Recovery should also work. -- Fajar ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
Jorgen Lundman wrote: I only have a 32bit PCI bus in the Intel Atom 330 board, so I have no choice than to be slower, but I can confirm that the Supermicro dac-sata-mv8 (SATA-1) card works just fine, and does display in cfgadm. (Hot-swapping is possible). I have been told aoc-sat2-mv8 does as well (SATA-II) but I have not personally tried it. Lund I have an AOC-SAT2-MV8 in an older Opteron-based system. It's a 2-socket, Opteron 252 system with 8GB of RAM, and PCI-X slots. It's one of the newer AOC cards with the latest Marvell chipset, and it works like a champ - very well, smooth, and I don't see any problems. Simple, out-of-the-box installation and works with no tinkering at all (with OS 2008.11 and 2009.05). That said, there's a couple of things you want to be aware of about the AOC: (1) it uses normal sata cables. This is really nice in terms of availability (you can get any length you want easily at any computer store), but it's a bit messy compared to the nice multi-lane ones. (2) It's a PCI-X card, and will run at 133Mhz. I have a second gigabit ethernet card in my motherboard, which limits the two PCI-X cards to 100Mhz.The down side is that with 8 drives and 2 gigabit ethernet interfaces, it's not hard to flood the PCI-X bus (which can pump 100Mhz x 64bit = 6400 Mbps max, but about 50% of that under real usage) (3) as a PCI-X card, it's a two-third's length, low-profile card. It will fit in any PCI-X slot you have. However, if you are trying to put it in a 32-bit PCI slot, be aware that it extends about 2 inches (50mm) beyond the back of the PCI slot. Make sure you have the proper clearances for such a card. Also, it's a 3.3v card (won't work in 5v slots). None of this should be a problem in any modern motherboard/case setup, only in really old stuff. (4) All the SATA connectors are on the end of the card, which means you'll need _at least_ another 1 inch (25mm) clearance to plug the cables in. As much as I like the card, these days, I'd chose the LSI PCI-E model. The PCI-E bus is just superior to PCI-X - you get much less bus contention which means it's easier to get full throughput from each card One more thing: I've found that the newest MLC-based SSDs with the newer barefoot SATA controller and 64MB or more of cache are more than suitable for use as Read cache, and they actually do OK as write-cache, too. Particularly for small business server machine (those that have have 8-12 data drives, total). And, these days, there's nice little funky dual-2.5 drives in a floppy form-factor things. http://www.addonics.com/products/mobile_rack/doubledrive.asp Example new SSD for Readzilla/Logzilla : http://www.ocztechnology.com/products/flash_drives/ocz_summit_series_sata_ii_2_5-ssd -- Erik Trimble Java System Support Mailstop: usca22-123 Phone: x17195 Santa Clara, CA ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] cutting up a SSD for read/log use...
I just looked at pricing for the higher-end MLC devices, and it looks like I'm better off getting a single drive of 2X capacity than two with X capacity. Leaving aside the issue that by using 2 drives I get 2 x 3.0Gbps SATA performance instead of 1 x 3.0Gbps, are there problems with using two slices instead of whole-drives? That is, one slice for Read and the other for ZIL? My main concern is exactly how the on-drive cache would be used in a two-slices configuration. In order to get decent performance, I really need the on-drive cache to be used properly. -- Erik Trimble Java System Support Mailstop: usca22-123 Phone: x17195 Santa Clara, CA ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Speeding up resilver on x4500
It is currently taking ~1 week to resilver an x4500 running S10U6, recently patched with~170M small files on ~170 datasets after a disk failure/replacement, i.e., scrub: resilver in progress for 53h47m, 30.72% done, 121h19m to go Is there anything that can be tuned to improve this performance, e.g., adding a faster cache device for reading and/or writing? I am also curious if anyone has a prediction on when the snapshot-restarting-resilvering bug will be patched in Solaris 10? http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6343667 Thanks. -- Stuart Anderson ander...@ligo.caltech.edu http://www.ligo.caltech.edu/~anderson ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] cutting up a SSD for read/log use...
Hi Erik, On 22/06/2009, at 1:15 PM, Erik Trimble wrote: I just looked at pricing for the higher-end MLC devices, and it looks like I'm better off getting a single drive of 2X capacity than two with X capacity. Leaving aside the issue that by using 2 drives I get 2 x 3.0Gbps SATA performance instead of 1 x 3.0Gbps, are there problems with using two slices instead of whole-drives? That is, one slice for Read and the other for ZIL? The benefit you will get using 2 drives instead of 1 will be doubling your IOPS which will improve your overall performance, especially when using those drives as ZILs. Are you planning on using these drives as primary data storage and ZIL for the same volumes or as primary storage for (say) your rpool and ZIL for a data pool on spinning metal? cheers, James ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs find list of properties per dataset
Harry Putnam wrote: James C. McPherson james.mcpher...@sun.com writes: your original question was somewhat vague - could you clarify what it is you need to find out? I wanted to see a list of all the properties available to zfs on a filesystem. NB, the user and group quotas are stored as properties, but get all does not return them. So it is not a true statement that get all returns all of the properties. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Speeding up resilver on x4500
Stuart Anderson wrote: It is currently taking ~1 week to resilver an x4500 running S10U6, recently patched with~170M small files on ~170 datasets after a disk failure/replacement, i.e., wow, that is impressive. There is zero chance of doing that with a manageable number of UFS file systems. scrub: resilver in progress for 53h47m, 30.72% done, 121h19m to go Is there anything that can be tuned to improve this performance, e.g., adding a faster cache device for reading and/or writing? Resilver tends to be bound by one of two limits: 1. sequential write performance of the resilvering device 2. random I/O performance of the non-resilvering devices A while back, I was doing some characterization of this, but the funding disappeared :-( So, it is unclear whether or how caching might help. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] cutting up a SSD for read/log use...
Erik Trimble wrote: I just looked at pricing for the higher-end MLC devices, and it looks like I'm better off getting a single drive of 2X capacity than two with X capacity. Leaving aside the issue that by using 2 drives I get 2 x 3.0Gbps SATA performance instead of 1 x 3.0Gbps, are there problems with using two slices instead of whole-drives? That is, one slice for Read and the other for ZIL? My main concern is exactly how the on-drive cache would be used in a two-slices configuration. In order to get decent performance, I really need the on-drive cache to be used properly. Is the on-disk cache volatile? For most SSDs I'm familiar with, the on-disk cache is non-volatile, so all of the rules pertaining to whole disks with volatile write buffers are nullified. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Speeding up resilver on x4500
On Jun 21, 2009, at 8:57 PM, Richard Elling wrote: Stuart Anderson wrote: It is currently taking ~1 week to resilver an x4500 running S10U6, recently patched with~170M small files on ~170 datasets after a disk failure/replacement, i.e., wow, that is impressive. There is zero chance of doing that with a manageable number of UFS file systems. However, it is a bit disconcerting to have to run with reduced data protection for an entire week. While I am certainly not going back to UFS, it seems like it should be at least theoretically possible to do this several orders of magnitude faster, e.g., what if every block on the replacement disk had its RAIDZ2 data recomputed from the degraded array regardless of whether the pool was using it or not. In that case I would expect it to be able to sequentially reconstruct in the same few hours it would take a HW RAID controller to do the same RAID6 job. Perhaps there needs to be an option to re-order the loops for resilvering on pools with lots of small files to resilver in device order rather than filesystem order? scrub: resilver in progress for 53h47m, 30.72% done, 121h19m to go Is there anything that can be tuned to improve this performance, e.g., adding a faster cache device for reading and/or writing? Resilver tends to be bound by one of two limits: 1. sequential write performance of the resilvering device 2. random I/O performance of the non-resilvering devices A quick look at iostat leads me to conjecture that the vdev rebuilding is taking a very low priority compared to ongoing application I/O (NFSD in this case). Are there any ZFS knobs that control the relative priority of resilvering to other disk I/O tasks? Thanks. -- Stuart Anderson ander...@ligo.caltech.edu http://www.ligo.caltech.edu/~anderson ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Backup schemes invoving rsync Linux_rmt = Osol
There is a lot more about snapshots and backup schemes that I don't know than there is that I do know... My needs are pretty simple and small compared to some of the heavy users here, but I'd like to run just the rough idea by the group. One backup I need to make is from to linux online servers to a zpool. Whats involved is two websites I use sometimes quite a bit but more often not so much and sometimes not at all for quite a stretch. I thought I'd rsync the base directory to a zpool using the `--inplace' flag to rsync that has been discussed here a time or two. It appears the snapshot setup was including the specific zpool I plan to use but at a frequency level way above what seems I'd need. But them I'm not really sure how to go at this. I see mnthly weekly hourly and frequent which appear to be every 15 min. I thought it might be better to turn that off completely. Then use a cron job to run the rsync once per week. The same little script would run a snapshot following the completion of the rsync run. I'd probably use something like this for rsync: Cron calls my script and runs: rsync -avvz --stats --delete --exclude-from='exclude_file' --delete-excluded --inplace m...@remote:/home/me/public_html/ /www/remote_www/ It would be run with the `authorized_keys' mechanism Then runs a snapshot of zfs z3/www I'm thinking that snapshot would catch every thing since the last run. Is that a reasonable plan? Sorry to ruminate on here... but would like any ideas you may have. I'd probably do something very similar to the www plan with the home lan machines... Two linux boxes are no problem, but I haven't really figured out the two windows XP machines yet. It seems there's always been bugs in Cifs implementation preventing reliable connections.. Currently I can only see top level of any windows shares from zfs host. A while back I had `retrospect' (a windows version) connecting to zfs machine, but that fell through so often... it was not working. Windows based backup tools (at least some of them) expect the admin to be able to navigate to appropriate shares or at least use UNK addressing. But I have yet to get the zfs machine to show up in windows network places. Even when I can force things by typing an UNK address on windows explorer there is a very good chance after either machine is rebooted ... the connection would not work thereafter. Seems like initiating the connections from zfs host is a better way to go. And I'm currently looking in to using Bacula. Using rsync against the windows machines seems to mostly be a non-starter due to something in ntfs or the OS that makes files appear to be all new every so often causing rsync to be fooled into pulling the same files across again. That, by itself is bad enough but not sure what impact it might have on zfs snapshots. .. maybe none... long as --inplace is being used. I've been pretty lazy about it and have continued to backup the windows machines with things like ghost... making complete drive images. But since it seems like nothing but headaches trying to connect to the zfs host reliably...(From windows), I run the backup onto an internal drive and then have to move the monster file again to get it on a zfs filesystem. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] bug access
Hi, This CR has been marked as incomplete by User 1-UM-1502 for the reason Need More Info. Please update the CR providing the information requested in the Evaluation and/or Comments field. hmmm - wondering, how to find out, what 'more info' means and how to provide this info. There is no URL in the bug-response and it even seems to be impossible, to obtain the current state via http://bugs.opensolaris.org/view_bug.do?bug_id= (Bug Database Search is even more bogus - in general it doesn't find any bugs by ID). Should I continue to ignore these responses and mark those bugs internally as 'gets probably never fixed'? Regards. jel. -- Otto-von-Guericke University http://www.cs.uni-magdeburg.de/ Department of Computer Science Geb. 29 R 027, Universitaetsplatz 2 39106 Magdeburg, Germany Tel: +49 391 67 12768 ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] cutting up a SSD for read/log use...
Richard Elling wrote: Erik Trimble wrote: I just looked at pricing for the higher-end MLC devices, and it looks like I'm better off getting a single drive of 2X capacity than two with X capacity. Leaving aside the issue that by using 2 drives I get 2 x 3.0Gbps SATA performance instead of 1 x 3.0Gbps, are there problems with using two slices instead of whole-drives? That is, one slice for Read and the other for ZIL? My main concern is exactly how the on-drive cache would be used in a two-slices configuration. In order to get decent performance, I really need the on-drive cache to be used properly. Is the on-disk cache volatile? For most SSDs I'm familiar with, the on-disk cache is non-volatile, so all of the rules pertaining to whole disks with volatile write buffers are nullified. -- richard I'm pretty sure it is volatile. It's a single DRAM chip. http://benchmarkreviews.com/index.php?option=com_contenttask=viewid=333Itemid=60limit=1limitstart=3 (this review is specific to the OCZ Summit, but as the parts are pretty much Samsung-standard, it's identical to several other brand's versions) Are you sure most other SSDs have nvram as cache? I'm looking around, and they _seem_ to be using standard DRAM, just like Hard drives... Or maybe the SLC-based ones use nvram, and the MLC-based ones dram... -- Erik Trimble Java System Support Mailstop: usca22-123 Phone: x17195 Santa Clara, CA ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] cutting up a SSD for read/log use...
James Lever wrote: Hi Erik, On 22/06/2009, at 1:15 PM, Erik Trimble wrote: I just looked at pricing for the higher-end MLC devices, and it looks like I'm better off getting a single drive of 2X capacity than two with X capacity. Leaving aside the issue that by using 2 drives I get 2 x 3.0Gbps SATA performance instead of 1 x 3.0Gbps, are there problems with using two slices instead of whole-drives? That is, one slice for Read and the other for ZIL? The benefit you will get using 2 drives instead of 1 will be doubling your IOPS which will improve your overall performance, especially when using those drives as ZILs. Are you planning on using these drives as primary data storage and ZIL for the same volumes or as primary storage for (say) your rpool and ZIL for a data pool on spinning metal? cheers, James ZIL and Read cache for a data pool of HDs. i.e. zpool create tank mirror c1t0d0 c1t1d0 [...] log c2t0d0 cache c3t0d0 or zpool create tank mirror c1t0d0 c1t1d0 [...] log c2t0d0s0 cache c2t0d0s1 -- Erik Trimble Java System Support Mailstop: usca22-123 Phone: x17195 Santa Clara, CA ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Best controller card for 8 SATA drives ?
On Mon, Jun 22 at 12:05, Andre van Eyssen wrote: I'll add another vote for the LSI products. I have a four port PCI-X card in my V880, and the performance is good and the product is well behaved. The only caveats: 1. Make sure you upgrade the firmware ASAP 2. You may need to use lsiutil to fiddle the target mappings We bought a Dell T610 as a fileserver, and it comes with an LSI 1068E based board (PERC6/i SAS). Worked out of the box, no special drivers or anything to install, everything autodetected just fine. Hotplug works great too, I've yanked drives (Came with WD RE3 SASA devices) while the box was under load without issues, took ~5 seconds to timeout the device and give me full interactivity at the console. They then show right back up when hot plugged back in, and I can resilver without problems. --eric -- Eric D. Mudama edmud...@mail.bounceswoosh.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] PicoLCD Was: Best controller card for 8 SATA drives ?
I hesitate to post this question here, since the relation to ZFS is tenuous at best (zfs to sata controller to LCD panel). But maybe someone has already been down this path before me. Looking at building a RAID, with osol and zfs, I naturally want a front-panel. I was looking at something like; http://www.mini-box.com/picoLCD-256x64-Sideshow-CDROM-Bay Since it appears to come with OpenSource drivers. Based on lcd4linux, which I can compile with marginal massaging. Has anyone run this successfully with osol? It appears to handle mrtg directly, so I should be able to graph a whole load of ZFS data. Has someone already been down this road too? -- Jorgen Lundman | lund...@lundman.net Unix Administrator | +81 (0)3 -5456-2687 ext 1017 (work) Shibuya-ku, Tokyo| +81 (0)90-5578-8500 (cell) Japan| +81 (0)3 -3375-1767 (home) ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Why Oracle process open(2)/ioctl(2) /dev/dtrace/helper?
Hi all, Under what circumstance would an exiting process open(2) and ioctl(2) on /dev/dtrace/helper? I have an issue here where CU was complaining that their Oracle processes were taking a long time in ioctl(2) on /dev/dtrace/helper during shutdown. Removing group/world readable bit on /dev/dtrace/helper avoided the long shutdown time. So, the puzzling question is why /dev/dtrace/helper was being ioctl(2)ed on? I doubt this is a binary with USDT. truss(1M) shown the following: Base time stamp: 1244728105.2977 [ Thu Jun 11 22:48:25 KST 2009 ] 16003/1:psargs: oraclePOM1 (LOCAL=NO) 16003/1:read(31, 0x1069F3796, 2064) (sleeping...) ... 16003/1:1468.91560.0005 open(/dev/dtrace/helper, O_RDWR) = 8 16003/1:1477.37738.4617 ioctl(8, (('d'24)|('t'16)|('h'8)|2), 0x) = 0 16003/1:1477.37810.0008 Received signal #14, SIGALRM [caught] 16003/1:1477.37830.0002 lwp_sigmask(SIG_SETMASK, 0x9FBEF057, 0xFFF7) = 0xFFBFFEFF [0x] 1277 */ 1278 #define DTRACEHIOC (('d' 24) | ('t' 16) | ('h' 8)) 1279 #define DTRACEHIOC_ADD (DTRACEHIOC | 1)/* add helper */ 1280 #define DTRACEHIOC_REMOVE (DTRACEHIOC | 2)/* remove helper */ ustack() from dtrace(1M) shown this was from within the exit hanlder in exit(2). open(/dev/dtrace/helper) libc.so.1`open libCrun.so.1`0x7a50aed8 libCrun.so.1`0x7a50b0f4 ld.so.1`call_fini+0xd0 ld.so.1`atexit_fini+0x80 libc.so.1`_exithandle+0x48 libc.so.1`exit+0x4 oracle`_start+0x184 It would be great if anyone can help extend my understanding of dtrace. Thanks. Regards Sang Thong - 1788: oraclePOM1 (LOCAL=NO) data model = _LP64 flags = ORPHAN|MSACCT|MSFORK /1:flags = ASLEEP read(0x11,0x106346b16,0x810) 1788: oracleEOM1 (LOCAL=NO) /lib/sparcv9/libumem.so.1 /oraom/app/DBMS/lib/libskgxp10.so /oraom/app/DBMS/lib/libhasgen10.so /opt/ORCLcluster/lib/libskgxn2.so /oraom/app/DBMS/lib/libocr10.so /oraom/app/DBMS/lib/libocrb10.so /oraom/app/DBMS/lib/libocrutl10.so /oraom/app/DBMS/lib/libjox10.so /oraom/app/DBMS/lib/libclsra10.so /oraom/app/DBMS/lib/libdbcfg10.so /oraom/app/DBMS/lib/libnnz10.so /lib/sparcv9/libkstat.so.1 /lib/sparcv9/libnsl.so.1 /lib/sparcv9/libsocket.so.1 /lib/sparcv9/libgen.so.1 /lib/sparcv9/libdl.so.1 /usr/lib/sparcv9/libsched.so.1 /lib/sparcv9/libc.so.1 /lib/sparcv9/libaio.so.1 /lib/sparcv9/librt.so.1 /lib/sparcv9/libm.so.2 /opt/SUNWcluster/lib/sparcv9/libudlm.so /usr/cluster/lib/sparcv9/libhaops.so.1 /lib/sparcv9/libmd.so.1 /usr/cluster/lib/sparcv9/libscha.so.1 /usr/cluster/lib/sparcv9/libsecurity.so.1 /usr/cluster/lib/sparcv9/libclos.so.1 /lib/sparcv9/libdoor.so.1 /usr/lib/sparcv9/libzonecfg.so.1 /lib/sparcv9/libsecdb.so.1 /usr/lib/sparcv9/libCstd.so.1 /usr/lib/sparcv9/libCrun.so.1 /lib/sparcv9/libuuid.so.1 /lib/sparcv9/libnvpair.so.1 /lib/sparcv9/libsysevent.so.1 /lib/sparcv9/libsec.so.1 /usr/lib/sparcv9/libbrand.so.1 /usr/lib/sparcv9/libpool.so.1 /lib/sparcv9/libscf.so.1 /lib/sparcv9/libproc.so.1 /lib/sparcv9/libuutil.so.1 /usr/lib/sparcv9/libxml2.so.2 /lib/sparcv9/libcmd.so.1 /lib/sparcv9/libavl.so.1 /usr/lib/sparcv9/libexacct.so.1 /lib/sparcv9/librtld_db.so.1 /lib/sparcv9/libelf.so.1 /lib/sparcv9/libctf.so.1 /lib/sparcv9/libpthread.so.1 /usr/lib/sparcv9/libz.so.1 /platform/sun4u-us3/lib/sparcv9/libc_psr.so.1 Mon, 22 Jun 2009 09:06:09 KST open(/dev/dtrace/helper) libc.so.1`open libCrun.so.1`0x7a50aed8 libCrun.so.1`0x7a50b0f4 ld.so.1`call_fini+0xd0 ld.so.1`atexit_fini+0x80 libc.so.1`_exithandle+0x48 libc.so.1`exit+0x4 oracle`_start+0x184 *** ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Sorry, wrong alias Re: Why Oracle process open(2)/ioctl(2) /dev/dtrace/helper?
My apologies... sent to wrong alias. Sang-Thong Chang wrote: Hi all, Under what circumstance would an exiting process open(2) and ioctl(2) on /dev/dtrace/helper? I have an issue here where CU was complaining that their Oracle processes were taking a long time in ioctl(2) on /dev/dtrace/helper during shutdown. Removing group/world readable bit on /dev/dtrace/helper avoided the long shutdown time. So, the puzzling question is why /dev/dtrace/helper was being ioctl(2)ed on? I doubt this is a binary with USDT. truss(1M) shown the following: Base time stamp: 1244728105.2977 [ Thu Jun 11 22:48:25 KST 2009 ] 16003/1:psargs: oraclePOM1 (LOCAL=NO) 16003/1:read(31, 0x1069F3796, 2064) (sleeping...) ... 16003/1:1468.91560.0005 open(/dev/dtrace/helper, O_RDWR) = 8 16003/1:1477.37738.4617 ioctl(8, (('d'24)|('t'16)|('h'8)|2), 0x) = 0 16003/1:1477.37810.0008 Received signal #14, SIGALRM [caught] 16003/1:1477.37830.0002 lwp_sigmask(SIG_SETMASK, 0x9FBEF057, 0xFFF7) = 0xFFBFFEFF [0x] 1277 */ 1278 #define DTRACEHIOC (('d' 24) | ('t' 16) | ('h' 8)) 1279 #define DTRACEHIOC_ADD (DTRACEHIOC | 1)/* add helper */ 1280 #define DTRACEHIOC_REMOVE (DTRACEHIOC | 2)/* remove helper */ ustack() from dtrace(1M) shown this was from within the exit hanlder in exit(2). open(/dev/dtrace/helper) libc.so.1`open libCrun.so.1`0x7a50aed8 libCrun.so.1`0x7a50b0f4 ld.so.1`call_fini+0xd0 ld.so.1`atexit_fini+0x80 libc.so.1`_exithandle+0x48 libc.so.1`exit+0x4 oracle`_start+0x184 It would be great if anyone can help extend my understanding of dtrace. Thanks. Regards Sang Thong - 1788: oraclePOM1 (LOCAL=NO) data model = _LP64 flags = ORPHAN|MSACCT|MSFORK /1:flags = ASLEEP read(0x11,0x106346b16,0x810) 1788: oracleEOM1 (LOCAL=NO) /lib/sparcv9/libumem.so.1 /oraom/app/DBMS/lib/libskgxp10.so /oraom/app/DBMS/lib/libhasgen10.so /opt/ORCLcluster/lib/libskgxn2.so /oraom/app/DBMS/lib/libocr10.so /oraom/app/DBMS/lib/libocrb10.so /oraom/app/DBMS/lib/libocrutl10.so /oraom/app/DBMS/lib/libjox10.so /oraom/app/DBMS/lib/libclsra10.so /oraom/app/DBMS/lib/libdbcfg10.so /oraom/app/DBMS/lib/libnnz10.so /lib/sparcv9/libkstat.so.1 /lib/sparcv9/libnsl.so.1 /lib/sparcv9/libsocket.so.1 /lib/sparcv9/libgen.so.1 /lib/sparcv9/libdl.so.1 /usr/lib/sparcv9/libsched.so.1 /lib/sparcv9/libc.so.1 /lib/sparcv9/libaio.so.1 /lib/sparcv9/librt.so.1 /lib/sparcv9/libm.so.2 /opt/SUNWcluster/lib/sparcv9/libudlm.so /usr/cluster/lib/sparcv9/libhaops.so.1 /lib/sparcv9/libmd.so.1 /usr/cluster/lib/sparcv9/libscha.so.1 /usr/cluster/lib/sparcv9/libsecurity.so.1 /usr/cluster/lib/sparcv9/libclos.so.1 /lib/sparcv9/libdoor.so.1 /usr/lib/sparcv9/libzonecfg.so.1 /lib/sparcv9/libsecdb.so.1 /usr/lib/sparcv9/libCstd.so.1 /usr/lib/sparcv9/libCrun.so.1 /lib/sparcv9/libuuid.so.1 /lib/sparcv9/libnvpair.so.1 /lib/sparcv9/libsysevent.so.1 /lib/sparcv9/libsec.so.1 /usr/lib/sparcv9/libbrand.so.1 /usr/lib/sparcv9/libpool.so.1 /lib/sparcv9/libscf.so.1 /lib/sparcv9/libproc.so.1 /lib/sparcv9/libuutil.so.1 /usr/lib/sparcv9/libxml2.so.2 /lib/sparcv9/libcmd.so.1 /lib/sparcv9/libavl.so.1 /usr/lib/sparcv9/libexacct.so.1 /lib/sparcv9/librtld_db.so.1 /lib/sparcv9/libelf.so.1 /lib/sparcv9/libctf.so.1 /lib/sparcv9/libpthread.so.1 /usr/lib/sparcv9/libz.so.1 /platform/sun4u-us3/lib/sparcv9/libc_psr.so.1 Mon, 22 Jun 2009 09:06:09 KST open(/dev/dtrace/helper) libc.so.1`open libCrun.so.1`0x7a50aed8 libCrun.so.1`0x7a50b0f4 ld.so.1`call_fini+0xd0 ld.so.1`atexit_fini+0x80 libc.so.1`_exithandle+0x48 libc.so.1`exit+0x4 oracle`_start+0x184 *** ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Speeding up resilver on x4500
On Mon, Jun 22, 2009 at 4:24 PM, Stuart Anderson ander...@ligo.caltech.eduwrote: However, it is a bit disconcerting to have to run with reduced data protection for an entire week. While I am certainly not going back to UFS, it seems like it should be at least theoretically possible to do this several orders of magnitude faster, e.g., what if every block on the replacement disk had its RAIDZ2 data recomputed from the degraded Maybe this is also saying - that for large disk sets a single RAIDZ2 provides a false sense of security. Nicholas ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss