Re: [zfs-discuss] marvell88sx2 driver build126
On Mon, Nov 9, 2009 at 2:51 PM, Cindy Swearingen cindy.swearin...@sun.comwrote: Hi, I can't find any bug-related issues with marvell88sx2 in b126. I looked over Dave Hollister's shoulder while he searched for marvell in his webrevs of this putback and nothing came up: driver change with build 126? not for the SATA framework, but for HBAs there is: http://hub.opensolaris.org/bin/view/Community+Group+on/2009093001 I will find a thumper, load build 125, create a raidz pool, and upgrade to b126. I'll also send the error messages that Tim provided to someone who works in the driver group. Thanks, Cindy I tried the build 125 driver and it didn't make a difference. The odd part I've just noticed is that it's port 4 on both cards that have been giving me issues. I guess it's possible it's just a coincidence/bad luck. I've grabbed the b125 ISO from genunix and am going to try booting off the livecd to see if it produces different results. --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] marvell88sx2 driver build126
Does this mean that there are no driver changes in marvell88sx2, between b125 and b126? If no driver changes, then it means that we both had extreme unluck with our drives, because we both had checksum errors? And my discs were brand new. How probable is this? Something is weird here. What is your opinion on this? Should we agree that there was a hardware error, and it was just a coincidence? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] ZFS and oracle on SAN disks
I believe the best practice is to use seperate disks/zpool for oracle database files as the record size needs to be set the same as the db block size - when using a jbod or internal disks. If the server is using a large SAN LUN can anybody see any issues if there is only one zpool and the dataset is set to have the record size set rather than at the zpool level? Thanks Ian -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] FreeNAS 0.7 with zfs out
Apparently went live on 6th November. This isn't FreeBSD 8.x zfs, but at least raidz2 is there. http://www.freenas.org/ FreeNAS 0.7 (Khasadar) Sunday, 21 June 2009 Majors changes: * Add ability to configure the login shell for a user. * Upgrade Samba to 3.0.37. * Upgrade transmission to 1.72. * Local users must join the group 'ftp' to be able to login via FTP if 'Local users only' in 'Services|FTP' is enabled. * Upgrade lighttpd to 1.4.23. * Add a user portal. This allows a local user to login and change it's password. The user must have access permissions to login to the user portal. This can be configured in the user configuration WebGUI. Please note that the administrator needs to apply changes done by the local users because there have to be restarted several services (which is not allowed to be done by a user). * Upgrade ProFTPD to 1.3.2a. * Upgrade iSCSI initiator to 2.2.3. * Upgrade fusefs-ntfs/ntfs-3g to 2009.4.4. * Announce AFP shares using Bonjour/ZeroConf (FR 2839592). Thanks to Morton Jonuschat. * Add AFP FP_SYNCFORK command support (FR 2836955). Thanks to Morton Jonuschat for the patch. * Upgrade e2fsprogs to 1.41.8. * Add Adaptec AACRAID 32/64-bit driver to v5.2.0 Build 17517. * Upgrade inadyn-mt to 02.14.10. * Upgrade fuppes to SVN-0.640. Minors changes: * Set transmission umask to 0002 per default. This can be customize via the 'Services|BitTorrent' WebGUI or the rc.conf variable 'transmission_umask' (FR 2813791). * Add ixgbe driver to i386 kernel. * Add ixgb driver to AMD64 kernel (BR 2813759). * Add support for Blowfish 448 bits encryption (FR 2816028). * Add configuration option in 'Services|BitTorrent' to enable/disable usage of distributed hash table (DHT). * Add /usr/bin/getopt command (FR 2824548). * Add extra options for S.M.A.R.T. in 'Disks|Management|Disk|Edit' (FR 2824730). * Add RAID1 balance algorithm 'prefer' (FR 2833989). * Add latvian language support. Thanks to the translators. * Update Quixplorer russian translation (BR 2841900). Thanks to Alexey Sannikov. * Add 'Max. login attempts' to 'Services|FTP' (FR 2844193). * Get AFP dbd cnid scheme working (BR 2844900). * Set 'dir-listing.encoding = utf-8' for the webserver directory listing (FR 2872624). * Display volume serial number in 'Disks|Management' (FR 2881880). * Now it is possible to configure iSCSI-Targets for export: removable media (static dynamic size), pass-through devices. Thanks to Vasily Chalykh. Bug fixes: * Prohibit user 'transmission' to login via FTP. * ZPool disk space usage isn't displayed correctly (BR 2810584). * Improved Unison WebGUI to be able to configure ZFS shares as working directory (BR 2795084). * Synchronizing ZFS configuration fails (BR 2814324). * Restrict bittorrent administrative WebGUI port to [1024,65535] (BR 2835342). * The 'Unmount disk/partition' checkbox on 'Disks|Mount Point|Fsck' was ignored (BR 2860297). FreeNAS 0.7RC1 (Sardaukar) Sunday, 21 June 2009 Majors changes: * Upgrade to FreeBSD 7.2. * Include ZFS support. Thanks to Nelson Silva for doing core coding and Falk Menzel for testing and giving some tipps and ideas. * Upgrade iSCSI initiator to 2.1.1. * Replace iSCSI target by istgt. Thanks to Daisuke Aoyama for the WebGUI adaption. Please note, if you have used devices with the previous iSCSI target software you have to recreate your target. * Add WOL patch for nVidia(nfe(4)) and 3Com(xl(4)). Thanks to Tobias Reber. * Upgrade mt-daapd/firefly to svn-1696. * Refactor 'Diagnostics|Log' WebGUI. * Add kernel patches to get ARTiGO A2000 hardware working. Thanks to David Davis for the patches. * Respect the modified log file location (via rc.conf for syslog, fuppes, mt-daapd, ...) in the WebGUI (FR 2778803/2791772). * Upgrade transmission to 1.61. Add 'Watch directory' and 'Extra options' fields to 'Services|BitTorrent' WebGUI. * Add entry 'FTP - Ban module' to the list of log files in 'Diagnostics|Log' if the module is enabled (FR 2797652). * Add 'iperf', a tool to measure maximum TCP and UDP bandwidth (FR 2785038). * Add 'bsnmp-ucd' module that implements parts of UCD-SNMP-MIB. * Add SNMP client tools: bsnmpget, bsnmpset and bsnmpwalk * Add 'Auxiliary parameters' to 'Services|SNMP' that will be added to the end of the snmpd.config file. * Upgrade e2fsprogs to 1.41.5. * Upgrade rsync to 3.0.6. * Upgrade tftp-hpa to 0.49. Bug fixes: * Hardening WebGUI to prevent cross-site request forgery attacks (JPCERT/CC JVN#15267895). FreeNAS 0.69.2 (Muad'Dib) Monday, 20 April 2009 Majors changes: * Add another WOL patch. It is tested for nfe(4) und xl(4). Thanks to Tobias Reber. * Add switch in 'System|Advanced' WebGUI to enable the console screensaver (FR 2777301). * Upgrade Adaptec SCSI RAID administration tool to
Re: [zfs-discuss] Finding SATA cards for ZFS; was Lundman home NAS
On 01/09/09 08:26, James Andrewartha wrote: Jorgen Lundman wrote: The mv8 is a marvell based chipset, and it appears there are no Solaris drivers for it. There doesn't appear to be any movement from Sun or marvell to provide any either. Do you mean specifically Marvell 6480 drivers? I use both DAC-SATA-MV8 and AOC-SAT2-MV8, which use Marvell MV88SX and works very well in Solaris. (Package SUNWmv88sx). They're PCI-X SATA cards, the AOC-SASLP-MV8 is a PCIe SAS card and has no (Open)Solaris driver. Shame, I was just thinking that this was a nice looking card that could replace my AOC-SAT2-MV8s. Ah well... R. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] marvell88sx2 driver build126
Hi Orvar, Correct, I don't see any marvell8sx2 driver changes between b125-126. So far, only you and Tim are reporting these issues. Generally, we see bugs filed by the internal test teams if they see similar problems. I will try to reproduce the RAIDZ checksum errors separately from the marvell88sx2 issue. Thanks, Cindy On 11/10/09 02:25, Orvar Korvar wrote: Does this mean that there are no driver changes in marvell88sx2, between b125 and b126? If no driver changes, then it means that we both had extreme unluck with our drives, because we both had checksum errors? And my discs were brand new. How probable is this? Something is weird here. What is your opinion on this? Should we agree that there was a hardware error, and it was just a coincidence? ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Odd sparing problem
So, I currently have a pool with 12 disks raid-z2 (12+2). As you may have seen in the other thread, I've been having on and off issues with b126 randomly dropping drives. Well, I think after changing several cables, and doing about 20 reboots plugging one drive in at a time (I only booted to the marvell bios, not the whole way into the OS), I've gotten the marvell cards to settle down. The problem is, I'm now seeing this in a zpool output: pool: fserv state: ONLINE scrub: resilver completed after 0h0m with 0 errors on Tue Nov 10 09:15:12 2009 config: NAME STATE READ WRITE CKSUM fserv ONLINE 0 0 0 raidz2-0ONLINE 0 0 0 c8t0d0ONLINE 0 0 0 c8t1d0ONLINE 0 0 0 c8t2d0ONLINE 0 0 0 c8t3d0ONLINE 0 0 0 c8t4d0ONLINE 0 0 0 c8t5d0ONLINE 0 0 0 c7t0d0ONLINE 0 0 0 c7t1d0ONLINE 0 0 0 c7t2d0ONLINE 0 0 0 c7t3d0ONLINE 0 0 0 c7t4d0ONLINE 0 0 0 spare-11 ONLINE 0 0 5 c7t5d0 ONLINE 0 0 0 30K resilvered c7t6d0 ONLINE 0 0 0 spares c7t6d0 INUSE currently in use Anyone have any thoughts? I'm trying to figure out how to get c7t6d0 back to being a hotspare since c7t5d0 is installed, there, and happy. It's almost as if it's using both disks for spare-11 right now. --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ..and now ZFS send dedupe
James C. McPherson wrote, On 09-11-09 04:40 PM: Roman Naumenko wrote: Interesting stuff. By the way, is there a place to watch lated news like this on zfs/opensolaris? rss maybe? You could subscribe to onnv-not...@opensolaris.org... James C. McPherson -- Senior Kernel Software Engineer, Solaris Sun Microsystems http://blogs.sun.com/jmcp http://www.jmcp.homeunix.com/blog Thanks, James. What is the subscription process? Just to send email? -- Roman Naumenko ro...@frontline.ca ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] RAID-Z and virtualization
Toby Thain wrote: On 8-Nov-09, at 12:20 PM, Joe Auty wrote: Tim Cook wrote: On Sun, Nov 8, 2009 at 2:03 AM, besson3c j...@netmusician.org mailto:j...@netmusician.org wrote: ... Why not just convert the VM's to run in virtualbox and run Solaris directly on the hardware? That's another possibility, but it depends on how Virtualbox stacks up against VMWare Server. At this point a lot of planning would be necessary to switch to something else, although this is possibility. How would Virtualbox stack up against VMWare Server? Last I checked it doesn't have a remote console of any sort, which would be a deal breaker. Can I disable allocating virtual memory to Virtualbox VMs? Can I get my VMs to auto boot in a specific order at runlevel 3? Can I control my VMs via the command line? Yes you certainly can. Works well, even for GUI based guests, as there is vm-level VRDP (VNC/Remote Desktop) access as well as whatever remote access the guest provides. I thought Virtualbox was GUI only, designed for Desktop use primarily? Not at all. Read up on VBoxHeadless. I take it that Virtualbox, being Qemu/KVM based will support 64 bit versions of FreeBSD guests, unlike Xen based solutions? --Toby This switch will only make sense if all of this points to a net positive. --Tim -- Joe Auty NetMusician: web publishing software for musicians http://www.netmusician.org j...@netmusician.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org mailto:zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Joe Auty NetMusician: web publishing software for musicians http://www.netmusician.org j...@netmusician.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS and oracle on SAN disks
On Nov 10, 2009, at 5:32 AM, Ian Garbutt wrote: I believe the best practice is to use seperate disks/zpool for oracle database files as the record size needs to be set the same as the db block size - when using a jbod or internal disks. recordsize is not a pool property, it is a dataset (zvol or file system) property. You can have different recordsize or volblocksize property settings for each dataset in the pool. You can also have different cache policies, which will also be useful. Roch reports a 40% performance improvement by adjusting the logbias property. http://blogs.sun.com/roch/entry/synchronous_write_bias_property However, in that case, a large SAN LUN was a hybrid storage pool. If you are using a traditional RAID array, then the nonvolatile write cache can also improve performance. -- richard If the server is using a large SAN LUN can anybody see any issues if there is only one zpool and the dataset is set to have the record size set rather than at the zpool level? Thanks Ian -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] marvell88sx2 driver build126
On Nov 10, 2009, at 1:25 AM, Orvar Korvar wrote: Does this mean that there are no driver changes in marvell88sx2, between b125 and b126? If no driver changes, then it means that we both had extreme unluck with our drives, because we both had checksum errors? And my discs were brand new. There are other drivers in the software stack that may have changed. -- richard How probable is this? Something is weird here. What is your opinion on this? Should we agree that there was a hardware error, and it was just a coincidence? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS + fsck
Hi, *everybody* is interested in the flag days page. Including me. Asking me to raise the priority is not helpful. From my perspective, it's a surprise that 'everybody' is interested, as I'm not seeing a lot of people complaining that the flag day page is not updating. Only a couple of people on this list, and one of those is me! Perhaps I'm looking in the wrong places. I used this page frequently, too. But now i'm just using the twitter account feeded by onnv-notify . You can look to it at http://twitter.com/codenews Regards Joerg ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs inotify?
On Nov 10, 2009, at 10:23 AM, Andrew Daugherity wrote: For example: rsync -avn --delete-before /export/ims/.zfs/snapshot/zfs-auto- snap_daily-2009-11-09-1900/ /export/ims/.zfs/snapshot/zfs-auto- snap_daily-2009-11-08-1900/ [...] If you cared to see changes within files (I don't), toss the -W flag on there then and it should run even faster, methinks, as it won't compare file contents at all. -Jeremy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs inotify?
Thanks for info, although the audit system seems a lot more complex than what I need. Would still be nice if they fixed bart to work on large filesystems, though. Turns out the solution was right under my nose -- rsync in dry-run mode works quite well as a snapshot diff tool. I'll share this with the list, in case it helps anyone else. For example: rsync -avn --delete-before /export/ims/.zfs/snapshot/zfs-auto-snap_daily-2009-11-09-1900/ /export/ims/.zfs/snapshot/zfs-auto-snap_daily-2009-11-08-1900/ This makes a list of what files have been changed, added, or deleted between these two snapshots, and runs in about 10 seconds. If you cared to see changes within files (I don't), it would be trivial to add a loop along the lines of '$rsync_cmd | while read file; do diff $snap1/$file $snap2/file; done'. Note the trailing slashes (otherwise rsync works one directory higher and considers the snapshot directory name, which we don't want) and that the newer snapshot is the source and the older snapshot, the destination. I'm [ab]using rsync to have it tell me exactly how it would make the destination be a replica of the source. FWIW, I'm using rsync 3.0.6 from opencsw. Older rsync should work fine but may take longer to run. -Andrew Richard Elling richard.ell...@gmail.com 11/9/2009 7:33 PM Seems to me that you really want auditing. You can configure the audit system to only record the events you are interested in. http://docs.sun.com/app/docs/doc/816-4557/auditov-1?l=ena=view -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] PSARC recover files?
Say I end up with a handful of unrecoverable bad blocks that just so happen to be referenced by ALL of my snapshots (in some file that's been around forever). Say I don't care about the file or two in which the bad blocks exist. Is there any way to purge those blocks from the pool (and all snapshots) without having to restore the whole pool from backup? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] PSARC recover files?
On Tue, Nov 10, 2009 at 2:40 PM, BJ Quinn bjqu...@seidal.com wrote: Say I end up with a handful of unrecoverable bad blocks that just so happen to be referenced by ALL of my snapshots (in some file that's been around forever). Say I don't care about the file or two in which the bad blocks exist. Is there any way to purge those blocks from the pool (and all snapshots) without having to restore the whole pool from backup? No. The whole point of a snapshot is to keep a consistent on-disk state from a certain point in time. I'm not entirely sure how you managed to corrupt blocks that are part of an existing snapshot though, as they'd be read-only. The only way that should even be able to happen is if you took a snapshot after the blocks were already corrupted. Any new writes would be allocated from new blocks. --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] PSARC recover files?
On Tue, Nov 10, 2009 at 03:04:24PM -0600, Tim Cook wrote: No. The whole point of a snapshot is to keep a consistent on-disk state from a certain point in time. I'm not entirely sure how you managed to corrupt blocks that are part of an existing snapshot though, as they'd be read-only. Physical corruption of the media Something outside of ZFS diddling bits on storage The only way that should even be able to happen is if you took a snapshot after the blocks were already corrupted. Any new writes would be allocated from new blocks. It can be corrupted while it sits on disk. Since it's read-only, you can't force it to allocate anything and clean itself up. -- Darren ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] PSARC recover files?
On Tue, Nov 10, 2009 at 3:19 PM, A Darren Dunham ddun...@taos.com wrote: On Tue, Nov 10, 2009 at 03:04:24PM -0600, Tim Cook wrote: No. The whole point of a snapshot is to keep a consistent on-disk state from a certain point in time. I'm not entirely sure how you managed to corrupt blocks that are part of an existing snapshot though, as they'd be read-only. Physical corruption of the media Something outside of ZFS diddling bits on storage The only way that should even be able to happen is if you took a snapshot after the blocks were already corrupted. Any new writes would be allocated from new blocks. It can be corrupted while it sits on disk. Since it's read-only, you can't force it to allocate anything and clean itself up. You're telling me a scrub won't actively clean up corruption in snapshots? That sounds absolutely absurd to me. --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] PSARC recover files?
On Tue, Nov 10, 2009 at 03:33:22PM -0600, Tim Cook wrote: You're telling me a scrub won't actively clean up corruption in snapshots? That sounds absolutely absurd to me. Depends on how much redundancy you have in your pool. If you have no mirrors, no RAID-Z, and no ditto blocks for data, well, you have no redundancy, and ZFS won't be able to recover affected files. Nico -- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Rebooting while Scrubing in snv_126
Greetings folks, Something funny happened to my amd64 box last night. I shut it down while a a scrub was running on rpool. This was not a fast reboot or anything like that. Since then, the system does not come up any more. I still can boot in single user mode but /sbin/zfs mount -va hangs while displaying: Reading ZFS config: * I went a bit further in single user mode: truss -f ksh /lib/svc/method/fs-local's last sign of life is a memcntl, shortly after a ZFS_IOC_POOL_CONFIGS successful ioctl. Any word of wisdom here? I really hope this recoverable since this was my primary system (no serial console though, so I can't really capture the essence of the drama :) ) I think rpool was version 19 but I'm not 100% positive here. Regards. Francois ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zle compression ?
by some posting on zfs-fuse mailinglist, i came across zle compression which seems to be part of the dedupe-commit some days ago: http://hg.genunix.org/onnv-gate.hg/diff/e2081f502306/usr/src/uts/common/fs/zfs/zle.c --snipp 31 + * Zero-length encoding. This is a fast and simple algorithm to eliminate 32 + * runs of zeroes. Each chunk of compressed data begins with a length byte, b. 33 + * If b n (where n is the compression parameter) then the next b + 1 bytes 34 + * are literal values. If b = n then the next (256 - b + 1) bytes are zero. --snipp i`m curious - what does that mean? does zfs have another compression scheme named zle now ? if yes, why ? wasn´t zero-length encoding already there and just a builtin feature ? maybe that builtin has now become an option ? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] This is the scrub that never ends...
On Fri, 2009-09-11 at 13:51 -0400, Will Murnane wrote: On Thu, Sep 10, 2009 at 13:06, Will Murnane will.murn...@gmail.com wrote: On Wed, Sep 9, 2009 at 21:29, Bill Sommerfeld sommerf...@sun.com wrote: Any suggestions? Let it run for another day. I'll let it keep running as long as it wants this time. scrub: scrub completed after 42h32m with 0 errors on Thu Sep 10 17:20:19 2009 And the people rejoiced. So I guess the issue is more scrubs may report ETA very inaccurately than scrubs never finish. Thanks for the suggestions and support. One of my pools routinely does this -- the scrub gets to 100% after about 50 hours but keeps going for another day or more after that. It turns out that zpool reports number of blocks visited vs number of blocks allocated, but clamps the ratio at 100%. If there is substantial turnover in the pool, it appears you may end up needing to visit more blocks than are actually allocated at any one point in time. I made a modified version of the zpool command and this is what it prints for me: ... scrub: scrub in progress for 74h25m, 119.90% done, 0h0m to go 5428197411840 blocks examined, 4527262118912 blocks allocated ... This is the (trivial) source change I made to see what's going on under the covers: diff -r 12fb4fb507d6 usr/src/cmd/zpool/zpool_main.c --- a/usr/src/cmd/zpool/zpool_main.cMon Oct 26 22:25:39 2009 -0700 +++ b/usr/src/cmd/zpool/zpool_main.cTue Nov 10 17:07:59 2009 -0500 @@ -2941,12 +2941,15 @@ if (examined == 0) examined = 1; - if (examined total) - total = examined; fraction_done = (double)examined / total; - minutes_left = (uint64_t)((now - start) * - (1 - fraction_done) / fraction_done / 60); + if (fraction_done 1) { + minutes_left = (uint64_t)((now - start) * + (1 - fraction_done) / fraction_done / 60); + } else { + minutes_left = 0; + } + minutes_taken = (uint64_t)((now - start) / 60); (void) printf(gettext(%s in progress for %lluh%um, %.2f%% done, @@ -2954,6 +2957,9 @@ scrub_type, (u_longlong_t)(minutes_taken / 60), (uint_t)(minutes_taken % 60), 100 * fraction_done, (u_longlong_t)(minutes_left / 60), (uint_t)(minutes_left % 60)); + (void) printf(gettext(\t %lld blocks examined, %lld blocks allocated\n), + examined, + total); } static void ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Odd sparing problem
Hi Tim, I'm not sure I understand this output completely, but have you tried detaching the spare? Cindy On 11/10/09 09:21, Tim Cook wrote: So, I currently have a pool with 12 disks raid-z2 (12+2). As you may have seen in the other thread, I've been having on and off issues with b126 randomly dropping drives. Well, I think after changing several cables, and doing about 20 reboots plugging one drive in at a time (I only booted to the marvell bios, not the whole way into the OS), I've gotten the marvell cards to settle down. The problem is, I'm now seeing this in a zpool output: pool: fserv state: ONLINE scrub: resilver completed after 0h0m with 0 errors on Tue Nov 10 09:15:12 2009 config: NAME STATE READ WRITE CKSUM fserv ONLINE 0 0 0 raidz2-0ONLINE 0 0 0 c8t0d0ONLINE 0 0 0 c8t1d0ONLINE 0 0 0 c8t2d0ONLINE 0 0 0 c8t3d0ONLINE 0 0 0 c8t4d0ONLINE 0 0 0 c8t5d0ONLINE 0 0 0 c7t0d0ONLINE 0 0 0 c7t1d0ONLINE 0 0 0 c7t2d0ONLINE 0 0 0 c7t3d0ONLINE 0 0 0 c7t4d0ONLINE 0 0 0 spare-11 ONLINE 0 0 5 c7t5d0 ONLINE 0 0 0 30K resilvered c7t6d0 ONLINE 0 0 0 spares c7t6d0 INUSE currently in use Anyone have any thoughts? I'm trying to figure out how to get c7t6d0 back to being a hotspare since c7t5d0 is installed, there, and happy. It's almost as if it's using both disks for spare-11 right now. --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Odd sparing problem
On Tue, Nov 10, 2009 at 4:38 PM, Cindy Swearingen cindy.swearin...@sun.comwrote: Hi Tim, I'm not sure I understand this output completely, but have you tried detaching the spare? Cindy Hey Cindy, Detaching did in fact solve the issue. During my previous issues when the spare kicked in, it actually automatically detached itself once I replaced the failed drive, so I didn't understand what was going on this time around. Thanks! --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] marvell88sx2 driver build126
On Tue, Nov 10, 2009 at 10:55 AM, Richard Elling richard.ell...@gmail.comwrote: On Nov 10, 2009, at 1:25 AM, Orvar Korvar wrote: Does this mean that there are no driver changes in marvell88sx2, between b125 and b126? If no driver changes, then it means that we both had extreme unluck with our drives, because we both had checksum errors? And my discs were brand new. There are other drivers in the software stack that may have changed. -- richard How probable is this? Something is weird here. What is your opinion on this? Should we agree that there was a hardware error, and it was just a coincidence? So... I do appear to have reached somewhat of a truce with the system and b126 at the moment. I'm now going through and replacing the last of my old maxtor 300GB drives with brand new hitachi 1TB drives. One thing I'm noticing is a lot of checksum errors being generated during the resilver. Is this normal? Furthermore, since I see no known data errors, is it safe to assume it's all being corrected, and I'm not losing any data? I still do have a separate copy of this data on a box at work that should be completely consistent... but I will need to re-purpose that storage soon, and will be without a known good backup for a while (I know, I know). I'd rather do a fresh zfs send/receive than find out 6 months from now I lost something. pool: fserv state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scrub: resilver in progress for 0h8m, 0.89% done, 15h14m to go config: NAMESTATE READ WRITE CKSUM fserv DEGRADED 0 0 0 raidz2-0 DEGRADED 0 0 0 c8t0d0 ONLINE 0 0 0 c8t1d0 ONLINE 0 0 0 c8t2d0 ONLINE 0 0 0 c8t3d0 ONLINE 0 0 0 c8t4d0 ONLINE 0 0 0 c8t5d0 ONLINE 0 0 0 c7t0d0 ONLINE 0 0 0 c7t1d0 ONLINE 0 0 0 c7t2d0 ONLINE 0 0 0 replacing-9 DEGRADED 0 0 161K 14274451003165180679 FAULTED 0 0 0 was /dev/dsk/c7t3d0s0/old c7t3d0ONLINE 0 0 0 2.05G resilvered c7t4d0 ONLINE 0 0 0 c7t5d0 ONLINE 0 0 0 spares c7t6d0AVAIL errors: No known data errors --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] PSARC recover files?
I believe it was physical corruption of the media. Strange thing is last time it happened to me it also managed to replicate the bad blocks over to my backup server replicated with SNDR... And yes, it IS read only, and a scrub will NOT actively clean up corruption in snapshots. It will DETECT corruption, but if it's unrecoverable, that's that. It's unrecoverable. If there's not enough redundancy in the pool, I'm ok with the data not being recoverable. But wouldn't there be a way to purge out the bad blocks if for example it was only in a single bad file out of millions of files, and I didn't care about the file in question? I don't want to recover the file, I want to have a working version of my pool+snapshots minus the tiny bit that was obviously corrupt. Barring another solution, I'd have to take the pool in question, delete the bad file, and delete ALL the snapshots. Then restore the old snapshots from backup to another pool, and copy over the current data from the pool that had a problem over to the new pool. I can get most of my snapshots back that way, with the best known current data sitting on top as the active data set. Problem is with hundreds of snapshots plus compression, zfs send/recv takes over 24 hours to restore a full backup like that to a new storage device. Last time this happened to me, I just had to say goodbye to all my snapshots and deal with it, all over a couple of kilobytes of temp files. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Fwd: [ilugb] Does ZFS support Hole Punching/Discard
I've been following the use of SSD with ZFS and HSPs for some time now, and I am working (in an architectural capacity) with one of our IT guys to set up our own ZFS HSP (using a J4200 connected to an X2270). The best practice seems to be to use an Intel X25-M for the L2ARC (Readzilla) and an Intel X25-E for the ZIL/SLOG (Logzilla). However, whilst being a BIG thing in the Windows 7 world - I have pretty much heard nothing about Intel's G2 devices and updated firmware when Intel's SSDs are used in a ZFS HSP. In particular, does ZFS use or support the TRIM command? Is it even relevant or useful in a hierarchical (vs. primary) storage context? Any comment would be appreciated. Some comment from the Fishworks guys in particular would be great! -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zfs eradication
Hi, I was discussing the common practice of disk eradication used by many firms for security. I was thinking this may be a useful feature of ZFS to have an option to eradicate data as its removed, meaning after the last reference/snapshot is done and a block is freed, then write the eradication patterns back to the removed blocks. By any chance, has this been discussed or considered before? Thanks, Brian ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Fwd: [ilugb] Does ZFS support Hole Punching/Discard
On Tue, Nov 10, 2009 at 6:51 PM, George Janczuk geor...@objectconsulting.com.au wrote: I've been following the use of SSD with ZFS and HSPs for some time now, and I am working (in an architectural capacity) with one of our IT guys to set up our own ZFS HSP (using a J4200 connected to an X2270). The best practice seems to be to use an Intel X25-M for the L2ARC (Readzilla) and an Intel X25-E for the ZIL/SLOG (Logzilla). However, whilst being a BIG thing in the Windows 7 world - I have pretty much heard nothing about Intel's G2 devices and updated firmware when Intel's SSDs are used in a ZFS HSP. In particular, does ZFS use or support the TRIM command? Is it even relevant or useful in a hierarchical (vs. primary) storage context? Any comment would be appreciated. Some comment from the Fishworks guys in particular would be great! My personal thought would be that it doesn't really make sense to even have it, at least for readzilla. In theory, you always want the SSD to be full, or nearly full, as it's a cache. The whole point of TRIM, from my understanding, is to speed up the drive by zeroing out unused blocks so they next time you try to write to them, they don't have to be cleared, then written to. When dealing with a cache, there shouldn't (again in theory) be any free blocks, a warmed cache should be full of data. Logzilla is kind of in the same boat, it should constantly be filling and emptying as new data comes in. I'd imagine the TRIM would just add unnecessary overhead. It could in theory help there by zeroing out blocks ahead of time before a new batch of writes come in if you have a period of little I/O. My thought is it would be far more work than it's worth, but I'll let the coders decide that one. --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs eradication
Typically this is called Sanitization and could be done as part of an evacuation of data from the disk in preparation for removal. You would want to specify the patterns to write and the number of passes. -- mark Brian Kolaci wrote: Hi, I was discussing the common practice of disk eradication used by many firms for security. I was thinking this may be a useful feature of ZFS to have an option to eradicate data as its removed, meaning after the last reference/snapshot is done and a block is freed, then write the eradication patterns back to the removed blocks. By any chance, has this been discussed or considered before? Thanks, Brian ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- http://www.sun.com * Mark A. Carlson * Sr. Architect *Systems Group* Phone x69559 / 303-223-6139 Email mark.carl...@sun.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Zpool hosed during testing
This didn't occur on a production server, but I thought I'd post this anyway because it might be interesting. I'm currently testing a ZFS NAS machine consisting of a Dell R710 with two Dell 5/E SAS HBAs. Right now I'm in the middle of torture testing the system, simulating drive failures, exporting the storage pool, rearranging the disks in different slots, and what have you. Up until now, everything has been going swimmingly. Here was my original zpool configuration: NAME STATE READ WRITE CKSUM tank ONLINE 0 0 0 raidz2 ONLINE 0 0 0 c1t1d0 ONLINE 0 0 0 c1t2d0 ONLINE 0 0 0 c1t3d0 ONLINE 0 0 0 c1t4d0 ONLINE 0 0 0 c1t5d0 ONLINE 0 0 0 c1t6d0 ONLINE 0 0 0 c1t7d0 ONLINE 0 0 0 c1t8d0 ONLINE 0 0 0 c1t9d0 ONLINE 0 0 0 c1t10d0 ONLINE 0 0 0 c1t11d0 ONLINE 0 0 0 c1t12d0 ONLINE 0 0 0 raidz2 ONLINE 0 0 0 c2t25d0 ONLINE 0 0 0 c2t26d0 ONLINE 0 0 0 c2t27d0 ONLINE 0 0 0 c2t28d0 ONLINE 0 0 0 c2t29d0 ONLINE 0 0 0 c2t30d0 ONLINE 0 0 0 c2t31d0 ONLINE 0 0 0 c2t32d0 ONLINE 0 0 0 c2t33d0 ONLINE 0 0 0 c2t34d0 ONLINE 0 0 0 c2t35d0 ONLINE 0 0 0 c2t36d0 ONLINE 0 0 0 I exported the tank zpool, and rearranged drives in the chassis and reimported it - I ended up with this: NAME STATE READ WRITE CKSUM tank ONLINE 0 0 0 raidz2 ONLINE 0 0 0 c2t31d0 ONLINE 0 0 0 c1t2d0 ONLINE 0 0 0 c1t3d0 ONLINE 0 0 0 c1t1d0 ONLINE 0 0 0 c1t12d0 ONLINE 0 0 0 c1t6d0 ONLINE 0 0 0 c1t7d0 ONLINE 0 0 0 c1t8d0 ONLINE 0 0 0 c1t9d0 ONLINE 0 0 0 c1t5d0 ONLINE 0 0 0 c1t11d0 ONLINE 0 0 0 c2t25d0 ONLINE 0 0 0 raidz2 ONLINE 0 0 0 c1t4d0 ONLINE 0 0 0 c2t26d0 ONLINE 0 0 0 c2t27d0 ONLINE 0 0 0 c2t28d0 ONLINE 0 0 0 c2t29d0 ONLINE 0 0 0 c2t30d0 ONLINE 0 0 0 c1t10d0 ONLINE 0 0 0 c2t32d0 ONLINE 0 0 0 c2t33d0 ONLINE 0 0 0 c2t34d0 ONLINE 0 0 0 c2t35d0 ONLINE 0 0 0 c2t48d0 ONLINE 0 0 0 Great. No problems. Next, I took c2t48d0 offline and then unconfigured it with cfgadm. # zpool offline tank c2t48d0 # cfgadm -c unconfigure c2::dsk/c2t48d0 I checked the status next. # zpool status tank pool: tank state: DEGRADED status: One or more devices has been taken offline by the administrator. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Online the device using 'zpool online' or replace the device with 'zpool replace'. scrub: none requested config: NAME STATE READ WRITE CKSUM tank DEGRADED 0 0 0 raidz2 ONLINE 0 0 0 c2t31d0 ONLINE 0 0 0 c1t2d0 ONLINE 0 0 0 c1t3d0 ONLINE 0 0 0 c1t1d0 ONLINE 0 0 0 c1t12d0 ONLINE 0 0 0 c1t6d0 ONLINE 0 0 0 c1t7d0 ONLINE 0 0 0 c1t8d0 ONLINE 0 0 0 c1t9d0 ONLINE 0 0 0 c1t5d0 ONLINE 0 0 0 c1t11d0 ONLINE 0 0 0 c2t25d0 ONLINE 0 0 0 raidz2 DEGRADED 0 0 0 c1t4d0 ONLINE 0 0 0 c2t26d0 ONLINE 0 0 0 c2t27d0 ONLINE 0 0 0 c2t28d0 ONLINE 0 0 0 c2t29d0 ONLINE 0 0 0 c2t30d0 ONLINE 0 0 0 c1t10d0 ONLINE 0 0 0 c2t32d0 ONLINE 0 0 0
Re: [zfs-discuss] zfs eradication
On Nov 10, 2009, at 20:55, Mark A. Carlson wrote: Typically this is called Sanitization and could be done as part of an evacuation of data from the disk in preparation for removal. You would want to specify the patterns to write and the number of passes. See also remanence: http://en.wikipedia.org/wiki/Data_remanence (S)ATA actually has a protocol command (secure erase) that will cause the disk to over write all of its sectors, and not be usable until its done. This doesn't exist in SCSI / SAS / FC as far as I know. Generally speaking one over write is sufficient to prevent data from being accessible, but various government standards specify anywhere between one and four passes: http://en.wikipedia.org/wiki/Data_erasure Degaussing or complete destruction is usually necessary for the top secret stuff. DBAN is a useful (open-source) utility that I tend to recommend for regular folk: http://www.dban.org/ While it could be useful, there are penalties in various jurisdictions for leaking data (especially with government-related stuff), so I'm not sure if Sun would want to potentially expose itself to inappropriate use that doesn't clean everything properly. With ZFS encryption coming up, it could be sufficient to have your data sets encrypted and then simply forget the key. The data is still technically there, but (theoretically) completely inaccessible. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zfs eradication
Excuse me for mentioning it but why not just use the format command? format(1M) analyze Run read, write, compare tests, and data purge. The data purge function implements the National Computer Security Center Guide to Understanding Data Remnance (NCSC-TG-025 version 2) Overwriting Algorithm. See NOTES. The NCSC-TG-025 algorithm for overwriting meets the DoD 5200.28-M (ADP Security Manual) Eraser Procedures specification. The NIST Guidelines for Media Sanitization (NIST SP 800-88) also reference this algorithm.. And if the disk is buggered (a very technical term). A great big hammer! Mark A. Carlson wrote: Typically this is called "Sanitization" and could be done as part of an evacuation of data from the disk in preparation for removal. You would want to specify the patterns to write and the number of passes. -- mark Brian Kolaci wrote: Hi, I was discussing the common practice of disk eradication used by many firms for security. I was thinking this may be a useful feature of ZFS to have an option to eradicate data as its removed, meaning after the last reference/snapshot is done and a block is freed, then write the eradication patterns back to the removed blocks. By any chance, has this been discussed or considered before? Thanks, Brian ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Mark A. Carlson Sr. Architect Systems Group Phone x69559 / 303-223-6139 Email mark.carl...@sun.com www.eagle.co.nz This email is confidential and may be legally privileged. If received in error please destroy and immediately notify us. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] CIFS crashes when accessed with Adobe Photoshop Elements 6.0 via Vista
upgrade to the latest dev release fixed the problem for me. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] marvell88sx2 driver build126
On Tue, Nov 10, 2009 at 5:15 PM, Tim Cook t...@cook.ms wrote: On Tue, Nov 10, 2009 at 10:55 AM, Richard Elling richard.ell...@gmail.com wrote: On Nov 10, 2009, at 1:25 AM, Orvar Korvar wrote: Does this mean that there are no driver changes in marvell88sx2, between b125 and b126? If no driver changes, then it means that we both had extreme unluck with our drives, because we both had checksum errors? And my discs were brand new. There are other drivers in the software stack that may have changed. -- richard How probable is this? Something is weird here. What is your opinion on this? Should we agree that there was a hardware error, and it was just a coincidence? So... I do appear to have reached somewhat of a truce with the system and b126 at the moment. I'm now going through and replacing the last of my old maxtor 300GB drives with brand new hitachi 1TB drives. One thing I'm noticing is a lot of checksum errors being generated during the resilver. Is this normal? Furthermore, since I see no known data errors, is it safe to assume it's all being corrected, and I'm not losing any data? I still do have a separate copy of this data on a box at work that should be completely consistent... but I will need to re-purpose that storage soon, and will be without a known good backup for a while (I know, I know). I'd rather do a fresh zfs send/receive than find out 6 months from now I lost something. pool: fserv state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scrub: resilver in progress for 0h8m, 0.89% done, 15h14m to go config: NAMESTATE READ WRITE CKSUM fserv DEGRADED 0 0 0 raidz2-0 DEGRADED 0 0 0 c8t0d0 ONLINE 0 0 0 c8t1d0 ONLINE 0 0 0 c8t2d0 ONLINE 0 0 0 c8t3d0 ONLINE 0 0 0 c8t4d0 ONLINE 0 0 0 c8t5d0 ONLINE 0 0 0 c7t0d0 ONLINE 0 0 0 c7t1d0 ONLINE 0 0 0 c7t2d0 ONLINE 0 0 0 replacing-9 DEGRADED 0 0 161K 14274451003165180679 FAULTED 0 0 0 was /dev/dsk/c7t3d0s0/old c7t3d0ONLINE 0 0 0 2.05G resilvered c7t4d0 ONLINE 0 0 0 c7t5d0 ONLINE 0 0 0 spares c7t6d0AVAIL errors: No known data errors --Tim Anyone? It's up to 7.35M checksum errors and it's rebuilding extremely slowly (as evidenced by the 10 hour time). The errors are only showing on the replacing-9 line, not the individual drive. pool: fserv state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scrub: resilver in progress for 6h56m, 39.61% done, 10h34m to go config: NAMESTATE READ WRITE CKSUM fserv DEGRADED 0 0 0 raidz2-0 DEGRADED 0 0 0 c8t0d0 ONLINE 0 0 0 c8t1d0 ONLINE 0 0 0 c8t2d0 ONLINE 0 0 0 c8t3d0 ONLINE 0 0 0 c8t4d0 ONLINE 0 0 0 c8t5d0 ONLINE 0 0 0 c7t0d0 ONLINE 0 0 0 c7t1d0 ONLINE 0 0 0 c7t2d0 ONLINE 0 0 0 replacing-9 DEGRADED 0 0 7.35M 14274451003165180679 FAULTED 0 0 0 was /dev/dsk/c7t3d0s0/old c7t3d0ONLINE 0 0 0 91.9G resilvered c7t4d0 ONLINE 0 0 0 c7t5d0 ONLINE 0 0 0 spares c7t6d0AVAIL errors: No known data errors --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss