[zfs-discuss] ZFS bug - should I be worried about this?
I found this today: http://blog.lastinfirstout.net/2010/06/sunoracle-finally-announces-zfs-data.html?utm_source=feedburnerutm_medium=feedutm_campaign=Feed%3A+LastInFirstOut+%28Last+In%2C+First+Out%29utm_content=FriendFeed+Bot How can I be sure my Solaris 10 systems are fine? Is latest OpenSolaris (134) safe? Thx Gabriele. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS root recovery SMI/EFI label weirdness
Thanks I don't know how I missed it. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS bug - should I be worried about this?
On 06/28/10 08:15 PM, Gabriele Bulfon wrote: I found this today: http://blog.lastinfirstout.net/2010/06/sunoracle-finally-announces-zfs-data.html?utm_source=feedburnerutm_medium=feedutm_campaign=Feed%3A+LastInFirstOut+%28Last+In%2C+First+Out%29utm_content=FriendFeed+Bot How can I be sure my Solaris 10 systems are fine? Is latest OpenSolaris (134) safe? Did you read the Sunsolve document? b134 is not vulnerable. -- Ian. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS on Ubuntu
All true, I just saw too many need ubuntu and zfs and thought to state the obvious in case the patch set for nexenta happen to differ enough to provide a working set. I've had nexenta succeed where opensolaris quarter releases failed and vice versa On Jun 27, 2010, at 9:54 PM, Erik Trimble erik.trim...@oracle.com wrote: On 6/27/2010 9:07 PM, Richard Elling wrote: On Jun 27, 2010, at 8:52 PM, Erik Trimble wrote: But that won't solve the OP's problem, which was that OpenSolaris doesn't support his hardware. Nexenta has the same hardware limitations as OpenSolaris. AFAICT, the OP's problem is with a keyboard. The vagaries of keyboards is well documented, but there is no silver bullet. Indeed, I have one box that seems to be more or less happy with PS-2 vs USB for every other OS or hypervisor. My advice, have one of each handy, just in case. -- richard Right. I was just pointing out the fallacy of thinking that Nexenta might work on hardware that OpenSolaris doesn't (or has problems with). -- Erik Trimble Java System Support Mailstop: usca22-123 Phone: x17195 Santa Clara, CA ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Kernel panic on zpool status -v (build 143)
I ran 'zpool scrub' and will report what happens once it's finished. (It will take pretty long.) The scrub finished successfully (with no errors) and 'zpool status -v' doesn't crash the kernel any more. Andrej smime.p7s Description: S/MIME Cryptographic Signature ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS bug - should I be worried about this?
Yes, I did read it. And what worries me is patches availability... -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS on Ubuntu
I think zfs on ubuntu currently is a rather bad idea. See test below with ubuntu Lucid 10.04 (amd64) r...@bigone:~# cat /proc/partitions major minor #blocks name 80 312571224 sda 81 979933 sda1 823911827 sda2 83 48829567 sda3 84 1 sda4 85 49287388 sda5 86 49287388 sda6 87 49287388 sda7 88 49287388 sda8 89 49287388 sda9 8 10 12410181 sda10 r...@bigone:~# zpool create zowhat raidz2 sda5 sda6 sda7 sda8 sda9 cannot open 'zowhat': dataset does not exist r...@bigone:~# zpool status pool: zowhat state: ONLINE scrub: none requested config: NAMESTATE READ WRITE CKSUM zowhat ONLINE 0 0 0 raidz2ONLINE 0 0 0 sda5ONLINE 0 0 0 sda6ONLINE 0 0 0 sda7ONLINE 0 0 0 sda8ONLINE 0 0 0 sda9ONLINE 0 0 0 errors: No known data errors r...@bigone:~# zpool list NAME SIZE USED AVAILCAP HEALTH ALTROOT - - - - - - - r...@bigone:~# zfs list no datasets available r...@bigone:~# - Original Message - All true, I just saw too many need ubuntu and zfs and thought to state the obvious in case the patch set for nexenta happen to differ enough to provide a working set. I've had nexenta succeed where opensolaris quarter releases failed and vice versa On Jun 27, 2010, at 9:54 PM, Erik Trimble erik.trim...@oracle.com wrote: On 6/27/2010 9:07 PM, Richard Elling wrote: On Jun 27, 2010, at 8:52 PM, Erik Trimble wrote: But that won't solve the OP's problem, which was that OpenSolaris doesn't support his hardware. Nexenta has the same hardware limitations as OpenSolaris. AFAICT, the OP's problem is with a keyboard. The vagaries of keyboards is well documented, but there is no silver bullet. Indeed, I have one box that seems to be more or less happy with PS-2 vs USB for every other OS or hypervisor. My advice, have one of each handy, just in case. -- richard Right. I was just pointing out the fallacy of thinking that Nexenta might work on hardware that OpenSolaris doesn't (or has problems with). -- Erik Trimble Java System Support Mailstop: usca22-123 Phone: x17195 Santa Clara, CA ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS bug - should I be worried about this?
mmmI double checked some of the running systems. Most of them have the first patch (sparc-122640-05 and x86-122641-06), but not the second one (sparc-142900-09 and x86-142901-09)... ...I feel I'm right in the middle of the problem... How much am I risking?! These systems are all mirrored via zpool... Would this really make me safe without patching?? : set zfs:zfs_immediate_write_sz=10 set zfs:zvol_immediate_write_sz=10 Or a Log would be preferred? *sweat* These systems are all running for years nowand I considered them safe... Have I been at risk all this time?! -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS bug - should I be worried about this?
On 28-6-2010 12:13, Gabriele Bulfon wrote: *sweat* These systems are all running for years nowand I considered them safe... Have I been at risk all this time?! They're still running, are they not? So, stop sweating. g But you're right about the changed patching service from Oracle. It sucks big time. Safety patches should be available, even it the OS is free. You can't expect users to run unsafe systems just because they have not payed for the OS. After all, it's Oracle (SUN) who gives away the OS. -- + All that's really worth doing is what we do for others (Lewis Carrol) ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS bug - should I be worried about this?
Yes...they're still running...but being aware that a power failure causing an unexpected poweroff may make the pool unreadable is a pain Yes. Patches should be available. Or adoption may be lowering a lot... -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS bug - should I be worried about this?
On 28.06.10 16:16, Gabriele Bulfon wrote: Yes...they're still running...but being aware that a power failure causing an unexpected poweroff may make the pool unreadable is a pain Pool integrity is not affected by this issue. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] OCZ Vertex 2 Pro performance numbers
On 6/26/10 9:47 AM -0400 David Magda wrote: Crickey. Who's the genius who thinks of these URLs? SEOs ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS bug - should I be worried about this?
On Mon, 2010-06-28 at 05:16 -0700, Gabriele Bulfon wrote: Yes...they're still running...but being aware that a power failure causing an unexpected poweroff may make the pool unreadable is a pain Yes. Patches should be available. Or adoption may be lowering a lot... I don't have access to the information, but if this problem is the same one I think it is, then the pool does not become unreadable. Rather, its state after such an event represents a *consistent* state from some point of time *earlier* than that confirmed fsync() (or a write on a file opened with O_SYNC or O_DSYNC). For most users, this is not a critical failing. For users using databases or requiring transactional integrity for data stored on ZFS, then yes, this is a very nasty problem indeed. I suspect that this is the problem I reported earlier in my blog (http://gdamore.blogspot.com) about certain kernels having O_SYNC and O_DSYNC problems. I can't confirm this though, because I don't have access to the SunSolve database to read the report. (This is something I'll have to check into fixing... it seems like my employer ought to have access to that information...) - Garrett ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Announce: zfsdump
For quite some time I have been using zfs send -R fsn...@snapname | dd of=/dev/rmt/1ln to make a tape backup of my zfs file system. A few weeks back the size of the file system grew to larger than would fit on a single DAT72 tape, and I once again searched for a simple solution to allow dumping of a zfs file system to multiple tapes. Once again I was disappointed... I expect there are plenty of other ways this could have been handled, but none leapt out at me. I didn't want to pay large sums of cash for a commercial backup product, and I didn't see that Amanda would be an easy thing to fit into my existing scripts. In particular, (and I could well be reading this incorrectly) it seems that the commercial products, Amanda, star, all are dumping the zfs file system file by file (with or without ACLs). I found none which would allow me to dump the file system and its snapshots, unless I used zfs send to a scratch disk, and dumped to tape from there. But, of course, that assumes I have a scratch disk large enough. So, I have implemented zfsdump as a ksh script. The method is as follows: 1. Make a bunch of fifos. 2. Pipe the stream from zfs send to split, with split writing to the fifos (in sequence). 3. Use dd to copy from the fifos to tape(s). When the first tape is complete, zfsdump returns. One then calls it again, specifying that the second tape is to be used, and so on. From the man page: Example 1. Dump the @Tues snapshot of the tank filesystem to the non-rewinding, non-compressing tape, with a 36GB capacity: zfsdump -z t...@tues -a -R -f /dev/rmt/1ln -s 36864 -t 0 For the second tape: zfsdump -z t...@tues -a -R -f /dev/rmt/1ln -s 36864 -t 1 If you would like to try it out, download the package from: http://www.quantmodels.co.uk/zfsdump/ I have packaged it up, so do the usual pkgadd stuff to install. Please, though, [b]try this out with caution[/b]. Build a few test file systems, and see that it works for you. [b]It comes without warranty of any kind.[/b] Tristram -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Announce: zfsdump
I use Bacula which works very well (much better than Amanda did). You may be able to customize it to do direct zfs send/receive, however I find that although they are great for copying file systems to other machines, they are inadequate for backups unless you always intend to restore the whole file system. Most people want to restore a file or directory tree of files, not a whole file system. In the past 25 years of backups and restores, I've never had to restore a whole file system. I get requests for a few files, or somebody's mailbox or somebody's http document root. You can directly install it from CSW (or blastwave). On 6/28/2010 11:26 AM, Tristram Scott wrote: For quite some time I have been using zfs send -R fsn...@snapname | dd of=/dev/rmt/1ln to make a tape backup of my zfs file system. A few weeks back the size of the file system grew to larger than would fit on a single DAT72 tape, and I once again searched for a simple solution to allow dumping of a zfs file system to multiple tapes. Once again I was disappointed... I expect there are plenty of other ways this could have been handled, but none leapt out at me. I didn't want to pay large sums of cash for a commercial backup product, and I didn't see that Amanda would be an easy thing to fit into my existing scripts. In particular, (and I could well be reading this incorrectly) it seems that the commercial products, Amanda, star, all are dumping the zfs file system file by file (with or without ACLs). I found none which would allow me to dump the file system and its snapshots, unless I used zfs send to a scratch disk, and dumped to tape from there. But, of course, that assumes I have a scratch disk large enough. So, I have implemented zfsdump as a ksh script. The method is as follows: 1. Make a bunch of fifos. 2. Pipe the stream from zfs send to split, with split writing to the fifos (in sequence). 3. Use dd to copy from the fifos to tape(s). When the first tape is complete, zfsdump returns. One then calls it again, specifying that the second tape is to be used, and so on. From the man page: Example 1. Dump the @Tues snapshot of the tank filesystem to the non-rewinding, non-compressing tape, with a 36GB capacity: zfsdump -z t...@tues -a -R -f /dev/rmt/1ln -s 36864 -t 0 For the second tape: zfsdump -z t...@tues -a -R -f /dev/rmt/1ln -s 36864 -t 1 If you would like to try it out, download the package from: http://www.quantmodels.co.uk/zfsdump/ I have packaged it up, so do the usual pkgadd stuff to install. Please, though, [b]try this out with caution[/b]. Build a few test file systems, and see that it works for you. [b]It comes without warranty of any kind.[/b] Tristram ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
Now at 36 hours since zdb process start and: PID USERNAME SIZE RSS STATE PRI NICE TIME CPU PROCESS/NLWP 827 root 4936M 4931M sleep 590 0:50:47 0.2% zdb/209 Idling at 0.2% processor for nearly the past 24 hours... feels very stuck. Thoughts on how to determine where and why? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Announce: zfsdump
I use Bacula which works very well (much better than Amanda did). You may be able to customize it to do direct zfs send/receive, however I find that although they are great for copying file systems to other machines, they are inadequate for backups unless you always intend to restore the whole file system. Most people want to restore a file or directory tree of files, not a whole file system. In the past 25 years of backups and restores, I've never had to restore a whole file system. I get requests for a few files, or somebody's mailbox or somebody's http document root. You can directly install it from CSW (or blastwave). Thanks for your comments, Brian. I should look at Bacula in more detail. As for full restore versus ad hoc requests for files I just deleted, my experience is mostly similar to yours, although I have had need for full system restore more than once. For the restore of a few files here and there, I believe this is now well handled with zfs snapshots. I have always found these requests to be down to human actions. The need for full system restore has (almost) always been hardware failure. If the file was there an hour ago, or yesterday, or last week, or last month, then we have it in a snapshot. If the disk died horribly during a power outage (grrr!) then it would be very nice to be able to restore not only the full file system, but also the snapshots too. The only way I know of achieving that is by using zfs send etc. On 6/28/2010 11:26 AM, Tristram Scott wrote: [snip] Tristram ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discu ss -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Announce: zfsdump
On Jun 28, 2010, at 12:26 PM, Tristram Scott wrote: I use Bacula which works very well (much better than Amanda did). You may be able to customize it to do direct zfs send/receive, however I find that although they are great for copying file systems to other machines, they are inadequate for backups unless you always intend to restore the whole file system. Most people want to restore a file or directory tree of files, not a whole file system. In the past 25 years of backups and restores, I've never had to restore a whole file system. I get requests for a few files, or somebody's mailbox or somebody's http document root. You can directly install it from CSW (or blastwave). Thanks for your comments, Brian. I should look at Bacula in more detail. As for full restore versus ad hoc requests for files I just deleted, my experience is mostly similar to yours, although I have had need for full system restore more than once. For the restore of a few files here and there, I believe this is now well handled with zfs snapshots. I have always found these requests to be down to human actions. The need for full system restore has (almost) always been hardware failure. If the file was there an hour ago, or yesterday, or last week, or last month, then we have it in a snapshot. If the disk died horribly during a power outage (grrr!) then it would be very nice to be able to restore not only the full file system, but also the snapshots too. The only way I know of achieving that is by using zfs send etc. I like snapshots when I'm making a major change to the system or for cloning. So to me, snapshots are good for transaction based operations. Such as stopping flushing a database, take a snapshot, then resume the database. Then you can back up the snapshot with Bacula and destroy the snapshot when the backup is complete. I have Bacula configured with a pre-backup and post-backup scripts to do just that. When you do the restore, it will create something that looks like a snapshot from the file system perspective, but isn't really one. But if you're looking for a copy of a file from a specific date, Bacula retains that. In fact you specify the retention period you want and you'll have access to any/all individual files on a per date basis. You can retain the files for months or years if you like, and you specify that in the Bacula config file as to how long you want to keep the tapes around. So it really comes down to your use-case. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
Update: have given up on the zdb write mode repair effort, as least for now. Hoping for any guidance / direction anyone's willing to offer... Re-running 'zpool import -F -f tank' with some stack trace debug, as suggested in similar threads elsewhere. Note that this appears hung at near idle. ff03e278c520 ff03e9c60038 ff03ef109490 1 60 ff0530db4680 PC: _resume_from_idle+0xf1CMD: zpool import -F -f tank stack pointer for thread ff03e278c520: ff00182bbff0 [ ff00182bbff0 _resume_from_idle+0xf1() ] swtch+0x145() cv_wait+0x61() zio_wait+0x5d() dbuf_read+0x1e8() dnode_next_offset_level+0x129() dnode_next_offset+0xa2() get_next_chunk+0xa5() dmu_free_long_range_impl+0x9e() dmu_free_object+0xe6() dsl_dataset_destroy+0x122() dsl_destroy_inconsistent+0x5f() findfunc+0x23() dmu_objset_find_spa+0x38c() dmu_objset_find_spa+0x153() dmu_objset_find+0x40() spa_load_impl+0xb23() spa_load+0x117() spa_load_best+0x78() spa_import+0xee() zfs_ioc_pool_import+0xc0() zfsdev_ioctl+0x177() cdev_ioctl+0x45() spec_ioctl+0x5a() fop_ioctl+0x7b() ioctl+0x18e() dtrace_systrace_syscall32+0x11a() _sys_sysenter_post_swapgs+0x149() -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
- Original Message - Now at 36 hours since zdb process start and: PID USERNAME SIZE RSS STATE PRI NICE TIME CPU PROCESS/NLWP 827 root 4936M 4931M sleep 59 0 0:50:47 0.2% zdb/209 Idling at 0.2% processor for nearly the past 24 hours... feels very stuck. Thoughts on how to determine where and why? Just a hunch, is this pool using dedup? Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
I had a similar issue on boot after upgrade in the past and it was due to the large number of snapshots I had... don't know if that could be related or not... Malachi de Ælfweald http://www.google.com/profiles/malachid On Mon, Jun 28, 2010 at 8:59 AM, Andrew Jones andrewnjo...@gmail.comwrote: Now at 36 hours since zdb process start and: PID USERNAME SIZE RSS STATE PRI NICE TIME CPU PROCESS/NLWP 827 root 4936M 4931M sleep 590 0:50:47 0.2% zdb/209 Idling at 0.2% processor for nearly the past 24 hours... feels very stuck. Thoughts on how to determine where and why? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
Dedup had been turned on in the past for some of the volumes, but I had turned it off altogether before entering production due to performance issues. GZIP compression was turned on for the volume I was trying to delete. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
Malachi, Thanks for the reply. There were no snapshots for the CSV1 volume that I recall... very few snapshots on the any volume in the tank. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Kernel Panic on zpool clean
Hi, I have a machine running 2009.06 with 8 SATA drives in SCSI connected enclosure. I had a drive fail and accidentally replaced the wrong one, which unsurprisingly caused the rebuild to fail. The status of the zpool then ended up as: pool: storage2 state: FAULTED status: An intent log record could not be read. Waiting for adminstrator intervention to fix the faulted pool. action: Either restore the affected device(s) and run 'zpool online', or ignore the intent log records by running 'zpool clear'. see: http://www.sun.com/msg/ZFS-8000-K4 scrub: none requested config: NAME STATE READ WRITE CKSUM storage2 FAULTED 0 0 1 bad intent log raidz1 ONLINE 0 0 0 c9t4d2 ONLINE 0 0 0 c9t4d3 ONLINE 0 0 0 c10t4d2ONLINE 0 0 0 c10t4d4ONLINE 0 0 0 raidz1 DEGRADED 0 0 6 c10t4d0UNAVAIL 0 0 0 cannot open replacing ONLINE 0 0 0 c9t4d0 ONLINE 0 0 0 c10t4d3 ONLINE 0 0 0 c10t4d1ONLINE 0 0 0 c9t4d1 ONLINE 0 0 0 running zpool clear storage2 caused the machine to dump and reboot. I've tried removing the spare and putting back the faulty drive to give: pool: storage2 state: FAULTED status: An intent log record could not be read. Waiting for adminstrator intervention to fix the faulted pool. action: Either restore the affected device(s) and run 'zpool online', or ignore the intent log records by running 'zpool clear'. see: http://www.sun.com/msg/ZFS-8000-K4 scrub: none requested config: NAME STATE READ WRITE CKSUM storage2 FAULTED 0 0 1 bad intent log raidz1 ONLINE 0 0 0 c9t4d2 ONLINE 0 0 0 c9t4d3 ONLINE 0 0 0 c10t4d2ONLINE 0 0 0 c10t4d4ONLINE 0 0 0 raidz1 DEGRADED 0 0 6 c10t4d0FAULTED 0 0 0 corrupted data replacing DEGRADED 0 0 0 c9t4d0 ONLINE 0 0 0 c9t4d4 UNAVAIL 0 0 0 cannot open c10t4d1ONLINE 0 0 0 c9t4d1 ONLINE 0 0 0 Again this core dumps when I try to do zpool clear storage2 Does anyone have any suggestions what would be the best course of action now? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Dedup RAM requirements, vs. L2ARC?
I'm putting together a new server, based on a Dell PowerEdge T410. I have simple SAS controller, with six 2TB Hitachi DeskStar 7200 RPM SATA drives. The processor is a quad-core 2 GHz Core i7-based Xeon. I will run the drives as one set of three mirror pairs striped together, for 6 TB of homogeneous storage. I'd like to run Dedup, but right now the server has only 4 GB of RAM. It has been pointed out to me several times that this is far too little. So how much should I buy? A few considerations: 1. I would like to run dedup on old copies of backups (dedup ratio for these filesystems are 3+). Basically I have a few years of backups onto tape, and will consolidate these. I need to have the data there on disk, but I rarely need to access it (maybe once a month). So those filesystems can be exported, and effectively shut off. Am I correct in guessing that, if a filesystem has been exported, its dedup table is not in RAM, and therefore is not relevant to RAM requirements? I don't mind if it's really slow to do the first and only copy to the file system, as I can let it run for a week without a problem. 2. Are the RAM requirements for ZFS with dedup based on the total available zpool size (I'm not using thin provisioning), or just on how much data is in the filesystem being deduped? That is, if I have 500 GB of deduped data but 6 TB of possible storage, which number is relevant for calculating RAM requirements? 3. What are the RAM requirements for ZFS in the absence of dedup? That is, if I only have deduped filesystems in an exported state, and all that is active is non-deduped, is 4 GB enough? 4. How does the L2ARC come into play? I can afford to buy a fast Intel X25M G2, for instance, or any of the newer SandForce-based MLC SSDs to cache the dedup table. But does it work that way? It's not really affordable for me to get more than 16 GB of RAM on this system, because there are only four slots available, and the 8 GB DIMMS are a bit pricey. 5. Could I use one of the PCIe-based SSD cards for this purpose, such as the brand-new OCZ Revo? That should be somewhere between a SATA-based SSD and RAM. Thanks in advance for all of your advice and help. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
- Original Message - Dedup had been turned on in the past for some of the volumes, but I had turned it off altogether before entering production due to performance issues. GZIP compression was turned on for the volume I was trying to delete. Was there a lot of deduped data still on disk before it was put into production? Turning off dedup won't dedup the data, just inhibit deduplication of new data... Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
Just re-ran 'zdb -e tank' to confirm the CSV1 volume is still exhibiting error 16: snip Could not open tank/CSV1, error 16 snip Considering my attempt to delete the CSV1 volume lead to the failure in the first place, I have to think that if I can either 1) complete the deletion of this volume or 2) roll back to a transaction prior to this based on logging or 3) repair whatever corruption has been caused by this partial deletion, that I will then be able to import the pool. What does 'error 16' mean in the ZDB output, any suggestions? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Dedup RAM requirements, vs. L2ARC?
2. Are the RAM requirements for ZFS with dedup based on the total available zpool size (I'm not using thin provisioning), or just on how much data is in the filesystem being deduped? That is, if I have 500 GB of deduped data but 6 TB of possible storage, which number is relevant for calculating RAM requirements? It's based on the data stored in the zpool. You'll need about 200 bytes of per DDT (data deduplication table) entry, meaning about 1,2GB per 1TB stored on 128kB blocks. With smaller blocks (smaller files are stored in smaller blocks), that means more memory. With only large files, 1,2GB or 1,5GB per 1TB stored data should sufffice. 3. What are the RAM requirements for ZFS in the absence of dedup? That is, if I only have deduped filesystems in an exported state, and all that is active is non-deduped, is 4 GB enough? Probably not, see above. 4. How does the L2ARC come into play? I can afford to buy a fast Intel X25M G2, for instance, or any of the newer SandForce-based MLC SSDs to cache the dedup table. But does it work that way? It's not really affordable for me to get more than 16 GB of RAM on this system, because there are only four slots available, and the 8 GB DIMMS are a bit pricey. L2ARC will buffer the DDT along with the data, so if you get some good SSDs (such as Crucial RealSSD C300), this will speed things up quite a bit. 5. Could I use one of the PCIe-based SSD cards for this purpose, such as the brand-new OCZ Revo? That should be somewhere between a SATA-based SSD and RAM. If your budget is low, as it may seem, good SATA SSDs will probably be the best. They can help out quite a bit. Just remember that dedup on opensolaris is not thoroughly tested yet. It works, but AFAIK there are still issues with long hangs in case of unexpected reboots. Disclaimer: I'm not an Oracle (nor Sun) employee - this is just my advice to you based on testing dedup on my test systems. Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Dedup RAM requirements, vs. L2ARC?
On 6/28/2010 12:33 PM, valrh...@gmail.com wrote: I'm putting together a new server, based on a Dell PowerEdge T410. I have simple SAS controller, with six 2TB Hitachi DeskStar 7200 RPM SATA drives. The processor is a quad-core 2 GHz Core i7-based Xeon. I will run the drives as one set of three mirror pairs striped together, for 6 TB of homogeneous storage. I'd like to run Dedup, but right now the server has only 4 GB of RAM. It has been pointed out to me several times that this is far too little. So how much should I buy? A few considerations: 1. I would like to run dedup on old copies of backups (dedup ratio for these filesystems are 3+). Basically I have a few years of backups onto tape, and will consolidate these. I need to have the data there on disk, but I rarely need to access it (maybe once a month). So those filesystems can be exported, and effectively shut off. Am I correct in guessing that, if a filesystem has been exported, its dedup table is not in RAM, and therefore is not relevant to RAM requirements? I don't mind if it's really slow to do the first and only copy to the file system, as I can let it run for a week without a problem. That's correct. An exported pool is effectively ignored by the system, so it won't contribute to any ARC requirements. 2. Are the RAM requirements for ZFS with dedup based on the total available zpool size (I'm not using thin provisioning), or just on how much data is in the filesystem being deduped? That is, if I have 500 GB of deduped data but 6 TB of possible storage, which number is relevant for calculating RAM requirements? Requirements are based on *current* BLOCK usage, after dedup has occurred. That is, ZFS needs an entry in the DDT for each block actually allocated in the filesystem. The number of times that block is referenced won't influence the DDT size, nor will the *potential* size of the pool matter (other than for capacity planning). Remember that ZFS uses variable size blocks, so you need to determine what your average block size is in order to estimate your DDT usage. 3. What are the RAM requirements for ZFS in the absence of dedup? That is, if I only have deduped filesystems in an exported state, and all that is active is non-deduped, is 4 GB enough? It of course depends heavily on your usage pattern, and the kind of files you are serving up. ZFS requires at a bare minimum a couple of dozen MB for its own usage. Everything above that is caching. Heavy write I/O will also eat up RAM, as ZFS needs to cache the writes in RAM before doing a large write I/O to backing store. Take a look at the amount of data you expect to be using heavily - your RAM should probably exceed this amount, plus an additional 1GB or so for the OS/ZFS/kernel/etc use. That is assuming you are doing nothing but fileserving on the system. 4. How does the L2ARC come into play? I can afford to buy a fast Intel X25M G2, for instance, or any of the newer SandForce-based MLC SSDs to cache the dedup table. But does it work that way? It's not really affordable for me to get more than 16 GB of RAM on this system, because there are only four slots available, and the 8 GB DIMMS are a bit pricey. L2ARC is secondary ARC. ZFS attempts to cache all reads in the ARC (Adaptive Read Cache) - should it find that it doesn't have enough space in the ARC (which is RAM-resident), it will evict some data over to the L2ARC (which in turn will simply dump the least-recently-used data when it runs out of space). Remember, however, every time something gets written to the L2ARC, a little bit of space is taken up in the ARC itself (a pointer to the L2ARC entry needs to be kept in ARC). So, it's not possible to have a giant L2ARC and tiny ARC. As a rule of thump, I try not to have my L2ARC exceed my main RAM by more than 10-15x (with really bigMem machines, I'm a bit looser and allow 20-25x or so, but still...). So, if you are thinking of getting a 160GB SSD, it would be wise to go for at minimum 8GB of RAM. Once again, the amount of ARC space reserved for a L2ARC entry is fixed, and independent of the actual block size stored in L2ARC. The jist of this is that tiny files eat up a disproportionate amount of systems resources for their size (smaller size = larger % overhead vis-a-vis large files). 5. Could I use one of the PCIe-based SSD cards for this purpose, such as the brand-new OCZ Revo? That should be somewhere between a SATA-based SSD and RAM. Thanks in advance for all of your advice and help. ZFS doesn't care what you use for the L2ARC. Some of us actually use Hard drives, so a PCI-E Flash card is entirely possible. The Revo is possibly the first PCI-E Flash card that wasn't massively expensive, otherwise, I don't think they'd be a good option. They're going to be more expensive than even an SLC SSD, however. In addition, given that L2ARC is heavily read-biased, cheap MLC SSDs are
Re: [zfs-discuss] Dedup RAM requirements, vs. L2ARC?
On 6/28/2010 12:53 PM, Roy Sigurd Karlsbakk wrote: 2. Are the RAM requirements for ZFS with dedup based on the total available zpool size (I'm not using thin provisioning), or just on how much data is in the filesystem being deduped? That is, if I have 500 GB of deduped data but 6 TB of possible storage, which number is relevant for calculating RAM requirements? It's based on the data stored in the zpool. You'll need about 200 bytes of per DDT (data deduplication table) entry, meaning about 1,2GB per 1TB stored on 128kB blocks. With smaller blocks (smaller files are stored in smaller blocks), that means more memory. With only large files, 1,2GB or 1,5GB per 1TB stored data should sufffice. Actually, I think the rule-of-thumb is 270 bytes/DDT entry. It's 200 bytes of ARC for every L2ARC entry. DDT doesn't count for this ARC space usage E.g.:I have 1TB of 4k files that are to be deduped, and it turns out that I have about a 5:1 dedup ratio. I'd also like to see how much ARC usage I eat up with a 160GB L2ARC. (1)How many entries are there in the DDT: 1TB of 4k files means there are 2^30 files (about 1 billion). However, at a 5:1 dedup ratio, I'm only actually storing 20% of that, so I have about 214 million blocks. Thus, I need a DDT of about 270 * 214 million =~ 58GB in size (2)My L2ARC is 160GB in size, but I'm using 58GB for the DDT. Thus, I have 102GB free for use as a data cache. 102GB / 4k =~ 27 million blocks can be stored in the remaining L2ARC space. However, 26 million files takes up: 200 * 27 million =~ 5.5GB of space in ARC Thus, I'd better have at least 5.5GB of RAM allocated solely for L2ARC reference pointers, and no other use. 4. How does the L2ARC come into play? I can afford to buy a fast Intel X25M G2, for instance, or any of the newer SandForce-based MLC SSDs to cache the dedup table. But does it work that way? It's not really affordable for me to get more than 16 GB of RAM on this system, because there are only four slots available, and the 8 GB DIMMS are a bit pricey. L2ARC will buffer the DDT along with the data, so if you get some good SSDs (such as Crucial RealSSD C300), this will speed things up quite a bit. 5. Could I use one of the PCIe-based SSD cards for this purpose, such as the brand-new OCZ Revo? That should be somewhere between a SATA-based SSD and RAM. If your budget is low, as it may seem, good SATA SSDs will probably be the best. They can help out quite a bit. Just remember that dedup on opensolaris is not thoroughly tested yet. It works, but AFAIK there are still issues with long hangs in case of unexpected reboots. Disclaimer: I'm not an Oracle (nor Sun) employee - this is just my advice to you based on testing dedup on my test systems. Vennlige hilsener / Best regards roy While I'm an Oracle employee, but I don't have any insider knowledge on this. It's solely my experience talking. -- Erik Trimble Java System Support Mailstop: usca22-123 Phone: x17195 Santa Clara, CA ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
On Jun 28, 2010, at 9:32 PM, Andrew Jones wrote: Update: have given up on the zdb write mode repair effort, as least for now. Hoping for any guidance / direction anyone's willing to offer... Re-running 'zpool import -F -f tank' with some stack trace debug, as suggested in similar threads elsewhere. Note that this appears hung at near idle. It looks like it is processing huge inconsistent data set that was destroyed previously. So you need to wait a bit longer. regards victor ff03e278c520 ff03e9c60038 ff03ef109490 1 60 ff0530db4680 PC: _resume_from_idle+0xf1CMD: zpool import -F -f tank stack pointer for thread ff03e278c520: ff00182bbff0 [ ff00182bbff0 _resume_from_idle+0xf1() ] swtch+0x145() cv_wait+0x61() zio_wait+0x5d() dbuf_read+0x1e8() dnode_next_offset_level+0x129() dnode_next_offset+0xa2() get_next_chunk+0xa5() dmu_free_long_range_impl+0x9e() dmu_free_object+0xe6() dsl_dataset_destroy+0x122() dsl_destroy_inconsistent+0x5f() findfunc+0x23() dmu_objset_find_spa+0x38c() dmu_objset_find_spa+0x153() dmu_objset_find+0x40() spa_load_impl+0xb23() spa_load+0x117() spa_load_best+0x78() spa_import+0xee() zfs_ioc_pool_import+0xc0() zfsdev_ioctl+0x177() cdev_ioctl+0x45() spec_ioctl+0x5a() fop_ioctl+0x7b() ioctl+0x18e() dtrace_systrace_syscall32+0x11a() _sys_sysenter_post_swapgs+0x149() -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Resilvering onto a spare - degraded because of read and cksum errors
Hi Donald, I think this is just a reporting error in the zpool status output, depending on what Solaris release is. Thanks, Cindy On 06/27/10 15:13, Donald Murray, P.Eng. wrote: Hi, I awoke this morning to a panic'd opensolaris zfs box. I rebooted it and confirmed it would panic each time it tried to import the 'tank' pool. Once I disconnected half of one of the mirrored disks, the box booted cleanly and the pool imported without a panic. Because this box has a hot spare, it began resilvering automatically. This is the first time I've resilvered to a hot spare, so I'm not sure whether the output below [1] is normal. In particular, I think it's odd that the spare has an equal number of read and cksum errors. Is this normal? Is my spare a piece of junk, just like the disk it replaced? [1] r...@weyl:~# zpool status tank pool: tank state: DEGRADED status: One or more devices could not be opened. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Attach the missing device and online it using 'zpool online'. see: http://www.sun.com/msg/ZFS-8000-2Q scrub: resilver in progress for 3h42m, 97.34% done, 0h6m to go config: NAME STATE READ WRITE CKSUM tank DEGRADED 0 0 0 mirror DEGRADED 0 0 0 spare DEGRADED 1.36M 0 0 9828443264686839751 UNAVAIL 0 0 0 was /dev/dsk/c6t1d0s0 c7t1d0 DEGRADED 0 0 1.36M too many errors c9t0d0 ONLINE 0 0 0 mirror ONLINE 0 0 0 c7t0d0 ONLINE 0 0 0 c5t1d0 ONLINE 0 0 0 spares c7t1d0 INUSE currently in use errors: No known data errors r...@weyl:~# ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Kernel Panic on zpool clean
On Jun 28, 2010, at 11:27 PM, George wrote: I've tried removing the spare and putting back the faulty drive to give: pool: storage2 state: FAULTED status: An intent log record could not be read. Waiting for adminstrator intervention to fix the faulted pool. action: Either restore the affected device(s) and run 'zpool online', or ignore the intent log records by running 'zpool clear'. see: http://www.sun.com/msg/ZFS-8000-K4 scrub: none requested config: NAME STATE READ WRITE CKSUM storage2 FAULTED 0 0 1 bad intent log raidz1 ONLINE 0 0 0 c9t4d2 ONLINE 0 0 0 c9t4d3 ONLINE 0 0 0 c10t4d2ONLINE 0 0 0 c10t4d4ONLINE 0 0 0 raidz1 DEGRADED 0 0 6 c10t4d0FAULTED 0 0 0 corrupted data replacing DEGRADED 0 0 0 c9t4d0 ONLINE 0 0 0 c9t4d4 UNAVAIL 0 0 0 cannot open c10t4d1ONLINE 0 0 0 c9t4d1 ONLINE 0 0 0 Again this core dumps when I try to do zpool clear storage2 Does anyone have any suggestions what would be the best course of action now? I think first we need to understand why it does not like 'zpool clear', as that may provide better understanding of what is wrong. For that you need to create directory for saving crashdumps e.g. like this mkdir -p /var/crash/`uname -n` then run savecore and see if it would save a crash dump into that directory. If crashdump is there, then you need to perform some basic investigation: cd /var/crash/`uname -n` mdb dump number ::status ::stack ::spa -c ::spa -v ::spa -ve $q for a start. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Kernel Panic on zpool clean
I've attached the output of those commands. The machine is a v20z if that makes any difference. Thanks, George -- This message posted from opensolaris.orgmdb: logging to debug.txt ::status debugging crash dump vmcore.0 (64-bit) from crypt operating system: 5.11 snv_111b (i86pc) panic message: BAD TRAP: type=e (#pf Page fault) rp=ff00084fc660 addr=0 occurred in module unix due to a NULL pointer dereference dump content: kernel pages only ::stack mutex_enter+0xb() metaslab_free+0x12e(ff01c9fb3800, ff01cce64668, 1b9528, 0) zio_dva_free+0x26(ff01cce64608) zio_execute+0xa0(ff01cce64608) zio_nowait+0x5a(ff01cce64608) arc_free+0x197(ff01cf0c80c0, ff01c9fb3800, 1b9528, ff01d389bcf0, 0, 0) dsl_free+0x30(ff01cf0c80c0, ff01d389bcc0, 1b9528, ff01d389bcf0, 0, 0 ) dsl_dataset_block_kill+0x293(0, ff01d389bcf0, ff01cf0c80c0, ff01d18cfd80) dmu_objset_sync+0xc4(ff01cffe0080, ff01cf0c80c0, ff01d18cfd80) dsl_pool_sync+0x1ee(ff01d389bcc0, 1b9528) spa_sync+0x32a(ff01c9fb3800, 1b9528) txg_sync_thread+0x265(ff01d389bcc0) thread_start+8() ::spa -c ADDR STATE NAME ff01c8df3000ACTIVE rpool version=000e name='rpool' state= txg=056a6ad1 pool_guid=53825ef3c58abc97 hostid=00820b9b hostname='crypt' vdev_tree type='root' id= guid=53825ef3c58abc97 children[0] type='mirror' id= guid=e9b8daed37492cfe whole_disk= metaslab_array=0017 metaslab_shift=001d ashift=0009 asize=001114e0 is_log= children[0] type='disk' id= guid=ad7e5022f804365a path='/dev/dsk/c8t0d0s0' devid='id1,s...@sseagate_st373307lc__3hz76yyd743809wm/a' phys_path='/p...@0,0/pci1022,7...@a/pci17c2,1...@4/s...@0,0:a' whole_disk= DTL=0052 children[1] type='disk' id=0001 guid=2f7a03c75a4931ac path='/dev/dsk/c8t1d0s0' devid='id1,s...@sseagate_st373307lc__3hz80bdp743793pa/a' phys_path='/p...@0,0/pci1022,7...@a/pci17c2,1...@4/s...@1,0:a' whole_disk= DTL=0050 ff01c9fb3800ACTIVE storage2 version=000e name='storage2' state= txg=001b9406 pool_guid=cc049c0f1321fc28 hostid=00820b9b hostname='crypt' vdev_tree type='root' id= guid=cc049c0f1321fc28 children[0] type='raidz' id= guid=dc1ecf18721028c1 nparity=0001 metaslab_array=000e metaslab_shift=0023 ashift=0009 asize=03a33f10 is_log= children[0] type='disk' id= guid=c7b64596709ebdef path='/dev/dsk/c9t4d2s0' devid='id1,s...@n600d0230006c8a5f0c3fd863ea736d00/a' phys_path='/p...@0,0/pci1022,7...@b/pci9005,4...@1/s...@4,2:a' whole_disk=0001 DTL=012d children[1] type='disk' id=0001 guid=cd7ba5d38162fe0d path='/dev/dsk/c9t4d3s0' devid='id1,s...@n600d0230006c8a5f0c3fd8514ed8d900/a' phys_path='/p...@0,0/pci1022,7...@b/pci9005,4...@1/s...@4,3:a' whole_disk=0001 DTL=012c children[2] type='disk' id=0002 guid=3b499fb48e06460b path='/dev/dsk/c10t4d2s0' devid='id1,s...@n600d0230006c8a5f0c3fd84312aa6d00/a' phys_path='/p...@0,0/pci1022,7...@b/pci9005,4...@1,1/s...@4,2:a' whole_disk=0001 DTL=012b children[3] type='disk' id=0003 guid=e205849496e5e447 path='/dev/dsk/c10t4d4s0' devid='id1,s...@n600d0230006c8a5f0c3fd8415c62ae00/a' phys_path='/p...@0,0/pci1022,7...@b/pci9005,4...@1,1/s...@4,4:a' whole_disk=0001 DTL=0128 children[1] type='raidz'
[zfs-discuss] COMSTAR ISCSI - configuration export/impo rt
Hi all, Having osol b134 exporting a couple of iscsi targets to some hosts,how can the COMSTAR configuration be migrated to other host? I can use the ZFS send/receive to replicate the luns but how can I replicate the target,views from serverA to serverB ? Is there any best procedures to follow to accomplish this? Thanks for all your time, Bruno Sent from my HTC -- This message has been scanned for viruses and dangerous content by MailScanner, and is believed to be clean. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
Thanks Victor. I will give it another 24 hrs or so and will let you know how it goes... You are right, a large 2TB volume (CSV1) was not in the process of being deleted, as described above. It is showing error 16 on 'zdb -e' -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] COMSTAR ISCSI - configuration export/import
I havnt tried it yet, but supposedly this will backup/restore the comstar config: $ svccfg export -a stmf comstar.bak.${DATE} If you ever need to restore the configuration, you can attach the storage and run an import: $ svccfg import comstar.bak.${DATE} - Mike On 6/28/10, bso...@epinfante.com bso...@epinfante.com wrote: Hi all, Having osol b134 exporting a couple of iscsi targets to some hosts,how can the COMSTAR configuration be migrated to other host? I can use the ZFS send/receive to replicate the luns but how can I replicate the target,views from serverA to serverB ? Is there any best procedures to follow to accomplish this? Thanks for all your time, Bruno Sent from my HTC -- This message has been scanned for viruses and dangerous content by MailScanner, and is believed to be clean. -- Sent from my mobile device ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS bug - should I be worried about this?
Oh well, thanks for this answer. It makes me feel much better! What are eventual risks? Gabriele Bulfon - Sonicle S.r.l. Tel +39 028246016 Int. 30 - Fax +39 028243880 Via Felice Cavallotti 16 - 20089, Rozzano - Milano - ITALY http://www.sonicle.com -- Da: Victor Latushkin A: Gabriele Bulfon Cc: zfs-discuss@opensolaris.org Data: 28 giugno 2010 16.14.12 CEST Oggetto: Re: [zfs-discuss] ZFS bug - should I be worried about this? On 28.06.10 16:16, Gabriele Bulfon wrote: Yes...they're still running...but being aware that a power failure causing an unexpected poweroff may make the pool unreadable is a pain Pool integrity is not affected by this issue. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import hangs indefinitely (retry post in parts; too long?)
Andrew, Looks like the zpool is telling you the devices are still doing work of some kind, or that there are locks still held. From man of section 2 intro page the errors are listed. Number 16 looks to be an EBUSY. 16 EBUSYDevice busy An attempt was made to mount a dev- ice that was already mounted or an attempt was made to unmount a device on which there is an active file (open file, current directory, mounted-on file, active text seg- ment). It will also occur if an attempt is made to enable accounting when it is already enabled. The device or resource is currently una- vailable. EBUSY is also used by mutexes, semaphores, condition vari- ables, and r/w locks, to indicate that a lock is held, and by the processor control function P_ONLINE. On 06/28/10 01:50 PM, Andrew Jones wrote: Just re-ran 'zdb -e tank' to confirm the CSV1 volume is still exhibiting error 16: snip Could not open tank/CSV1, error 16 snip Considering my attempt to delete the CSV1 volume lead to the failure in the first place, I have to think that if I can either 1) complete the deletion of this volume or 2) roll back to a transaction prior to this based on logging or 3) repair whatever corruption has been caused by this partial deletion, that I will then be able to import the pool. What does 'error 16' mean in the ZDB output, any suggestions? -- Geoff Shipman | Senior Technical Support Engineer Phone: +13034644710 Oracle Global Customer Services 500 Eldorado Blvd. UBRM-04 | Broomfield, CO 80021 Email: geoff.ship...@sun.com | Hours:9am-5pm MT,Monday-Friday ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Announce: zfsdump
From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- boun...@opensolaris.org] On Behalf Of Tristram Scott If you would like to try it out, download the package from: http://www.quantmodels.co.uk/zfsdump/ I haven't tried this yet, but thank you very much! Other people have pointed out bacula is able to handle multiple tapes, and individual file restores. However, the disadvantage of bacula/tar/cpio/rsync etc is that they all have to walk the entire filesystem searching for things that have changed. The advantage of zfs send (assuming incremental backups) is that it already knows what's changed, and it can generate a continuous datastream almost instantly. Something like 1-2 orders of magnitude faster per incremental backup. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Resilvering onto a spare - degraded because of read and cksum errors
Thanks Cindy. I'm running 111b at the moment. I ran a scrub last night, and it still reports the same status. r...@weyl:~# uname -a SunOS weyl 5.11 snv_111b i86pc i386 i86pc Solaris r...@weyl:~# zpool status -x pool: tank state: DEGRADED status: One or more devices could not be opened. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Attach the missing device and online it using 'zpool online'. see: http://www.sun.com/msg/ZFS-8000-2Q scrub: scrub completed after 2h40m with 0 errors on Mon Jun 28 01:23:12 2010 config: NAME STATE READ WRITE CKSUM tank DEGRADED 0 0 0 mirror DEGRADED 0 0 0 spare DEGRADED 1.37M 0 0 9828443264686839751 UNAVAIL 0 0 0 was /dev/dsk/c6t1d0s0 c7t1d0 DEGRADED 0 0 1.37M too many errors c9t0d0 ONLINE 0 0 0 mirror ONLINE 0 0 0 c7t0d0 ONLINE 0 0 0 c5t1d0 ONLINE 0 0 0 spares c7t1d0 INUSE currently in use errors: No known data errors r...@weyl:~# On Mon, Jun 28, 2010 at 14:55, Cindy Swearingen cindy.swearin...@oracle.com wrote: Hi Donald, I think this is just a reporting error in the zpool status output, depending on what Solaris release is. Thanks, Cindy On 06/27/10 15:13, Donald Murray, P.Eng. wrote: Hi, I awoke this morning to a panic'd opensolaris zfs box. I rebooted it and confirmed it would panic each time it tried to import the 'tank' pool. Once I disconnected half of one of the mirrored disks, the box booted cleanly and the pool imported without a panic. Because this box has a hot spare, it began resilvering automatically. This is the first time I've resilvered to a hot spare, so I'm not sure whether the output below [1] is normal. In particular, I think it's odd that the spare has an equal number of read and cksum errors. Is this normal? Is my spare a piece of junk, just like the disk it replaced? [1] r...@weyl:~# zpool status tank pool: tank state: DEGRADED status: One or more devices could not be opened. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Attach the missing device and online it using 'zpool online'. see: http://www.sun.com/msg/ZFS-8000-2Q scrub: resilver in progress for 3h42m, 97.34% done, 0h6m to go config: NAME STATE READ WRITE CKSUM tank DEGRADED 0 0 0 mirror DEGRADED 0 0 0 spare DEGRADED 1.36M 0 0 9828443264686839751 UNAVAIL 0 0 0 was /dev/dsk/c6t1d0s0 c7t1d0 DEGRADED 0 0 1.36M too many errors c9t0d0 ONLINE 0 0 0 mirror ONLINE 0 0 0 c7t0d0 ONLINE 0 0 0 c5t1d0 ONLINE 0 0 0 spares c7t1d0 INUSE currently in use errors: No known data errors r...@weyl:~# ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Announce: zfsdump
On Mon, Jun 28, 2010 at 11:26 AM, Tristram Scott tristram.sc...@quantmodels.co.uk wrote: For quite some time I have been using zfs send -R fsn...@snapname | dd of=/dev/rmt/1ln to make a tape backup of my zfs file system. A few weeks back the size of the file system grew to larger than would fit on a single DAT72 tape, and I once again searched for a simple solution to allow dumping of a zfs file system to multiple tapes. Once again I was disappointed... I expect there are plenty of other ways this could have been handled, but none leapt out at me. I didn't want to pay large sums of cash for a commercial backup product, and I didn't see that Amanda would be an easy thing to fit into my existing scripts. In particular, (and I could well be reading this incorrectly) it seems that the commercial products, Amanda, star, all are dumping the zfs file system file by file (with or without ACLs). I found none which would allow me to dump the file system and its snapshots, unless I used zfs send to a scratch disk, and dumped to tape from there. But, of course, that assumes I have a scratch disk large enough. So, I have implemented zfsdump as a ksh script. The method is as follows: 1. Make a bunch of fifos. 2. Pipe the stream from zfs send to split, with split writing to the fifos (in sequence). would be nice if i could pipe the zfs send stream to a split and then send of those splitted stream over the network to a remote system. it would help sending it over to remote system quicker. can your tool do that? something like this s | - | j zfs send p | - | o zfs recv (local) l | - | i(remote) t | - | n 3. Use dd to copy from the fifos to tape(s). When the first tape is complete, zfsdump returns. One then calls it again, specifying that the second tape is to be used, and so on. From the man page: Example 1. Dump the @Tues snapshot of the tank filesystem to the non-rewinding, non-compressing tape, with a 36GB capacity: zfsdump -z t...@tues -a -R -f /dev/rmt/1ln -s 36864 -t 0 For the second tape: zfsdump -z t...@tues -a -R -f /dev/rmt/1ln -s 36864 -t 1 If you would like to try it out, download the package from: http://www.quantmodels.co.uk/zfsdump/ I have packaged it up, so do the usual pkgadd stuff to install. Please, though, [b]try this out with caution[/b]. Build a few test file systems, and see that it works for you. [b]It comes without warranty of any kind.[/b] Tristram -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Asif Iqbal PGP Key: 0xE62693C5 KeyServer: pgp.mit.edu A: Because it messes up the order in which people normally read text. Q: Why is top-posting such a bad thing? ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Processes hang in /dev/zvol/dsk/poolname
After multiple power outages caused by storms coming through, I can no longer access /dev/zvol/dsk/poolname, which are hold l2arc and slog devices in another pool I don't think this is related, since I the pools are ofline pending access to the volumes. I tried running find /dev/zvol/dsk/poolname -type f and here is the stack, hopefully this someone a hint at what the issue is, I have scrubbed the pool and no errors were found, and zdb -l reports no issues that I can see. ::ps ! grep find R 1248 1243 1248 1243101 0x4a004000 ff02630d5728 find ff02630d5728::walk thread | ::findstack stack pointer for thread ff025f15b3e0: ff000cb54650 [ ff000cb54650 _resume_from_idle+0xf1() ] ff000cb54680 swtch+0x145() ff000cb546b0 cv_wait+0x61() ff000cb54700 txg_wait_synced+0x7c() ff000cb54770 zil_replay+0xe8() ff000cb54830 zvol_create_minor+0x227() ff000cb54850 sdev_zvol_create_minor+0x19() ff000cb549c0 devzvol_create_link+0x49() ff000cb54ad0 sdev_call_dircallback+0xfe() ff000cb54c20 devname_lookup_func+0x4cf() ff000cb54ca0 devzvol_lookup+0xf8() ff000cb54d20 sdev_iter_datasets+0xb0() ff000cb54da0 devzvol_readdir+0xd6() ff000cb54e20 fop_readdir+0xab() ff000cb54ec0 getdents64+0xbc() ff000cb54f10 sys_syscall32+0xff() -- DISK--- -bash-4.0$ sudo /usr/sbin/zdb -l /dev/dsk/c7t1d0s0 LABEL 0 version: 22 name: 'puddle' state: 0 txg: 3139 pool_guid: 13462109782214169516 hostid: 4421991 hostname: 'amd' top_guid: 15895240748538558983 guid: 15895240748538558983 vdev_children: 2 vdev_tree: type: 'disk' id: 0 guid: 15895240748538558983 path: '/dev/dsk/c7t1d0s0' devid: 'id1,s...@sata_hitachi_hdt72101__stf607mh3a3ksk/a' phys_path: '/p...@0,0/pci1043,8...@12/d...@1,0:a' whole_disk: 1 metaslab_array: 23 metaslab_shift: 33 ashift: 9 asize: 1000191557632 is_log: 0 DTL: 605 LABEL 1 version: 22 name: 'puddle' state: 0 txg: 3139 pool_guid: 13462109782214169516 hostid: 4421991 hostname: 'amd' top_guid: 15895240748538558983 guid: 15895240748538558983 vdev_children: 2 vdev_tree: type: 'disk' id: 0 guid: 15895240748538558983 path: '/dev/dsk/c7t1d0s0' devid: 'id1,s...@sata_hitachi_hdt72101__stf607mh3a3ksk/a' phys_path: '/p...@0,0/pci1043,8...@12/d...@1,0:a' whole_disk: 1 metaslab_array: 23 metaslab_shift: 33 ashift: 9 asize: 1000191557632 is_log: 0 DTL: 605 LABEL 2 version: 22 name: 'puddle' state: 0 txg: 3139 pool_guid: 13462109782214169516 hostid: 4421991 hostname: 'amd' top_guid: 15895240748538558983 guid: 15895240748538558983 vdev_children: 2 vdev_tree: type: 'disk' id: 0 guid: 15895240748538558983 path: '/dev/dsk/c7t1d0s0' devid: 'id1,s...@sata_hitachi_hdt72101__stf607mh3a3ksk/a' phys_path: '/p...@0,0/pci1043,8...@12/d...@1,0:a' whole_disk: 1 metaslab_array: 23 metaslab_shift: 33 ashift: 9 asize: 1000191557632 is_log: 0 DTL: 605 LABEL 3 version: 22 name: 'puddle' state: 0 txg: 3139 pool_guid: 13462109782214169516 hostid: 4421991 hostname: 'amd' top_guid: 15895240748538558983 guid: 15895240748538558983 vdev_children: 2 vdev_tree: type: 'disk' id: 0 guid: 15895240748538558983 path: '/dev/dsk/c7t1d0s0' devid: 'id1,s...@sata_hitachi_hdt72101__stf607mh3a3ksk/a' phys_path: '/p...@0,0/pci1043,8...@12/d...@1,0:a' whole_disk: 1 metaslab_array: 23 metaslab_shift: 33 ashift: 9 asize: 1000191557632 is_log: 0 DTL: 605 -bash-4.0$ there is also a volume attached that is part of the roolt pool that is accessed fine but zdb -l is not returning from zdb -l bash-4.0# /usr/sbin/zpool status -v pool: puddle state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM puddle ONLINE 0 0 0 c7t1d0 ONLINE 0 0 0 logs /dev/zvol/dsk/rpool/puddle_slog ONLINE 0 0 0 zfs list -rt volume puddle NAMEUSED AVAIL REFER MOUNTPOINT puddle/l2arc