[zfs-discuss] remove wrongly added device from zpool
Hi, For my carelessness, I added two disks into a raid-z2 zpool as normal data disk, but in fact I want to make them as zil devices. Any remedy solutions? Many thanks. Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- boun...@opensolaris.org] On Behalf Of Fred Liu For my carelessness, I added two disks into a raid-z2 zpool as normal data disk, but in fact I want to make them as zil devices. That's a huge bummer, and it's the main reason why device removal has been a priority request for such a long time... There is no solution. You can only destroy recreate your pool, or learn to live with it that way. Sorry... ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
That's a huge bummer, and it's the main reason why device removal has been a priority request for such a long time... There is no solution. You can only destroy recreate your pool, or learn to live with it that way. Sorry... Yeah, I also realized this when I send out this message. In NetApp, it is so easy to change raid group size. There is still a long way for zfs to go. Hope I can see that in the future. I also did another huge mistake which really brings me into the deep pain. I physically removed these two added devices for I though raidz2 can afford it. But now the whole pool corrupts. I don't know where I can go ... Any help will be tremendously appreciated. Thanks. Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
On 19 September, 2011 - Fred Liu sent me these 0,9K bytes: That's a huge bummer, and it's the main reason why device removal has been a priority request for such a long time... There is no solution. You can only destroy recreate your pool, or learn to live with it that way. Sorry... Yeah, I also realized this when I send out this message. In NetApp, it is so easy to change raid group size. There is still a long way for zfs to go. Hope I can see that in the future. I also did another huge mistake which really brings me into the deep pain. I physically removed these two added devices for I though raidz2 can afford it. But now the whole pool corrupts. I don't know where I can go ... Any help will be tremendously appreciated. You can add mirrors to those lonely disks. /Tomas -- Tomas Forsman, st...@acc.umu.se, http://www.acc.umu.se/~stric/ |- Student at Computing Science, University of Umeå `- Sysadmin at {cs,acc}.umu.se ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
From: Fred Liu [mailto:fred_...@issi.com] Yeah, I also realized this when I send out this message. In NetApp, it is so easy to change raid group size. There is still a long way for zfs to go. Hope I can see that in the future. This one missing feature of ZFS, IMHO, does not result in a long way for zfs to go in relation to netapp. I shut off my netapp 2 years ago in favor of ZFS, because ZFS performs so darn much better, and has such immensely greater robustness. Try doing ndmp, cifs, nfs, iscsi on netapp (all extra licenses). Try experimenting with the new version of netapp to see how good it is (you can't unless you buy a whole new box.) Try mirroring a production box onto a lower-cost secondary backup box (there is no such thing). Try storing your backup on disk and rotating your disks offsite. Try running any normal utilities - iostat, top, wireshark - you can't. Try backing up with commercial or otherwise modular (agent-based) backup software. You can't. You have to use CIFS/NFS/NDMP. Just try finding a public mailing list like this one where you can even so much as begin such a conversation about netapp... Been there done that, it's not even in the same ballpark. etc etc. (end rant.) I hate netapp. I also did another huge mistake which really brings me into the deep pain. I physically removed these two added devices for I though raidz2 can afford it. But now the whole pool corrupts. I don't know where I can go ... Any help will be tremendously appreciated. Um... Wanna post your zpool status and cat /etc/release and zpool upgrade ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
This one missing feature of ZFS, IMHO, does not result in a long way for zfs to go in relation to netapp. I shut off my netapp 2 years ago in favor of ZFS, because ZFS performs so darn much better, and has such immensely greater robustness. Try doing ndmp, cifs, nfs, iscsi on netapp (all extra licenses). Try experimenting with the new version of netapp to see how good it is (you can't unless you buy a whole new box.) Try mirroring a production box onto a lower-cost secondary backup box (there is no such thing). Try storing your backup on disk and rotating your disks offsite. Try running any normal utilities - iostat, top, wireshark - you can't. Try backing up with commercial or otherwise modular (agent-based) backup software. You can't. You have to use CIFS/NFS/NDMP. Just try finding a public mailing list like this one where you can even so much as begin such a conversation about netapp... Been there done that, it's not even in the same ballpark. etc etc. (end rant.) I hate netapp. Yeah, It is kind of touchy topic, we may discuss more in the future. I want to focus on how to repair my pool first. ;-( Um... Wanna post your zpool status and cat /etc/release and zpool upgrade I exported the pool for I want to use zpool import -F to fix it. But now I get one or more devices is currently unavailable Destroy and re-create the pool from a backup source. I use opensolaris b134 and zpool version 22. Thanks. Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
You can add mirrors to those lonely disks. Can it repair the pool? Thanks. Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
On Mon, September 19, 2011 08:07, Edward Ned Harvey wrote: This one missing feature of ZFS, IMHO, does not result in a long way for zfs to go in relation to netapp. I shut off my netapp 2 years ago in favor of ZFS, because ZFS performs so darn much better, and has such immensely greater robustness. Try doing ndmp, cifs, nfs, iscsi on netapp (all extra licenses). Try experimenting with the new version of netapp to see how good it is (you can't unless you buy a whole new box.) As another datum, at $WORK we're going to Isilon. Our NetApp is being retired by the end of the year as it just can't handle the load of HPC. We also have the regular assortment of web, mail, code repositories, etc., VMs that also live on Isilon. We're quite happy, especially with the more recent Isilon hardware that uses SSDs to store/cache metadata. NFS and CIFS are quite good, but we haven't really tried their iSCSI stuff yet; they don't have FC at all. We also have a bunch of Blue Arc, but find it much more finicky than Isilon. Perhaps Hitachi will help them stabilize things a bit. As for experimenting with NetApp, they do have a simulator that you can run in a VM if you wish (or actual hardware AFAICT). A bit more on topic, bp* rewrite has been a long-time coming, and AFAICT, it won't be in Solaris 11. As it stands, I don't care much about changing RAID levels, but not being able to remove a mistakenly added device is something is becoming more and more conspicuous. For better or worse I'm not doing as much Solaris stuff (esp. with the new Ellison pricing model), but still pay attention to what's going on, and this (missing) feature is one of those WTF? things that is the fly in the otherwise very tasty soup that is ZFS. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
On Mon, Sep 19, 2011 at 9:29 AM, Fred Liu fred_...@issi.com wrote: Yes. I have connected them back to server. But it does not help. I am really sad now... I cringed a little when I read the thread title. I did this on accident once as well, but lucky for me, I had enough scratch storage around in various sizes to cobble together a JBOD (risky) and use it as a holding area for my data while I remade the pool. I'm a home user and only have around 21TB or so, so it was feasible for me. Probably not so feasible for you enterprise guys with 1000s of users and 100s of filesystems! --khd ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
From: Krunal Desai [mailto:mov...@gmail.com] On Mon, Sep 19, 2011 at 9:29 AM, Fred Liu fred_...@issi.com wrote: Yes. I have connected them back to server. But it does not help. I am really sad now... I'll tell you what does not help. This email. Now that you know what you're trying to do, why don't you post the results of your zpool import command? How about an error message, and how you're trying to go about fixing your pool? Nobody here can help you without information. I cringed a little when I read the thread title. I did this on accident once as well, but lucky for me, I had enough scratch storage around in various sizes to cobble together a JBOD (risky) and use it as a holding area for my data while I remade the pool. I'm a home user and only have around 21TB or so, so it was feasible for me. Probably not so feasible for you enterprise guys with 1000s of users and 100s of filesystems! No enterprise guys with 1000s of users and 100s of filesystems are making this mistake. Even if it does happen, on a pool that significant, the obvious response is to add redundancy instead of recreating the pool. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
I'll tell you what does not help. This email. Now that you know what you're trying to do, why don't you post the results of your zpool import command? How about an error message, and how you're trying to go about fixing your pool? Nobody here can help you without information. User tty login@ idle JCPU PCPU what root console 9:25pm w root@cn03:~# df Filesystem 1K-blocks Used Available Use% Mounted on rpool/ROOT/opensolaris 94109412 6880699 87228713 8% / swap 108497952 344 108497608 1% /etc/svc/volatile /usr/lib/libc/libc_hwcap1.so.1 94109412 6880699 87228713 8% /lib/libc.so.1 swap 108497616 8 108497608 1% /tmp swap 10849768880 108497608 1% /var/run rpool/export 4686423 46841 1% /export rpool/export/home4686423 46841 1% /export/home rpool/export/home/fred 48710 5300 43410 11% /export/home/fred rpool10215515880 102155078 1% /rpool root@cn03:~# !z zpool import cn03 cannot import 'cn03': one or more devices is currently unavailable Destroy and re-create the pool from a backup source. Thanks. Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
I also used zpool import -fFX cn03 in b134 and b151a(via live SX11 live cd). It resulted a core dump and reboot after about 15 min. I can see all the leds are blinking on the HDD within this 15 min. Can replacing empty ZIL devices help? Thanks. Fred -Original Message- From: Fred Liu Sent: 星期一, 九月 19, 2011 21:54 To: 'Edward Ned Harvey'; 'Krunal Desai' Cc: zfs-discuss@opensolaris.org Subject: RE: [zfs-discuss] remove wrongly added device from zpool I'll tell you what does not help. This email. Now that you know what you're trying to do, why don't you post the results of your zpool import command? How about an error message, and how you're trying to go about fixing your pool? Nobody here can help you without information. User tty login@ idle JCPU PCPU what root console 9:25pm w root@cn03:~# df Filesystem 1K-blocks Used Available Use% Mounted on rpool/ROOT/opensolaris 94109412 6880699 87228713 8% / swap 108497952 344 108497608 1% /etc/svc/volatile /usr/lib/libc/libc_hwcap1.so.1 94109412 6880699 87228713 8% /lib/libc.so.1 swap 108497616 8 108497608 1% /tmp swap 10849768880 108497608 1% /var/run rpool/export 4686423 46841 1% /export rpool/export/home4686423 46841 1% /export/home rpool/export/home/fred 48710 5300 43410 11% /export/home/fred rpool10215515880 102155078 1% /rpool root@cn03:~# !z zpool import cn03 cannot import 'cn03': one or more devices is currently unavailable Destroy and re-create the pool from a backup source. Thanks. Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
The core dump: r10: ff19a5592000 r11:0 r12:0 r13:0 r14:0 r15: ff00ba4a5c60 fsb: fd7fff172a00 gsb: ff19a5592000 ds:0 es:0 fs:0 gs:0 trp:e err:0 rip: f782f81a cs: 30 rfl:10246 rsp: ff00b9bf0a40 ss: 38 ff00b9bf0830 unix:die+10f () ff00b9bf0940 unix:trap+177b () ff00b9bf0950 unix:cmntrap+e6 () ff00b9bf0ab0 procfs:prchoose+72 () ff00b9bf0b00 procfs:prgetpsinfo+2b () ff00b9bf0ce0 procfs:pr_read_psinfo+4e () ff00b9bf0d30 procfs:prread+72 () ff00b9bf0da0 genunix:fop_read+6b () ff00b9bf0f00 genunix:pread+22c () ff00b9bf0f10 unix:brand_sys_syscall+20d () syncing file systems... done dumping to /dev/zvol/dsk/rpool/dump, offset 65536, content: kernel 0:17 100% done 100% done: 1041082 pages dumped, dump succeeded rebooting... -Original Message- From: Fred Liu Sent: 星期一, 九月 19, 2011 22:00 To: Fred Liu; 'Edward Ned Harvey'; 'Krunal Desai' Cc: 'zfs-discuss@opensolaris.org' Subject: RE: [zfs-discuss] remove wrongly added device from zpool I also used zpool import -fFX cn03 in b134 and b151a(via live SX11 live cd). It resulted a core dump and reboot after about 15 min. I can see all the leds are blinking on the HDD within this 15 min. Can replacing empty ZIL devices help? Thanks. Fred -Original Message- From: Fred Liu Sent: 星期一, 九月 19, 2011 21:54 To: 'Edward Ned Harvey'; 'Krunal Desai' Cc: zfs-discuss@opensolaris.org Subject: RE: [zfs-discuss] remove wrongly added device from zpool I'll tell you what does not help. This email. Now that you know what you're trying to do, why don't you post the results of your zpool import command? How about an error message, and how you're trying to go about fixing your pool? Nobody here can help you without information. User tty login@ idle JCPU PCPU what root console 9:25pm w root@cn03:~# df Filesystem 1K-blocks Used Available Use% Mounted on rpool/ROOT/opensolaris 94109412 6880699 87228713 8% / swap 108497952 344 108497608 1% /etc/svc/volatile /usr/lib/libc/libc_hwcap1.so.1 94109412 6880699 87228713 8% /lib/libc.so.1 swap 108497616 8 108497608 1% /tmp swap 10849768880 108497608 1% /var/run rpool/export 4686423 46841 1% /export rpool/export/home4686423 46841 1% /export/home rpool/export/home/fred 48710 5300 43410 11% /export/home/fred rpool10215515880 102155078 1% /rpool root@cn03:~# !z zpool import cn03 cannot import 'cn03': one or more devices is currently unavailable Destroy and re-create the pool from a backup source. Thanks. Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
I use opensolaris b134. Thanks. Fred -Original Message- From: Richard Elling [mailto:richard.ell...@gmail.com] Sent: 星期一, 九月 19, 2011 22:21 To: Fred Liu Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] remove wrongly added device from zpool On Sep 19, 2011, at 12:10 AM, Fred Liu fred_...@issi.com wrote: Hi, For my carelessness, I added two disks into a raid-z2 zpool as normal data disk, but in fact I want to make them as zil devices. You don't mention which OS you are using, but for the past 5 years of [Open]Solaris releases, the system prints a warning message and will not allow this to occur without using the force option (-f). -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
You don't mention which OS you are using, but for the past 5 years of [Open]Solaris releases, the system prints a warning message and will not allow this to occur without using the force option (-f). -- richard Yes. There is a warning message, I used zpool add -f. Thanks. Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
I get some good progress like following: zpool import pool: cn03 id: 1907858070511204110 state: UNAVAIL status: One or more devices are missing from the system. action: The pool cannot be imported. Attach the missing devices and try again. see: http://www.sun.com/msg/ZFS-8000-6X config: cn03 UNAVAIL missing device raidz2-0 ONLINE c4t5000C5000970B70Bd0 ONLINE c4t5000C5000972C693d0 ONLINE c4t5000C500097009DBd0 ONLINE c4t5000C500097040BFd0 ONLINE c4t5000C5000970727Fd0 ONLINE c4t5000C50009707487d0 ONLINE c4t5000C50009724377d0 ONLINE c4t5000C50039F0B447d0 ONLINE c22t3d0 ONLINE c4t50015179591C238Fd0ONLINE logs c22t4d0 ONLINE c22t5d0 ONLINE Additional devices are known to be part of this pool, though their exact configuration cannot be determined. Any suggestions? Thanks. Fred -Original Message- From: Fred Liu Sent: 星期一, 九月 19, 2011 22:28 To: 'Richard Elling' Cc: zfs-discuss@opensolaris.org Subject: RE: [zfs-discuss] remove wrongly added device from zpool You don't mention which OS you are using, but for the past 5 years of [Open]Solaris releases, the system prints a warning message and will not allow this to occur without using the force option (-f). -- richard Yes. There is a warning message, I used zpool add -f. Thanks. Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
On Sep 19, 2011, at 9:16 AM, Fred Liu wrote: For each disk, look at the output of zdb -l /dev/rdsk/DISKNAMEs0. 1. Confirm that each disk provides 4 labels. 2. Build the vdev tree by hand and look to see which disk is missing This can be tedious and time consuming. Do I need to export the pool first? No, but your pool is not imported. Can you give more details about #2 -- Build the vdev tree by hand and look to see which disk is missing? The label, as displayed by zdb -l contains the heirarchy of the expected pool config. The contents are used to build the output you see in the zpool import or zpool status commands. zpool is complaining that it cannot find one of these disks, so look at the labels on the disks to determine what is or is not missing. The next steps depend on this knowledge. -- richard -- ZFS and performance consulting http://www.RichardElling.com VMworld Copenhagen, October 17-20 OpenStorage Summit, San Jose, CA, October 24-27 LISA '11, Boston, MA, December 4-9 ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
No, but your pool is not imported. YES. I see. and look to see which disk is missing? The label, as displayed by zdb -l contains the heirarchy of the expected pool config. The contents are used to build the output you see in the zpool import or zpool status commands. zpool is complaining that it cannot find one of these disks, so look at the labels on the disks to determine what is or is not missing. The next steps depend on this knowledge. zdb -l /dev/rdsk/c22t2d0s0 cannot open '/dev/rdsk/c22t2d0s0': I/O error root@cn03:~# zdb -l /dev/rdsk/c22t3d0s0 LABEL 0 version: 22 name: 'cn03' state: 0 txg: 18269872 pool_guid: 1907858070511204110 hostid: 13564652 hostname: 'cn03' top_guid: 11074483144412112931 guid: 11074483144412112931 vdev_children: 6 vdev_tree: type: 'disk' id: 1 guid: 11074483144412112931 path: '/dev/dsk/c22t3d0s0' devid: 'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a' phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a' whole_disk: 1 metaslab_array: 37414 metaslab_shift: 24 ashift: 9 asize: 1895563264 is_log: 0 create_txg: 18269863 LABEL 1 version: 22 name: 'cn03' state: 0 txg: 18269872 pool_guid: 1907858070511204110 hostid: 13564652 hostname: 'cn03' top_guid: 11074483144412112931 guid: 11074483144412112931 vdev_children: 6 vdev_tree: type: 'disk' id: 1 guid: 11074483144412112931 path: '/dev/dsk/c22t3d0s0' devid: 'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a' phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a' whole_disk: 1 metaslab_array: 37414 metaslab_shift: 24 ashift: 9 asize: 1895563264 is_log: 0 create_txg: 18269863 LABEL 1 version: 22 name: 'cn03' state: 0 txg: 18269872 pool_guid: 1907858070511204110 hostid: 13564652 hostname: 'cn03' top_guid: 11074483144412112931 guid: 11074483144412112931 vdev_children: 6 vdev_tree: type: 'disk' id: 1 guid: 11074483144412112931 path: '/dev/dsk/c22t3d0s0' devid: 'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a' phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a' whole_disk: 1 metaslab_array: 37414 metaslab_shift: 24 ashift: 9 asize: 1895563264 is_log: 0 create_txg: 18269863 LABEL 2 failed to unpack label 2 LABEL 3 failed to unpack label 3 c22t2d0 and c22t3d0 are the devices I physically removed and connected back to the server. How can I fix them? Thanks. Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Advice with SSD, ZIL and L2ARC
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 I have a new answer: interaction between dataset encryption and L2ARC and ZIL. 1. I am pretty sure (but not completely sure) that data stored in the ZIL is encrypted, if the destination dataset uses encryption. Can anybody confirm?. 2. What happens with L2ARC?. Since ARC is not encrypted (in RAM), is it encrypted when evicted to L2ARC?. Thanks for your time and attention!. - -- Jesus Cea Avion _/_/ _/_/_/_/_/_/ j...@jcea.es - http://www.jcea.es/ _/_/_/_/ _/_/_/_/ _/_/ jabber / xmpp:j...@jabber.org _/_/_/_/ _/_/_/_/_/ . _/_/ _/_/_/_/ _/_/ _/_/ Things are not so easy _/_/ _/_/_/_/ _/_/_/_/ _/_/ My name is Dump, Core Dump _/_/_/_/_/_/ _/_/ _/_/ El amor es poner tu felicidad en la felicidad de otro - Leibniz -BEGIN PGP SIGNATURE- Version: GnuPG v1.4.10 (GNU/Linux) Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/ iQCVAwUBTnd/zZlgi5GaxT1NAQLYOgQAiEbcrklzj/79u9DysRSb1YuMx2J/FsO2 0qgH7KKAVUY6g7QJ2oWB0jDVwLRqJJVzunx4MMmxc+U0eiLoETRVOnYpy6wRNdme fwC5vl0EW2xTQirQde3OSCyyBaN+mKAq+FSd+IA2jQn6y3MLVWq2ucv3d8B6VQDd 32rYKHe6+70= =tv4G -END PGP SIGNATURE- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Advice with SSD, ZIL and L2ARC
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 On 19/09/11 19:45, Jesus Cea wrote: I have a new answer: interaction between dataset encryption and L2ARC and ZIL. Question, a new question... :) - -- Jesus Cea Avion _/_/ _/_/_/_/_/_/ j...@jcea.es - http://www.jcea.es/ _/_/_/_/ _/_/_/_/ _/_/ jabber / xmpp:j...@jabber.org _/_/_/_/ _/_/_/_/_/ . _/_/ _/_/_/_/ _/_/ _/_/ Things are not so easy _/_/ _/_/_/_/ _/_/_/_/ _/_/ My name is Dump, Core Dump _/_/_/_/_/_/ _/_/ _/_/ El amor es poner tu felicidad en la felicidad de otro - Leibniz -BEGIN PGP SIGNATURE- Version: GnuPG v1.4.10 (GNU/Linux) Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/ iQCVAwUBTneAtplgi5GaxT1NAQJPXwP9EpD4tYoqFvsGGnAv51aNwNif1pcvuf6x swEqopVdZxIVHi5Sw7LXNCI6S4gL1WC0Fc2rSOhPj/2a98ihYQSCqn3g+RWlPiJy 0yHdcKfVkHVB3Wy/ReBBwI4RVZLIkjKXP8YN7yBBbPP9ZmrA9lMBFm7zkex6PLXe H2zgUQD8sa8= =tbUK -END PGP SIGNATURE- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
more below… On Sep 19, 2011, at 9:51 AM, Fred Liu wrote: No, but your pool is not imported. YES. I see. and look to see which disk is missing? The label, as displayed by zdb -l contains the heirarchy of the expected pool config. The contents are used to build the output you see in the zpool import or zpool status commands. zpool is complaining that it cannot find one of these disks, so look at the labels on the disks to determine what is or is not missing. The next steps depend on this knowledge. zdb -l /dev/rdsk/c22t2d0s0 cannot open '/dev/rdsk/c22t2d0s0': I/O error Is this disk supposed to be available? You might need to check the partition table, if one exists, to determine if s0 has a non-zero size. root@cn03:~# zdb -l /dev/rdsk/c22t3d0s0 LABEL 0 version: 22 name: 'cn03' state: 0 txg: 18269872 pool_guid: 1907858070511204110 hostid: 13564652 hostname: 'cn03' top_guid: 11074483144412112931 guid: 11074483144412112931 vdev_children: 6 vdev_tree: type: 'disk' id: 1 guid: 11074483144412112931 path: '/dev/dsk/c22t3d0s0' devid: 'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a' phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a' whole_disk: 1 metaslab_array: 37414 metaslab_shift: 24 ashift: 9 asize: 1895563264 is_log: 0 create_txg: 18269863 LABEL 1 version: 22 name: 'cn03' state: 0 txg: 18269872 pool_guid: 1907858070511204110 hostid: 13564652 hostname: 'cn03' top_guid: 11074483144412112931 guid: 11074483144412112931 vdev_children: 6 vdev_tree: type: 'disk' id: 1 guid: 11074483144412112931 path: '/dev/dsk/c22t3d0s0' devid: 'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a' phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a' whole_disk: 1 metaslab_array: 37414 metaslab_shift: 24 ashift: 9 asize: 1895563264 is_log: 0 create_txg: 18269863 LABEL 1 version: 22 name: 'cn03' state: 0 txg: 18269872 pool_guid: 1907858070511204110 hostid: 13564652 hostname: 'cn03' top_guid: 11074483144412112931 guid: 11074483144412112931 vdev_children: 6 vdev_tree: type: 'disk' id: 1 guid: 11074483144412112931 path: '/dev/dsk/c22t3d0s0' devid: 'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a' phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a' whole_disk: 1 metaslab_array: 37414 metaslab_shift: 24 ashift: 9 asize: 1895563264 is_log: 0 create_txg: 18269863 LABEL 2 failed to unpack label 2 LABEL 3 failed to unpack label 3 This is a bad sign, but can be recoverable, depending on how you got here. zdb is saying that it could not find labels at the end of the disk. Label 2 and label 3 are 256KB each, located at the end of the disk, aligned to 256KB boundary. zpool import is smarter than zdb in these cases, and can often recover from it -- up to the loss of all 4 labels, but you need to make sure that the partition tables look reasonable and haven't changed. c22t2d0 and c22t3d0 are the devices I physically removed and connected back to the server. How can I fix them? Unless I'm mistaken, these are ACARD SSDs that have an optional CF backup. Let's hope that the CF backup worked. -- richard -- ZFS and performance consulting http://www.RichardElling.com VMworld Copenhagen, October 17-20 OpenStorage Summit, San Jose, CA, October 24-27 LISA '11, Boston, MA, December 4-9 ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
-Original Message- From: Richard Elling [mailto:richard.ell...@gmail.com] Sent: 星期二, 九月 20, 2011 3:57 To: Fred Liu Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] remove wrongly added device from zpool more below… On Sep 19, 2011, at 9:51 AM, Fred Liu wrote: Is this disk supposed to be available? You might need to check the partition table, if one exists, to determine if s0 has a non-zero size. Yes. I use format to write an EFI label to it. Now this error is gone. But all four label are failed to unpack under zdb -l now. This is a bad sign, but can be recoverable, depending on how you got here. zdb is saying that it could not find labels at the end of the disk. Label 2 and label 3 are 256KB each, located at the end of the disk, aligned to 256KB boundary. zpool import is smarter than zdb in these cases, and can often recover from it -- up to the loss of all 4 labels, but you need to make sure that the partition tables look reasonable and haven't changed. I have tried zpool import -fFX cn03. But it will do core-dump and reboot about 1 hour later. Unless I'm mistaken, these are ACARD SSDs that have an optional CF backup. Let's hope that the CF backup worked. Yes. It is ACARD. You mean push the restore from CF button to see what will happen? Thanks for your nice help! Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] remove wrongly added device from zpool
zdb -l /dev/rdsk/c22t2d0s0 LABEL 0 failed to unpack label 0 LABEL 1 failed to unpack label 1 LABEL 2 failed to unpack label 2 LABEL 3 failed to unpack label 3 -Original Message- From: Fred Liu Sent: 星期二, 九月 20, 2011 4:06 To: 'Richard Elling' Cc: zfs-discuss@opensolaris.org Subject: RE: [zfs-discuss] remove wrongly added device from zpool -Original Message- From: Richard Elling [mailto:richard.ell...@gmail.com] Sent: 星期二, 九月 20, 2011 3:57 To: Fred Liu Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] remove wrongly added device from zpool more below… On Sep 19, 2011, at 9:51 AM, Fred Liu wrote: Is this disk supposed to be available? You might need to check the partition table, if one exists, to determine if s0 has a non-zero size. Yes. I use format to write an EFI label to it. Now this error is gone. But all four label are failed to unpack under zdb -l now. This is a bad sign, but can be recoverable, depending on how you got here. zdb is saying that it could not find labels at the end of the disk. Label 2 and label 3 are 256KB each, located at the end of the disk, aligned to 256KB boundary. zpool import is smarter than zdb in these cases, and can often recover from it -- up to the loss of all 4 labels, but you need to make sure that the partition tables look reasonable and haven't changed. I have tried zpool import -fFX cn03. But it will do core-dump and reboot about 1 hour later. Unless I'm mistaken, these are ACARD SSDs that have an optional CF backup. Let's hope that the CF backup worked. Yes. It is ACARD. You mean push the restore from CF button to see what will happen? Thanks for your nice help! Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] all the history
Hi, I did this: 1): prtvtoc /dev/rdsk/c22t3d0s0 | fmthard -s - /dev/rdsk/c22t2d0s0 2): zpool import cn03 3): zpool status pool: cn03 state: DEGRADED status: One or more devices has experienced an error resulting in data corruption. Applications may be affected. action: Restore the file in question if possible. Otherwise restore the entire pool from backup. see: http://www.sun.com/msg/ZFS-8000-8A scrub: none requested config: NAME STATE READ WRITE CKSUM cn03 DEGRADED 0 053 raidz2-0 ONLINE 0 0 0 c4t5000C5000970B70Bd0 ONLINE 0 0 0 c4t5000C5000972C693d0 ONLINE 0 0 0 c4t5000C500097009DBd0 ONLINE 0 0 0 c4t5000C500097040BFd0 ONLINE 0 0 0 c4t5000C5000970727Fd0 ONLINE 0 0 0 c4t5000C50009707487d0 ONLINE 0 0 0 c4t5000C50009724377d0 ONLINE 0 0 0 c4t5000C50039F0B447d0 ONLINE 0 0 0 c22t3d0 DEGRADED 0 0 120 too many errors c22t2d0 DEGRADED 0 040 too many errors c4t50015179591C238Fd0ONLINE 0 0 0 logs c22t4d0 ONLINE 0 0 0 c22t5d0 ONLINE 0 0 0 spares c4t5000C5003AC39D5Fd0UNAVAIL cannot open errors: 1 data errors, use '-v' for a list pool: rpool state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM rpool ONLINE 0 0 0 c4t500151795910D221d0s0 ONLINE 0 0 0 errors: No known data errors Thanks. Fred From: Fred Liu Sent: 星期二, 九月 20, 2011 9:23 To: Tony Kim; 'Richard Elling' Subject: all the history Hi, Following is the history: The whole history is I found a ZIL device offline at about 2:00PM today from syslog. And I removed and replaced it with a backup device. But I mis-typed the adding command like �C “zpool add cn03 c22t2d0” and correct command should be “ zpool add cn03 log c22t2d0”. The ZIL was wrongly added as a data device in cn03 pool. I noticed it so I physically removed this device from server as I think raidz2 can afford this. The commands I issued are “zpool add cn03 c22t2d0” and “zpool add �Cf cn03 c22t3d0” But the tragedy is coming, the cn03 whole pool is corrupted now: zpool import pool: cn03 id: 1907858070511204110 state: UNAVAIL status: One or more devices are missing from the system. action: The pool cannot be imported. Attach the missing devices and try again. see: http://www.sun.com/msg/ZFS-8000-6X config: cn03 UNAVAIL missing device raidz2-0 ONLINE c4t5000C5000970B70Bd0 ONLINE c4t5000C5000972C693d0 ONLINE c4t5000C500097009DBd0 ONLINE c4t5000C500097040BFd0 ONLINE c4t5000C5000970727Fd0 ONLINE c4t5000C50009707487d0 ONLINE c4t5000C50009724377d0 ONLINE c4t5000C50039F0B447d0 ONLINE c22t3d0 ONLINE c4t50015179591C238Fd0ONLINE logs c22t4d0 ONLINE c22t5d0 ONLINE Additional devices are known to be part of this pool, though their exact configuration cannot be determined. Thanks. Fred ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Advice with SSD, ZIL and L2ARC
On 9/19/11 11:45 AM, Jesus Cea wrote: -BEGIN PGP SIGNED MESSAGE- Hash: SHA1 I have a new answer: interaction between dataset encryption and L2ARC and ZIL. 1. I am pretty sure (but not completely sure) that data stored in the ZIL is encrypted, if the destination dataset uses encryption. Can anybody confirm?. If the data set (file system/zvol) is encrypted then the user data is also encrypted. The ZIL meta data used to parse blocks and records is kept in the clear (in order to claim the blocks) but the user data is encrypted. Neil. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss