[zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu
Hi,

For my carelessness, I added two disks into a raid-z2 zpool as normal data 
disk, but in fact
I want to make them as zil devices.

Any remedy solutions?


Many thanks.

Fred
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Fred Liu
 
 For my carelessness, I added two disks into a raid-z2 zpool as normal data
 disk, but in fact
 I want to make them as zil devices.

That's a huge bummer, and it's the main reason why device removal has been a
priority request for such a long time...  There is no solution.  You can
only destroy  recreate your pool, or learn to live with it that way.

Sorry...

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu
 
 That's a huge bummer, and it's the main reason why device removal has
 been a
 priority request for such a long time...  There is no solution.  You
 can
 only destroy  recreate your pool, or learn to live with it that way.
 
 Sorry...
 

Yeah, I also realized this when I send out this message. In NetApp, it is so
easy to change raid group size. There is still a long way for zfs to go.
Hope I can see that in the future.

I also did another huge mistake which really brings me into the deep pain.
I physically removed these two added devices for I though raidz2 can afford it.
But now the whole pool corrupts. I don't know where I can go ...
Any help will be tremendously appreciated.

Thanks.

Fred
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Tomas Forsman
On 19 September, 2011 - Fred Liu sent me these 0,9K bytes:

  
  That's a huge bummer, and it's the main reason why device removal has
  been a
  priority request for such a long time...  There is no solution.  You
  can
  only destroy  recreate your pool, or learn to live with it that way.
  
  Sorry...
  
 
 Yeah, I also realized this when I send out this message. In NetApp, it is so
 easy to change raid group size. There is still a long way for zfs to go.
 Hope I can see that in the future.
 
 I also did another huge mistake which really brings me into the deep pain.
 I physically removed these two added devices for I though raidz2 can afford 
 it.
 But now the whole pool corrupts. I don't know where I can go ...
 Any help will be tremendously appreciated.

You can add mirrors to those lonely disks.

/Tomas
-- 
Tomas Forsman, st...@acc.umu.se, http://www.acc.umu.se/~stric/
|- Student at Computing Science, University of Umeå
`- Sysadmin at {cs,acc}.umu.se
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Edward Ned Harvey
 From: Fred Liu [mailto:fred_...@issi.com]
 
 Yeah, I also realized this when I send out this message. In NetApp, it is
so
 easy to change raid group size. There is still a long way for zfs to go.
 Hope I can see that in the future.

This one missing feature of ZFS, IMHO, does not result in a long way for
zfs to go in relation to netapp.  I shut off my netapp 2 years ago in favor
of ZFS, because ZFS performs so darn much better, and has such immensely
greater robustness.  Try doing ndmp, cifs, nfs, iscsi on netapp (all extra
licenses).  Try experimenting with the new version of netapp to see how good
it is (you can't unless you buy a whole new box.)  Try mirroring a
production box onto a lower-cost secondary backup box (there is no such
thing).  Try storing your backup on disk and rotating your disks offsite.
Try running any normal utilities - iostat, top, wireshark - you can't.
Try backing up with commercial or otherwise modular (agent-based) backup
software.  You can't.  You have to use CIFS/NFS/NDMP.  

Just try finding a public mailing list like this one where you can even so
much as begin such a conversation about netapp...  Been there done that,
it's not even in the same ballpark.

etc etc.  (end rant.)  I hate netapp.


 I also did another huge mistake which really brings me into the deep
pain.
 I physically removed these two added devices for I though raidz2 can
afford
 it.
 But now the whole pool corrupts. I don't know where I can go ...
 Any help will be tremendously appreciated.

Um...

Wanna post your zpool status and cat /etc/release and zpool upgrade

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu
 
 This one missing feature of ZFS, IMHO, does not result in a long way
 for
 zfs to go in relation to netapp.  I shut off my netapp 2 years ago in
 favor
 of ZFS, because ZFS performs so darn much better, and has such
 immensely
 greater robustness.  Try doing ndmp, cifs, nfs, iscsi on netapp (all
 extra
 licenses).  Try experimenting with the new version of netapp to see how
 good
 it is (you can't unless you buy a whole new box.)  Try mirroring a
 production box onto a lower-cost secondary backup box (there is no such
 thing).  Try storing your backup on disk and rotating your disks
 offsite.
 Try running any normal utilities - iostat, top, wireshark - you can't.
 Try backing up with commercial or otherwise modular (agent-based)
 backup
 software.  You can't.  You have to use CIFS/NFS/NDMP.
 
 Just try finding a public mailing list like this one where you can even
 so
 much as begin such a conversation about netapp...  Been there done that,
 it's not even in the same ballpark.
 
 etc etc.  (end rant.)  I hate netapp.
 
 

Yeah, It is  kind of touchy topic, we may discuss more in the future.
I want to focus on how to repair my pool first. ;-(

 
 Um...
 
 Wanna post your zpool status and cat /etc/release and zpool
 upgrade
 

I exported the pool for I want to use zpool import -F to fix it.
But now I get  one or more devices is currently unavailable Destroy 
and re-create the pool from  a backup source.

I use opensolaris b134 and zpool version 22.


Thanks.

Fred
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu
 
 You can add mirrors to those lonely disks.
 

Can it repair the pool?

Thanks.

Fred
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread David Magda
On Mon, September 19, 2011 08:07, Edward Ned Harvey wrote:

 This one missing feature of ZFS, IMHO, does not result in a long way for
 zfs to go in relation to netapp.  I shut off my netapp 2 years ago in
 favor of ZFS, because ZFS performs so darn much better, and has such
 immensely greater robustness.  Try doing ndmp, cifs, nfs, iscsi on netapp
 (all extra licenses).  Try experimenting with the new version of netapp to
 see how good it is (you can't unless you buy a whole new box.)

As another datum, at $WORK we're going to Isilon. Our NetApp is being
retired by the end of the year as it just can't handle the load of HPC. We
also have the regular assortment of web, mail, code repositories, etc.,
VMs that also live on Isilon. We're quite happy, especially with the more
recent Isilon hardware that uses SSDs to store/cache metadata. NFS and
CIFS are quite good, but we haven't really tried their iSCSI stuff yet;
they don't have FC at all.

We also have a bunch of Blue Arc, but find it much more finicky than
Isilon. Perhaps Hitachi will help them stabilize things a bit.

As for experimenting with NetApp, they do have a simulator that you can
run in a VM if you wish (or actual hardware AFAICT).


A bit more on topic, bp* rewrite has been a long-time coming, and AFAICT,
it  won't be in Solaris 11. As it stands, I don't care much about changing
RAID levels, but not being able to remove a mistakenly added device is
something is becoming more and more conspicuous. For better or worse I'm
not doing as much Solaris stuff (esp. with the new Ellison pricing model),
but still pay attention to what's going on, and this (missing) feature is
one of those WTF? things that is the fly in the otherwise very tasty
soup that is ZFS.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Krunal Desai
On Mon, Sep 19, 2011 at 9:29 AM, Fred Liu fred_...@issi.com wrote:
 Yes. I have connected them back to server. But it does not help.
 I am really sad now...

I cringed a little when I read the thread title. I did this on
accident once as well, but lucky for me, I had enough scratch
storage around in various sizes to cobble together a JBOD (risky) and
use it as a holding area for my data while I remade the pool.

I'm a home user and only have around 21TB or so, so it was feasible
for me. Probably not so feasible for you enterprise guys with 1000s of
users and 100s of filesystems!

--khd
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Edward Ned Harvey
 From: Krunal Desai [mailto:mov...@gmail.com]
 
 On Mon, Sep 19, 2011 at 9:29 AM, Fred Liu fred_...@issi.com wrote:
  Yes. I have connected them back to server. But it does not help.
  I am really sad now...

I'll tell you what does not help.  This email.  Now that you know what you're 
trying to do, why don't you post the results of your zpool import command?  
How about an error message, and how you're trying to go about fixing your pool? 
 Nobody here can help you without information.


 I cringed a little when I read the thread title. I did this on
 accident once as well, but lucky for me, I had enough scratch
 storage around in various sizes to cobble together a JBOD (risky) and
 use it as a holding area for my data while I remade the pool.
 
 I'm a home user and only have around 21TB or so, so it was feasible
 for me. Probably not so feasible for you enterprise guys with 1000s of
 users and 100s of filesystems!

No enterprise guys with 1000s of users and 100s of filesystems are making this 
mistake.  Even if it does happen, on a pool that significant, the obvious 
response is to add redundancy instead of recreating the pool.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu
 
 I'll tell you what does not help.  This email.  Now that you know what
 you're trying to do, why don't you post the results of your zpool
 import command?  How about an error message, and how you're trying to
 go about fixing your pool?  Nobody here can help you without
 information.
 
 
User tty   login@  idle   JCPU   PCPU  what 
root console   9:25pm  w
root@cn03:~# df 
Filesystem   1K-blocks  Used Available Use% Mounted on  
rpool/ROOT/opensolaris  
  94109412   6880699  87228713   8% /   
swap 108497952   344 108497608   1% /etc/svc/volatile   
/usr/lib/libc/libc_hwcap1.so.1  
  94109412   6880699  87228713   8% /lib/libc.so.1  
swap 108497616 8 108497608   1% /tmp
swap 10849768880 108497608   1% /var/run
rpool/export 4686423 46841   1% /export 
rpool/export/home4686423 46841   1% /export/home
rpool/export/home/fred  
 48710  5300 43410  11% /export/home/fred   
rpool10215515880 102155078   1% /rpool  
root@cn03:~# !z 
zpool import cn03   
cannot import 'cn03': one or more devices is currently unavailable  
Destroy and re-create the pool from 
a backup source.

Thanks.

Fred  



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu
I also used zpool import -fFX cn03 in b134 and b151a(via live SX11 live cd). It 
resulted a core dump and reboot after about 15 min.
I can see all the leds are blinking on the HDD within  this 15 min.
Can replacing  empty ZIL devices help?

Thanks.

Fred
 -Original Message-
 From: Fred Liu
 Sent: 星期一, 九月 19, 2011 21:54
 To: 'Edward Ned Harvey'; 'Krunal Desai'
 Cc: zfs-discuss@opensolaris.org
 Subject: RE: [zfs-discuss] remove wrongly added device from zpool
 
 
  I'll tell you what does not help.  This email.  Now that you know
 what
  you're trying to do, why don't you post the results of your zpool
  import command?  How about an error message, and how you're trying
 to
  go about fixing your pool?  Nobody here can help you without
  information.
 
 
 User tty   login@  idle   JCPU   PCPU  what
 root console   9:25pm  w
 root@cn03:~# df
 Filesystem   1K-blocks  Used Available Use% Mounted on
 rpool/ROOT/opensolaris
   94109412   6880699  87228713   8% /
 swap 108497952   344 108497608   1%
 /etc/svc/volatile
 /usr/lib/libc/libc_hwcap1.so.1
   94109412   6880699  87228713   8% /lib/libc.so.1
 swap 108497616 8 108497608   1% /tmp
 swap 10849768880 108497608   1% /var/run
 rpool/export 4686423 46841   1% /export
 rpool/export/home4686423 46841   1% /export/home
 rpool/export/home/fred
  48710  5300 43410  11%
 /export/home/fred
 rpool10215515880 102155078   1% /rpool
 root@cn03:~# !z
 zpool import cn03
 cannot import 'cn03': one or more devices is currently unavailable
 Destroy and re-create the pool from
 a backup source.
 
 Thanks.
 
 Fred
 
 
 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu
The core dump:

r10: ff19a5592000 r11:0 r12:0   
r13:0 r14:0 r15: ff00ba4a5c60   
fsb: fd7fff172a00 gsb: ff19a5592000  ds:0   
 es:0  fs:0  gs:0   
trp:e err:0 rip: f782f81a   
 cs:   30 rfl:10246 rsp: ff00b9bf0a40   
 ss:   38   

ff00b9bf0830 unix:die+10f ()
ff00b9bf0940 unix:trap+177b ()  
ff00b9bf0950 unix:cmntrap+e6 () 
ff00b9bf0ab0 procfs:prchoose+72 ()  
ff00b9bf0b00 procfs:prgetpsinfo+2b ()   
ff00b9bf0ce0 procfs:pr_read_psinfo+4e ()
ff00b9bf0d30 procfs:prread+72 ()
ff00b9bf0da0 genunix:fop_read+6b () 
ff00b9bf0f00 genunix:pread+22c ()   
ff00b9bf0f10 unix:brand_sys_syscall+20d ()  

syncing file systems... done
dumping to /dev/zvol/dsk/rpool/dump, offset 65536, content: kernel  
 0:17 100% done 
100% done: 1041082 pages dumped, dump succeeded 
rebooting...

 -Original Message-
 From: Fred Liu
 Sent: 星期一, 九月 19, 2011 22:00
 To: Fred Liu; 'Edward Ned Harvey'; 'Krunal Desai'
 Cc: 'zfs-discuss@opensolaris.org'
 Subject: RE: [zfs-discuss] remove wrongly added device from zpool
 
 I also used zpool import -fFX cn03 in b134 and b151a(via live SX11 live
 cd). It resulted a core dump and reboot after about 15 min.
 I can see all the leds are blinking on the HDD within  this 15 min.
 Can replacing  empty ZIL devices help?
 
 Thanks.
 
 Fred
  -Original Message-
  From: Fred Liu
  Sent: 星期一, 九月 19, 2011 21:54
  To: 'Edward Ned Harvey'; 'Krunal Desai'
  Cc: zfs-discuss@opensolaris.org
  Subject: RE: [zfs-discuss] remove wrongly added device from zpool
 
  
   I'll tell you what does not help.  This email.  Now that you know
  what
   you're trying to do, why don't you post the results of your zpool
   import command?  How about an error message, and how you're trying
  to
   go about fixing your pool?  Nobody here can help you without
   information.
  
  
  User tty   login@  idle   JCPU   PCPU  what
  root console   9:25pm  w
  root@cn03:~# df
  Filesystem   1K-blocks  Used Available Use% Mounted on
  rpool/ROOT/opensolaris
94109412   6880699  87228713   8% /
  swap 108497952   344 108497608   1%
  /etc/svc/volatile
  /usr/lib/libc/libc_hwcap1.so.1
94109412   6880699  87228713   8%
 /lib/libc.so.1
  swap 108497616 8 108497608   1% /tmp
  swap 10849768880 108497608   1% /var/run
  rpool/export 4686423 46841   1% /export
  rpool/export/home4686423 46841   1% /export/home
  rpool/export/home/fred
   48710  5300 43410  11%
  /export/home/fred
  rpool10215515880 102155078   1% /rpool
  root@cn03:~# !z
  zpool import cn03
  cannot import 'cn03': one or more devices is currently unavailable
  Destroy and re-create the pool from
  a backup source.
 
  Thanks.
 
  Fred
 
 
 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu
I use opensolaris b134.

Thanks.

Fred

 -Original Message-
 From: Richard Elling [mailto:richard.ell...@gmail.com]
 Sent: 星期一, 九月 19, 2011 22:21
 To: Fred Liu
 Cc: zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] remove wrongly added device from zpool
 
 On Sep 19, 2011, at 12:10 AM, Fred Liu fred_...@issi.com wrote:
 
  Hi,
 
  For my carelessness, I added two disks into a raid-z2 zpool as normal
 data disk, but in fact
  I want to make them as zil devices.
 
 You don't mention which OS you are using, but for the past 5 years of
 [Open]Solaris
 releases, the system prints a warning message and will not allow this
 to occur
 without using the force option (-f).
   -- richard
 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu

 
 You don't mention which OS you are using, but for the past 5 years of
 [Open]Solaris
 releases, the system prints a warning message and will not allow this
 to occur
 without using the force option (-f).
   -- richard
 
 Yes. There is a warning message, I used zpool add -f.

Thanks.

Fred
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu
I get some good progress like following:

zpool import
  pool: cn03
id: 1907858070511204110
 state: UNAVAIL
status: One or more devices are missing from the system.
action: The pool cannot be imported. Attach the missing
devices and try again.
   see: http://www.sun.com/msg/ZFS-8000-6X
config:

cn03   UNAVAIL  missing device
  raidz2-0 ONLINE
c4t5000C5000970B70Bd0  ONLINE
c4t5000C5000972C693d0  ONLINE
c4t5000C500097009DBd0  ONLINE
c4t5000C500097040BFd0  ONLINE
c4t5000C5000970727Fd0  ONLINE
c4t5000C50009707487d0  ONLINE
c4t5000C50009724377d0  ONLINE
c4t5000C50039F0B447d0  ONLINE
  c22t3d0  ONLINE
  c4t50015179591C238Fd0ONLINE
logs
  c22t4d0  ONLINE
  c22t5d0  ONLINE

Additional devices are known to be part of this pool, though their
exact configuration cannot be determined.

Any suggestions?

Thanks.

Fred

 -Original Message-
 From: Fred Liu
 Sent: 星期一, 九月 19, 2011 22:28
 To: 'Richard Elling'
 Cc: zfs-discuss@opensolaris.org
 Subject: RE: [zfs-discuss] remove wrongly added device from zpool
 
 
 
  You don't mention which OS you are using, but for the past 5 years of
  [Open]Solaris
  releases, the system prints a warning message and will not allow this
  to occur
  without using the force option (-f).
-- richard
 
  Yes. There is a warning message, I used zpool add -f.
 
 Thanks.
 
 Fred
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Richard Elling
On Sep 19, 2011, at 9:16 AM, Fred Liu wrote:
 
 For each disk, look at the output of zdb -l /dev/rdsk/DISKNAMEs0.
 1. Confirm that each disk provides 4 labels.
 2. Build the vdev tree by hand and look to see which disk is missing
 
 This can be tedious and time consuming.
 
 Do I need to export the pool first?

No, but your pool is not imported.

 Can you give more details about #2 --  Build the vdev tree by hand and look 
 to see which disk is missing?

The label, as displayed by zdb -l contains the heirarchy of the expected pool 
config.
The contents are used to build the output you see in the zpool import or 
zpool status
commands. zpool is complaining that it cannot find one of these disks, so look 
at the
labels on the disks to determine what is or is not missing. The next steps 
depend on
this knowledge.
-- richard

-- 

ZFS and performance consulting
http://www.RichardElling.com
VMworld Copenhagen, October 17-20
OpenStorage Summit, San Jose, CA, October 24-27
LISA '11, Boston, MA, December 4-9 













___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu
 
 No, but your pool is not imported.
 

YES. I see.
 and look to see which disk is missing?
 
 The label, as displayed by zdb -l contains the heirarchy of the
 expected pool config.
 The contents are used to build the output you see in the zpool import
 or zpool status
 commands. zpool is complaining that it cannot find one of these disks,
 so look at the
 labels on the disks to determine what is or is not missing. The next
 steps depend on
 this knowledge.

zdb -l /dev/rdsk/c22t2d0s0
cannot open '/dev/rdsk/c22t2d0s0': I/O error

root@cn03:~# zdb -l /dev/rdsk/c22t3d0s0

LABEL 0

version: 22
name: 'cn03'
state: 0
txg: 18269872
pool_guid: 1907858070511204110
hostid: 13564652
hostname: 'cn03'
top_guid: 11074483144412112931
guid: 11074483144412112931
vdev_children: 6
vdev_tree:
type: 'disk'
id: 1
guid: 11074483144412112931
path: '/dev/dsk/c22t3d0s0'
devid: 
'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a'
phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a'
whole_disk: 1
metaslab_array: 37414
metaslab_shift: 24
ashift: 9
asize: 1895563264
is_log: 0
create_txg: 18269863

LABEL 1

version: 22
name: 'cn03'
state: 0
txg: 18269872
pool_guid: 1907858070511204110
hostid: 13564652
hostname: 'cn03'
top_guid: 11074483144412112931
guid: 11074483144412112931
vdev_children: 6
vdev_tree:
type: 'disk'
id: 1
guid: 11074483144412112931
path: '/dev/dsk/c22t3d0s0'
devid: 
'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a'
phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a'
whole_disk: 1
metaslab_array: 37414
metaslab_shift: 24
ashift: 9
asize: 1895563264
is_log: 0
create_txg: 18269863

LABEL 1

version: 22
name: 'cn03'
state: 0
txg: 18269872
pool_guid: 1907858070511204110
hostid: 13564652
hostname: 'cn03'
top_guid: 11074483144412112931
guid: 11074483144412112931
vdev_children: 6
vdev_tree:
type: 'disk'
id: 1
guid: 11074483144412112931
path: '/dev/dsk/c22t3d0s0'
devid: 
'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a'
phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a'
whole_disk: 1
metaslab_array: 37414
metaslab_shift: 24
ashift: 9
asize: 1895563264
is_log: 0
create_txg: 18269863

LABEL 2

failed to unpack label 2

LABEL 3

failed to unpack label 3


c22t2d0 and c22t3d0 are the devices I physically removed and connected back to 
the server.
How can I fix them?

Thanks.

Fred
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Advice with SSD, ZIL and L2ARC

2011-09-19 Thread Jesus Cea
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

I have a new answer: interaction between dataset encryption and L2ARC
and ZIL.

1. I am pretty sure (but not completely sure) that data stored in the
ZIL is encrypted, if the destination dataset uses encryption. Can
anybody confirm?.

2. What happens with L2ARC?. Since ARC is not encrypted (in RAM), is
it encrypted when evicted to L2ARC?.

Thanks for your time and attention!.

- -- 
Jesus Cea Avion _/_/  _/_/_/_/_/_/
j...@jcea.es - http://www.jcea.es/ _/_/_/_/  _/_/_/_/  _/_/
jabber / xmpp:j...@jabber.org _/_/_/_/  _/_/_/_/_/
.  _/_/  _/_/_/_/  _/_/  _/_/
Things are not so easy  _/_/  _/_/_/_/  _/_/_/_/  _/_/
My name is Dump, Core Dump   _/_/_/_/_/_/  _/_/  _/_/
El amor es poner tu felicidad en la felicidad de otro - Leibniz
-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.10 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/

iQCVAwUBTnd/zZlgi5GaxT1NAQLYOgQAiEbcrklzj/79u9DysRSb1YuMx2J/FsO2
0qgH7KKAVUY6g7QJ2oWB0jDVwLRqJJVzunx4MMmxc+U0eiLoETRVOnYpy6wRNdme
fwC5vl0EW2xTQirQde3OSCyyBaN+mKAq+FSd+IA2jQn6y3MLVWq2ucv3d8B6VQDd
32rYKHe6+70=
=tv4G
-END PGP SIGNATURE-
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Advice with SSD, ZIL and L2ARC

2011-09-19 Thread Jesus Cea
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

On 19/09/11 19:45, Jesus Cea wrote:
 I have a new answer: interaction between dataset encryption and
 L2ARC and ZIL.

Question, a new question... :)

- -- 
Jesus Cea Avion _/_/  _/_/_/_/_/_/
j...@jcea.es - http://www.jcea.es/ _/_/_/_/  _/_/_/_/  _/_/
jabber / xmpp:j...@jabber.org _/_/_/_/  _/_/_/_/_/
.  _/_/  _/_/_/_/  _/_/  _/_/
Things are not so easy  _/_/  _/_/_/_/  _/_/_/_/  _/_/
My name is Dump, Core Dump   _/_/_/_/_/_/  _/_/  _/_/
El amor es poner tu felicidad en la felicidad de otro - Leibniz
-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.10 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/

iQCVAwUBTneAtplgi5GaxT1NAQJPXwP9EpD4tYoqFvsGGnAv51aNwNif1pcvuf6x
swEqopVdZxIVHi5Sw7LXNCI6S4gL1WC0Fc2rSOhPj/2a98ihYQSCqn3g+RWlPiJy
0yHdcKfVkHVB3Wy/ReBBwI4RVZLIkjKXP8YN7yBBbPP9ZmrA9lMBFm7zkex6PLXe
H2zgUQD8sa8=
=tbUK
-END PGP SIGNATURE-
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Richard Elling
more below…

On Sep 19, 2011, at 9:51 AM, Fred Liu wrote:

 
 No, but your pool is not imported.
 
 
 YES. I see.
 and look to see which disk is missing?
 
 The label, as displayed by zdb -l contains the heirarchy of the
 expected pool config.
 The contents are used to build the output you see in the zpool import
 or zpool status
 commands. zpool is complaining that it cannot find one of these disks,
 so look at the
 labels on the disks to determine what is or is not missing. The next
 steps depend on
 this knowledge.
 
 zdb -l /dev/rdsk/c22t2d0s0
 cannot open '/dev/rdsk/c22t2d0s0': I/O error

Is this disk supposed to be available?
You might need to check the partition table, if one exists, to determine if
s0 has a non-zero size.

 root@cn03:~# zdb -l /dev/rdsk/c22t3d0s0
 
 LABEL 0
 
version: 22
name: 'cn03'
state: 0
txg: 18269872
pool_guid: 1907858070511204110
hostid: 13564652
hostname: 'cn03'
top_guid: 11074483144412112931
guid: 11074483144412112931
vdev_children: 6
vdev_tree:
type: 'disk'
id: 1
guid: 11074483144412112931
path: '/dev/dsk/c22t3d0s0'
devid: 
 'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a'
phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a'
whole_disk: 1
metaslab_array: 37414
metaslab_shift: 24
ashift: 9
asize: 1895563264
is_log: 0
create_txg: 18269863
 
 LABEL 1
 
version: 22
name: 'cn03'
state: 0
txg: 18269872
pool_guid: 1907858070511204110
hostid: 13564652
hostname: 'cn03'
top_guid: 11074483144412112931
guid: 11074483144412112931
vdev_children: 6
vdev_tree:
type: 'disk'
id: 1
guid: 11074483144412112931
path: '/dev/dsk/c22t3d0s0'
devid: 
 'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a'
phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a'
whole_disk: 1
metaslab_array: 37414
metaslab_shift: 24
ashift: 9
asize: 1895563264
is_log: 0
create_txg: 18269863
 
 LABEL 1
 
version: 22
name: 'cn03'
state: 0
txg: 18269872
pool_guid: 1907858070511204110
hostid: 13564652
hostname: 'cn03'
top_guid: 11074483144412112931
guid: 11074483144412112931
vdev_children: 6
vdev_tree:
type: 'disk'
id: 1
guid: 11074483144412112931
path: '/dev/dsk/c22t3d0s0'
devid: 
 'id1,sd@s4154412020202020414e53393031305f324e4e4e324e4e4e202020202020202035363238363739005f31/a'
phys_path: '/pci@0,0/pci15d9,400@1f,2/disk@3,0:a'
whole_disk: 1
metaslab_array: 37414
metaslab_shift: 24
ashift: 9
asize: 1895563264
is_log: 0
create_txg: 18269863
 
 LABEL 2
 
 failed to unpack label 2
 
 LABEL 3
 
 failed to unpack label 3

This is a bad sign, but can be recoverable, depending on how you got here. zdb 
is saying
that it could not find labels at the end of the disk. Label 2 and label 3 are 
256KB each, located
at the end of the disk, aligned to 256KB boundary. zpool import is smarter than 
zdb in these
cases, and can often recover from it -- up to the loss of all 4 labels, but you 
need to make sure 
that the partition tables look reasonable and haven't changed.

 c22t2d0 and c22t3d0 are the devices I physically removed and connected back 
 to the server.
 How can I fix them?

Unless I'm mistaken, these are ACARD SSDs that have an optional CF backup. 
Let's hope
that the CF backup worked.
 -- richard

-- 

ZFS and performance consulting
http://www.RichardElling.com
VMworld Copenhagen, October 17-20
OpenStorage Summit, San Jose, CA, October 24-27
LISA '11, Boston, MA, December 4-9 













___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu


 -Original Message-
 From: Richard Elling [mailto:richard.ell...@gmail.com]
 Sent: 星期二, 九月 20, 2011 3:57
 To: Fred Liu
 Cc: zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] remove wrongly added device from zpool
 
 more below…
 
 On Sep 19, 2011, at 9:51 AM, Fred Liu wrote:
 
 Is this disk supposed to be available?
 You might need to check the partition table, if one exists, to
 determine if
 s0 has a non-zero size.
 

Yes. I use format to write an EFI label to it. Now this error is gone.
But all four label are failed to unpack under zdb -l now.


 
 This is a bad sign, but can be recoverable, depending on how you got
 here. zdb is saying
 that it could not find labels at the end of the disk. Label 2 and label
 3 are 256KB each, located
 at the end of the disk, aligned to 256KB boundary. zpool import is
 smarter than zdb in these
 cases, and can often recover from it -- up to the loss of all 4 labels,
 but you need to make sure
 that the partition tables look reasonable and haven't changed.
 

I have tried zpool import -fFX cn03. But it will do core-dump and reboot about 
1 hour later.

 
 Unless I'm mistaken, these are ACARD SSDs that have an optional CF
 backup. Let's hope
 that the CF backup worked.

Yes. It is ACARD. You mean push the restore from CF button to see what will 
happen?


Thanks for your nice help!


Fred
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove wrongly added device from zpool

2011-09-19 Thread Fred Liu
zdb -l /dev/rdsk/c22t2d0s0

LABEL 0

failed to unpack label 0

LABEL 1

failed to unpack label 1

LABEL 2

failed to unpack label 2

LABEL 3

failed to unpack label 3

 -Original Message-
 From: Fred Liu
 Sent: 星期二, 九月 20, 2011 4:06
 To: 'Richard Elling'
 Cc: zfs-discuss@opensolaris.org
 Subject: RE: [zfs-discuss] remove wrongly added device from zpool
 
 
 
  -Original Message-
  From: Richard Elling [mailto:richard.ell...@gmail.com]
  Sent: 星期二, 九月 20, 2011 3:57
  To: Fred Liu
  Cc: zfs-discuss@opensolaris.org
  Subject: Re: [zfs-discuss] remove wrongly added device from zpool
 
  more below…
 
  On Sep 19, 2011, at 9:51 AM, Fred Liu wrote:
 
  Is this disk supposed to be available?
  You might need to check the partition table, if one exists, to
  determine if
  s0 has a non-zero size.
 
 
 Yes. I use format to write an EFI label to it. Now this error is gone.
 But all four label are failed to unpack under zdb -l now.
 
 
 
  This is a bad sign, but can be recoverable, depending on how you got
  here. zdb is saying
  that it could not find labels at the end of the disk. Label 2 and
 label
  3 are 256KB each, located
  at the end of the disk, aligned to 256KB boundary. zpool import is
  smarter than zdb in these
  cases, and can often recover from it -- up to the loss of all 4
 labels,
  but you need to make sure
  that the partition tables look reasonable and haven't changed.
 
 
 I have tried zpool import -fFX cn03. But it will do core-dump and
 reboot about 1 hour later.
 
 
  Unless I'm mistaken, these are ACARD SSDs that have an optional CF
  backup. Let's hope
  that the CF backup worked.
 
 Yes. It is ACARD. You mean push the restore from CF button to see
 what will happen?
 
 
 Thanks for your nice help!
 
 
 Fred
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] all the history

2011-09-19 Thread Fred Liu
Hi,

I did this:

1): prtvtoc /dev/rdsk/c22t3d0s0 | fmthard -s - /dev/rdsk/c22t2d0s0
2): zpool import cn03
3): zpool status
  pool: cn03
 state: DEGRADED
status: One or more devices has experienced an error resulting in data
corruption.  Applications may be affected.
action: Restore the file in question if possible.  Otherwise restore the
entire pool from backup.
   see: http://www.sun.com/msg/ZFS-8000-8A
 scrub: none requested
config:

NAME   STATE READ WRITE CKSUM
cn03   DEGRADED 0 053
  raidz2-0 ONLINE   0 0 0
c4t5000C5000970B70Bd0  ONLINE   0 0 0
c4t5000C5000972C693d0  ONLINE   0 0 0
c4t5000C500097009DBd0  ONLINE   0 0 0
c4t5000C500097040BFd0  ONLINE   0 0 0
c4t5000C5000970727Fd0  ONLINE   0 0 0
c4t5000C50009707487d0  ONLINE   0 0 0
c4t5000C50009724377d0  ONLINE   0 0 0
c4t5000C50039F0B447d0  ONLINE   0 0 0
  c22t3d0  DEGRADED 0 0   120  too many errors
  c22t2d0  DEGRADED 0 040  too many errors
  c4t50015179591C238Fd0ONLINE   0 0 0
logs
  c22t4d0  ONLINE   0 0 0
  c22t5d0  ONLINE   0 0 0
spares
  c4t5000C5003AC39D5Fd0UNAVAIL   cannot open

errors: 1 data errors, use '-v' for a list

  pool: rpool
 state: ONLINE
 scrub: none requested
config:

NAME   STATE READ WRITE CKSUM
rpool  ONLINE   0 0 0
  c4t500151795910D221d0s0  ONLINE   0 0 0

errors: No known data errors


Thanks.

Fred

From: Fred Liu
Sent: 星期二, 九月 20, 2011 9:23
To: Tony Kim; 'Richard Elling'
Subject: all the history

Hi,

Following is the history:

The whole history is I found a ZIL device offline at about 2:00PM today from 
syslog.
And I removed and replaced it with a backup device. But I mis-typed the adding
command like �C “zpool add cn03 c22t2d0” and correct command should be “ zpool 
add cn03 log
c22t2d0”.  The ZIL was wrongly added as a data device in cn03 pool. I noticed 
it so I physically
removed this device from server as I think raidz2 can afford this.

The commands I issued are “zpool add cn03 c22t2d0” and “zpool add �Cf cn03 
c22t3d0”

But the tragedy is coming, the cn03 whole pool is corrupted now:

zpool import
  pool: cn03
id: 1907858070511204110
 state: UNAVAIL
status: One or more devices are missing from the system.
action: The pool cannot be imported. Attach the missing
devices and try again.
   see: http://www.sun.com/msg/ZFS-8000-6X
config:

cn03   UNAVAIL  missing device
  raidz2-0 ONLINE
c4t5000C5000970B70Bd0  ONLINE
c4t5000C5000972C693d0  ONLINE
c4t5000C500097009DBd0  ONLINE
c4t5000C500097040BFd0  ONLINE
c4t5000C5000970727Fd0  ONLINE
c4t5000C50009707487d0  ONLINE
c4t5000C50009724377d0  ONLINE
c4t5000C50039F0B447d0  ONLINE
  c22t3d0  ONLINE
  c4t50015179591C238Fd0ONLINE
logs
  c22t4d0  ONLINE
  c22t5d0  ONLINE

Additional devices are known to be part of this pool, though their
exact configuration cannot be determined.

Thanks.

Fred
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Advice with SSD, ZIL and L2ARC

2011-09-19 Thread Neil Perrin

On 9/19/11 11:45 AM, Jesus Cea wrote:

-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

I have a new answer: interaction between dataset encryption and L2ARC
and ZIL.

1. I am pretty sure (but not completely sure) that data stored in the
ZIL is encrypted, if the destination dataset uses encryption. Can
anybody confirm?.



If the data set (file system/zvol) is encrypted then
the user data is also encrypted. The ZIL meta data
used to parse blocks and records is kept in the clear
(in order to claim the blocks) but the user data is
encrypted.

Neil.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss