Re: [zfs-discuss] Zpool import Problem ZFS-8000-EY
OK, the Pool is died and i had create a new one :-) regards ré -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Zpool import Problem ZFS-8000-EY
Hi Folks, i had an ZFS Pool running on an Hard-Raid5 Controller. I had play with the Vendor Maintenance Tool an now my Zpool are in trouble: ZFS-8000-EY r...@opensolaris:~# zpool import pool: data id: 15260857908242801044 state: FAULTED status: The pool was last accessed by another system. action: The pool cannot be imported due to damaged devices or data. The pool may be active on another system, but can be imported using the '-f' flag. see: http://www.sun.com/msg/ZFS-8000-EY config: dataFAULTED corrupted data c4t0d0p0 ONLINE r...@opensolaris:~# zpool import data cannot import 'data': pool may be in use from other system, it was last accessed by file.menny.local (hostid: 0xc0d3b779) on Sun Sep 5 20:40:27 2010 use '-f' to import anyway r...@opensolaris:~# zdb rpool: version: 22 name: 'rpool' state: 0 txg: 4 pool_guid: 964227353010346491 hostid: 756419 hostname: 'opensolaris' vdev_children: 1 vdev_tree: type: 'root' id: 0 guid: 964227353010346491 create_txg: 4 children[0]: type: 'disk' id: 0 guid: 5072264041238200753 path: '/dev/dsk/c5t0d0s0' devid: 'id1,s...@n600508b10010413953535a3038470008/a' phys_path: '/p...@0,0/pci8086,3...@2/pci8086,3...@0,2/pcie11,4...@2/s...@0,0:a' whole_disk: 0 metaslab_array: 23 metaslab_shift: 28 ashift: 9 asize: 36375625728 is_log: 0 create_txg: 4 please can you give me a little bit help to solve this problem? Thanks ré -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Zpool import Problem ZFS-8000-EY
so, i forgot the rest of information: ZFS v3 Zpool v15 Hard-Raid5 with CPQary3 the Pool was create in an early release and update to v15 with FreeBSD 8.1 AMD64, the FreeBSD Forum give me the tip to install OpenSolaris svn_134. Now its running ;-) regards ré -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zpool import problem / missing label / corrupted data
After a power outage due to a thunder storm my 3 disk raidz1 pool has become UNAVAILable. It is a ZFV v13 pool using the whole 3 disks created on FreeBSD current 8 x64 and worked well for over a month. Unfortunately I wasn't able to import the pool with neither a FreeBSD LiveCD or the current OpenSolaris LiveCD x86/x64. When I tried to import the pool with FreeBSD the system just hangs (I waited for 2 days for a response, which didn't come, not even an error message). Browsing the web informed me that this problem might be solved by using a newer ZFS revision. Therefore I gave opensolaris a try, resulting in the following zpool import -f output: pool: tank id: 582994605255991135 state: UNAVAIL status: The pool is formatted using an older on-disk version. action: The pool cannot be imported due to damaged devices or data. config: tank UNAVAILinsufficient replicas raidz1 UNAVAILcorrupted data c8t0d0p0 ONLINE c10d0p0ONLINE c11d0p0ONLINE The OpenSolaris LiveCD (snv_111b) uses ZFS v14, which would explain the status message. zdb -l shows full LABELs on c8t0d0p0, but the other disks are missing LABEL 2 and LABEL 3. c8t0d0p0 is connected via usb and the other disks via onboard SATA. Some people with similar problems where able to solve them by exporting and importing their pools. I can't export the pool, since I need to import it first in order to have a pool available to export, can't destroy it import it with the -D option for the same reason. At least the disks themselves seem to be okay, since dd finished on all disks without errors. Everything I found on the web so far was either not applicable or didn't lead to any results. I hope that you can help me to get at least a fraction of the data back. The 'important data' was the cifs share used by all in-house computers for central storage. There are partial backups of the data, on different disks, but those aren't up to date with the files on the pool. Snapshots have been made, but those are on the pool as well. I planned on adding another disk for the OS and use the raidz for storage only. The new disk arrived a few days after that incident with the power outage. So I am close to just setup the whole system from scratch with the new setup, but I want to give it a last try to recover it with your expertise. After all it would be pretty convenient to add the raidz pool to the fresh install without loosing any user data. Regards, sid -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import problem / missing label / corrupted data
hello, I cannot help for your problem but one of my drives wwas also destroyed by last week-end's storm. and what a storm it was! good luck with your restore. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import problem
Hi Terry Please could you post back to this forum the output from # zdb -l /dev/rdsk/... ... for each of the 5 drives in your raidz2. (maybe best as an attachment) Are you seeing labels with the error 'failed to unpack'? What is the reported 'status' of your zpool? (You have not provided a 'zpool status') Thanks Nigel Smith -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import problem
These are the symptoms of a shrinking device in a RAID-Z pool. You can try to run the attached script during the import to see if this the case. There's a bug filed on this, but I don't have it handy. - Eric On Sun, Oct 26, 2008 at 05:18:25PM -0700, Terry Heatlie wrote: Folks, I have a zpool with a raidz2 configuration which I've been switching between two machines - an old one with a hardware problem and a new one, which doesn't have hardware issues, but has a different configuration . I've been trying to import the pool on the new machine, so I can back up the data, because the old (broken) machine resets (I don't think it's panicking, because there are no logged messages) every time I try to tar off the data from the ZFS. Unfortunately, the first time I tried to import the pool on the new machine, I didn't have the right five drives in it, so it didn't work. After I figured out that I was confused about which was the boot drive, I did get the five drives into the new machine and asked it to import the pool. It said that the pool could not be imported due to damaged devices or data. Which is slightly odd, since it had been mounting the pool fine on the broken machine before. I then moved the drives back into the old machine, figuring I'd at least copy some small stuff onto a USB stick (it only dies reading large files, apparently), but now the old machine can't mount the pool either, and asking it to import gives the same message. It shows all five drives online, but says the pool is UNAVAIL due to insufficient replicas, and the raidz2 is UNAVAIL due to corrupted data. Must I resign myself to having lost this pool due to the hardware problems I've had, and restore such backups as I have on the new machine, or is there something that can be done to get the pool back online at least in degraded mode? Thanks in advance, --Terry. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Eric Schrock, Fishworkshttp://blogs.sun.com/eschrock #!/sbin/dtrace -s #pragma D option quiet BEGIN { printf(run 'zpool import' to generate trace\n\n); } vdev_raidz_open:entry { printf(%d BEGIN RAIDZ OPEN\n, timestamp); printf(%d config asize = %d\n, timestamp, args[0]-vdev_asize); printf(%d config ashift = %d\n, timestamp, args[0]-vdev_top-vdev_ashift); self-child = 1; self-asize = args[1]; self-ashift = args[2]; } vdev_disk_open:entry /self-child/ { self-disk_asize = args[1]; self-disk_ashift = args[2]; } vdev_disk_open:return /self-child/ { printf(%d child[%d]: asize = %d, ashift = %d\n, timestamp, self-child - 1, *self-disk_asize, *self-disk_ashift); self-disk_asize = 0; self-disk_ashift = 0; self-child++; } vdev_raidz_open:return { printf(%d asize = %d\n, timestamp, *self-asize); printf(%d ashift = %d\n, timestamp, *self-ashift); printf(%d END RAIDZ OPEN\n, timestamp); self-child = 0; self-asize = 0; self-ashift = 0; } ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import problem
Eric Schrock пишет: These are the symptoms of a shrinking device in a RAID-Z pool. You can try to run the attached script during the import to see if this the case. There's a bug filed on this, but I don't have it handy. it's 6753869 labeling/shrinking a disk in raid-z vdev makes pool un-importable victor - Eric On Sun, Oct 26, 2008 at 05:18:25PM -0700, Terry Heatlie wrote: Folks, I have a zpool with a raidz2 configuration which I've been switching between two machines - an old one with a hardware problem and a new one, which doesn't have hardware issues, but has a different configuration . I've been trying to import the pool on the new machine, so I can back up the data, because the old (broken) machine resets (I don't think it's panicking, because there are no logged messages) every time I try to tar off the data from the ZFS. Unfortunately, the first time I tried to import the pool on the new machine, I didn't have the right five drives in it, so it didn't work. After I figured out that I was confused about which was the boot drive, I did get the five drives into the new machine and asked it to import the pool. It said that the pool could not be imported due to damaged devices or data. Which is slightly odd, since it had been mounting the pool fine on the broken machine before. I then moved the drives back into the old machine, figuring I'd at least copy some small stuff onto a USB stick (it only dies reading large files, apparently), but now the old machine can't mount the pool either, and asking it to import gives the same message. It shows all five drives online, but says the pool is UNAVAIL due to insufficient replicas, and the raidz2 is UNAVAIL due to corrupted data. Must I resign myself to having lost this pool due to the hardware problems I've had, and restore such backups as I have on the new machine, or is there something that can be done to get the pool back online at least in degraded mode? Thanks in advance, --Terry. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import problem
I recently tried to import a b97 pool into a b98 upgraded version of that os, and it failed because of some bug. So maybe try eliminating that kind of problem by making sure to use the version that you know worked in the past. Maybe you already did this. div id=jive-html-wrapper-div Folks,divbr/divdivI have a zpool with a raidz2 configuration which I#39;ve been switching between two machines - an old one with a hardware problem and a new one, which doesn#39;t have hardware issues, but has a different configuration . nbsp; I#39;ve been trying to import the pool on the new machine, so I can back up the data, because the old (broken) machine resets (I don#39;t think it#39;s panicking, because there are no logged messages) every time I try to tar off the data from the ZFS. nbsp;/div divbr/divdivnbsp;Unfortunately, the first time I tried to import the pool on the new machine, I didn#39;t have the right five drives in it, so it didn#39;t work. nbsp;After I figured out that I was confused about which was the boot drive, I did get the five drives into the new machine and asked it to import the pool. nbsp;It said that the pool could not be imported due to damaged devices or data. nbsp; Which is slightly odd, since it had been mounting the pool fine on the broken machine before. nbsp;nbsp;/div divbr/divdivI then moved the drives back into the old machine, figuring I#39;d at least copy some small stuff onto a USB stick (it only dies reading large files, apparently), but now the old machine can#39;t mount the pool either, and asking it to import gives the same message. nbsp; It shows all five drives online, but says the pool is UNAVAIL due to insufficient replicas, and the nbsp;raidz2 is UNAVAIL due to corrupted data./div divbr/divdivMust I resign myself to having lost this pool due to the hardware problems I#39;ve had, and restore such backups as I have on the new machine, or is there something that can be done to get the pool back online at least in degraded mode?/div divbr/divdivThanks in advance,/divdivbr/divdiv--Terry./div /div___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discu ss -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zpool import problem
Folks, I have a zpool with a raidz2 configuration which I've been switching between two machines - an old one with a hardware problem and a new one, which doesn't have hardware issues, but has a different configuration . I've been trying to import the pool on the new machine, so I can back up the data, because the old (broken) machine resets (I don't think it's panicking, because there are no logged messages) every time I try to tar off the data from the ZFS. Unfortunately, the first time I tried to import the pool on the new machine, I didn't have the right five drives in it, so it didn't work. After I figured out that I was confused about which was the boot drive, I did get the five drives into the new machine and asked it to import the pool. It said that the pool could not be imported due to damaged devices or data. Which is slightly odd, since it had been mounting the pool fine on the broken machine before. I then moved the drives back into the old machine, figuring I'd at least copy some small stuff onto a USB stick (it only dies reading large files, apparently), but now the old machine can't mount the pool either, and asking it to import gives the same message. It shows all five drives online, but says the pool is UNAVAIL due to insufficient replicas, and the raidz2 is UNAVAIL due to corrupted data. Must I resign myself to having lost this pool due to the hardware problems I've had, and restore such backups as I have on the new machine, or is there something that can be done to get the pool back online at least in degraded mode? Thanks in advance, --Terry. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] zpool import problem
Terry Heatlie wrote: Folks, I have a zpool with a raidz2 configuration which I've been switching between two machines - an old one with a hardware problem and a new one, which doesn't have hardware issues, but has a different configuration . I've been trying to import the pool on the new machine, so I can back up the data, because the old (broken) machine resets (I don't think it's panicking, because there are no logged messages) every time I try to tar off the data from the ZFS. Unfortunately, the first time I tried to import the pool on the new machine, I didn't have the right five drives in it, so it didn't work. After I figured out that I was confused about which was the boot drive, I did get the five drives into the new machine and asked it to import the pool. It said that the pool could not be imported due to damaged devices or data. Which is slightly odd, since it had been mounting the pool fine on the broken machine before. I then moved the drives back into the old machine, figuring I'd at least copy some small stuff onto a USB stick (it only dies reading large files, apparently), but now the old machine can't mount the pool either, and asking it to import gives the same message. It shows all five drives online, but says the pool is UNAVAIL due to insufficient replicas, and the raidz2 is UNAVAIL due to corrupted data. Must I resign myself to having lost this pool due to the hardware problems I've had, and restore such backups as I have on the new machine, or is there something that can be done to get the pool back online at least in degraded mode? Note: we're also working on a troubleshooting wiki... need more days in the hour... You should try to read the labels from each device. zdb -l /dev/rdsk/... You should see 4 labels for each proper device. Here is my hypothesis: If you see a device which has only label 0 and 1, then it may be the case that the label has overlapping partitions. Why does this matter? Because under normal circumstances, the actual devices used for creating or importing the pool are stored in the /etc/zfs/zpool.cache file. When the system boots, it looks there first and will import the pools listed therein. When you export the pool, the zpool.cache entries for the pool are removed. If the pool is not in zpool.cache, then zpool import scans all of the devices found in /dev/dsk for valid pools. If you have overlapping partitions or slices, then a partially exposed vdev may be found. But since it won't be complete, due to perhaps not being able to see the end of the device which is where labels 2 3 are located, then it will be marked as bad. The solution would be to reconcile the partions/slices using format. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZPOOL Import Problem
On Sep 16, 2008, at 5:39 PM, Miles Nordin wrote: jd == Jim Dunham [EMAIL PROTECTED] writes: jd If at the time the SNDR replica is deleted the set was jd actively replicating, along with ZFS actively writing to the jd ZFS storage pool, I/O consistency will be lost, leaving ZFS jd storage pool in an indeterministic state on the remote node. jd To address this issue, prior to deleting the replicas, the jd replica should be placed into logging mode first. What if you stop the replication by breaking the network connection between primary and replica? consistent or inconsistent? Consistent. it sounds fishy, like ``we're always-consistent-on-disk with ZFS, but please use 'zpool offline' to avoid disastrous pool corruption.'' This is not the case at all. Maintaining I/O consistency of all volumes in a single I/O consistency group, is an attribute of replication. The instant an SNDR replica is deleted, that volume is no longer being replicated, and it becomes inconsistent with all other write-order volumes. By placing all volumes in the I/O consistency group in logging mode, not 'zpool offline', and then deleting the replica there is no means for any of the remote volumes to become I/O inconsistent. Yes, one will note that there is a group disable command sndradm -g group-name -d, but it was implemented for easy of administration, not for performing a write-order coordinated disable command. jd ndr_ii. This is an automatic snapshot taken before jd resynchronization starts, yeah that sounds fine, possibly better than DRBD in one way because it might allow the resync to go faster. From the PDF's it sounds like async replication isn't done the same way as the resync, it's done safely, and that it's even possible for async replication to accumulate hours of backlog in a ``disk queue'' without losing write ordering so long as you use the ``blocking mode'' variant of async. Correct reading of the documentation. ii might also be good for debugging a corrupt ZFS, so you can tinker with it but still roll back to the original corrupt copy. I'll read about it---I'm guessing I will need to prepare ahead of time if I want ii available in the toolbox after a disaster. jd AVS has the concept of I/O consistency groups, where all disks jd of a multi-volume filesystem (ZFS, QFS) or database (Oracle, jd Sybase) are kept write-order consistent when using either sync jd or async replication. Awesome, so long as people know to use it. so I guess that's the answer for the OP: use consistency groups! I use the name of the ZFS storage pool, as the name of the SNDR I/O consistency group. The one thing I worry about is, before, AVS was used between RAID and filesystem, which is impossible now because that inter-layer area n olonger exists. If you put the individual device members of a redundant zpool vdev into an AVS consistency group, what will AVS do when one of the devices fails? Nothing, as it is ZFS the reacts to the failed device Does it continue replicating the working devices and ignore the failed one? In this scenario ZFS knows he device failed, which means ZFS will stop writing to the disk, and thus the replica. This would sacrifice redundancy at the DR site. UFS-AVS-RAID would not do that in the same situation. Or hide the failed device from ZFS and slow things down by sending all read/writes of the failed device to the remote mirror? This would slwo down the primary site. UFS-AVS-RAID would not do that in the same situation. The latter ZFS-AVS behavior might be rescueable, if ZFS had the statistical read-preference feature. but writes would still be massively slowed with this scenario, while in UFS-AVS-RAID they would not be. To get back the level of control one used to have for writes, you'd need a different zpool-level way to achieve the intent of the AVS sync/async option. Maybe just a slog which is not AVS-replicated would be enough, modulo other ZFS fixes for hiding slow devices. ZFS-AVS is not UFS-AVS-RAID, and although one can foresee some downside to replicating ZFS with AVS, there are some big wins. Place SNDR in logging mode, and zpool scrub the secondary volumes for consistency, then resume replication. Compressed ZFS Storage pools, result in compressed replication Encrypted ZFS Storage pools, result in encrypted replication ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss Jim Dunham Engineering Manager Storage Platform Software Group Sun Microsystems, Inc. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZPOOL Import Problem
jd == Jim Dunham [EMAIL PROTECTED] writes: jd If at the time the SNDR replica is deleted the set was jd actively replicating, along with ZFS actively writing to the jd ZFS storage pool, I/O consistency will be lost, leaving ZFS jd storage pool in an indeterministic state on the remote node. jd To address this issue, prior to deleting the replicas, the jd replica should be placed into logging mode first. What if you stop the replication by breaking the network connection between primary and replica? consistent or inconsistent? it sounds fishy, like ``we're always-consistent-on-disk with ZFS, but please use 'zpool offline' to avoid disastrous pool corruption.'' jd ndr_ii. This is an automatic snapshot taken before jd resynchronization starts, yeah that sounds fine, possibly better than DRBD in one way because it might allow the resync to go faster. From the PDF's it sounds like async replication isn't done the same way as the resync, it's done safely, and that it's even possible for async replication to accumulate hours of backlog in a ``disk queue'' without losing write ordering so long as you use the ``blocking mode'' variant of async. ii might also be good for debugging a corrupt ZFS, so you can tinker with it but still roll back to the original corrupt copy. I'll read about it---I'm guessing I will need to prepare ahead of time if I want ii available in the toolbox after a disaster. jd AVS has the concept of I/O consistency groups, where all disks jd of a multi-volume filesystem (ZFS, QFS) or database (Oracle, jd Sybase) are kept write-order consistent when using either sync jd or async replication. Awesome, so long as people know to use it. so I guess that's the answer for the OP: use consistency groups! The one thing I worry about is, before, AVS was used between RAID and filesystem, which is impossible now because that inter-layer area n olonger exists. If you put the individual device members of a redundant zpool vdev into an AVS consistency group, what will AVS do when one of the devices fails? Does it continue replicating the working devices and ignore the failed one? This would sacrifice redundancy at the DR site. UFS-AVS-RAID would not do that in the same situation. Or hide the failed device from ZFS and slow things down by sending all read/writes of the failed device to the remote mirror? This would slwo down the primary site. UFS-AVS-RAID would not do that in the same situation. The latter ZFS-AVS behavior might be rescueable, if ZFS had the statistical read-preference feature. but writes would still be massively slowed with this scenario, while in UFS-AVS-RAID they would not be. To get back the level of control one used to have for writes, you'd need a different zpool-level way to achieve the intent of the AVS sync/async option. Maybe just a slog which is not AVS-replicated would be enough, modulo other ZFS fixes for hiding slow devices. pgpzm3T09CxRc.pgp Description: PGP signature ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZPOOL Import Problem
Corey -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] Sent: Friday, September 12, 2008 4:34 PM To: Leopold, Corey Cc: zfs-discuss@opensolaris.org; [EMAIL PROTECTED] Subject: Re: [zfs-discuss] ZPOOL Import Problem Corey, I ran into an odd problem importing a zpool while testing avs. I was trying to simulate a drive failure, break SNDR replication, and then import the pool on the secondary. To simulate the drive failure is just offlined one of the disks in the RAIDZ set. Are all constituent volumes of a single ZFS storage pool in the same SNDR I/O consistency group? Yes they were. One thing to note that shows some of my unfamiliarity with SNDR is that I actually deleted the replication set config before trying to mount on the secondary sndradm -g group -nd Since then I have been throwing them into logging mode sndradm -g group -nl and haven't had a problem in similar tests. (i.e. offlining a drive before making the secondary active). I don't believe that deleting the SNDR replication configuration should have made the ZPOOL invalid though. So there may still be a bug somewhere. If at the time the SNDR replica is deleted the set was actively replicating, along with ZFS actively writing to the ZFS storage pool, I/O consistency will be lost, leaving ZFS storage pool in an indeterministic state on the remote node. To address this issue, prior to deleting the replicas, the replica should be placed into logging mode first. Then ZFS will be left in I/O consistent after the disable is done. Corey Jim Dunham Engineering Manager Storage Platform Software Group Sun Microsystems, Inc. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZPOOL Import Problem
Corey, I ran into an odd problem importing a zpool while testing avs. I was trying to simulate a drive failure, break SNDR replication, and then import the pool on the secondary. To simulate the drive failure is just offlined one of the disks in the RAIDZ set. Are all constituent volumes of a single ZFS storage pool in the same SNDR I/O consistency group? -- pr1# zpool status pool: rpool state: ONLINE scrub: none requested config: NAMESTATE READ WRITE CKSUM rpool ONLINE 0 0 0 c3t0d0s0 ONLINE 0 0 0 errors: No known data errors pool: tank state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM tank ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c5t0d0s0 ONLINE 0 0 0 c5t1d0s0 ONLINE 0 0 0 c5t2d0s0 ONLINE 0 0 0 c5t3d0s0 ONLINE 0 0 0 errors: No known data errors pr1# zpool offline missing pool name usage: offline [-t] pool device ... pr1# zpool offline tank c5t0d0s0 pr1# zpool status pool: rpool state: ONLINE scrub: none requested config: NAMESTATE READ WRITE CKSUM rpool ONLINE 0 0 0 c3t0d0s0 ONLINE 0 0 0 errors: No known data errors pool: tank state: DEGRADED status: One or more devices has been taken offline by the administrator. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Online the device using 'zpool online' or replace the device with 'zpool replace'. scrub: none requested config: NAME STATE READ WRITE CKSUM tank DEGRADED 0 0 0 raidz1 DEGRADED 0 0 0 c5t0d0s0 OFFLINE 0 0 0 c5t1d0s0 ONLINE 0 0 0 c5t2d0s0 ONLINE 0 0 0 c5t3d0s0 ONLINE 0 0 0 errors: No known data errors pr1# zpool export tank --- I then disabled SNDR replication. pr1# sndradm -g zfs-tank -d Disable Remote Mirror? (Y/N) [N]: Y - Then I try to import the ZPOOL on the secondary. -- pr2# zpool import pool: tank id: 9795707198744908806 state: DEGRADED status: One or more devices are offlined. action: The pool can be imported despite missing or damaged devices. The fault tolerance of the pool may be compromised if imported. config: tank DEGRADED raidz1 DEGRADED c5t0d0s0 OFFLINE c5t1d0s0 ONLINE c5t2d0s0 ONLINE c5t3d0s0 ONLINE pr2# zpool import tank cannot import 'tank': one or more devices is currently unavailable pr2# zpool import -f tank cannot import 'tank': one or more devices is currently unavailable pr2# --- Importing on the primary gives the same error. Anyone have any ideas? Thanks Corey ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss Jim Dunham Engineering Manager Storage Platform Software Group Sun Microsystems, Inc. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZPOOL Import Problem
-Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] Sent: Friday, September 12, 2008 4:34 PM To: Leopold, Corey Cc: zfs-discuss@opensolaris.org; [EMAIL PROTECTED] Subject: Re: [zfs-discuss] ZPOOL Import Problem Corey, I ran into an odd problem importing a zpool while testing avs. I was trying to simulate a drive failure, break SNDR replication, and then import the pool on the secondary. To simulate the drive failure is just offlined one of the disks in the RAIDZ set. Are all constituent volumes of a single ZFS storage pool in the same SNDR I/O consistency group? Yes they were. One thing to note that shows some of my unfamiliarity with SNDR is that I actually deleted the replication set config before trying to mount on the secondary sndradm -g group -nd Since then I have been throwing them into logging mode sndradm -g group -nl and haven't had a problem in similar tests. (i.e. offlining a drive before making the secondary active). I don't believe that deleting the SNDR replication configuration should have made the ZPOOL invalid though. So there may still be a bug somewhere. Corey ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] ZPOOL Import Problem
I ran into an odd problem importing a zpool while testing avs. I was trying to simulate a drive failure, break SNDR replication, and then import the pool on the secondary. To simulate the drive failure is just offlined one of the disks in the RAIDZ set. -- pr1# zpool status pool: rpool state: ONLINE scrub: none requested config: NAMESTATE READ WRITE CKSUM rpool ONLINE 0 0 0 c3t0d0s0 ONLINE 0 0 0 errors: No known data errors pool: tank state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM tank ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c5t0d0s0 ONLINE 0 0 0 c5t1d0s0 ONLINE 0 0 0 c5t2d0s0 ONLINE 0 0 0 c5t3d0s0 ONLINE 0 0 0 errors: No known data errors pr1# zpool offline missing pool name usage: offline [-t] pool device ... pr1# zpool offline tank c5t0d0s0 pr1# zpool status pool: rpool state: ONLINE scrub: none requested config: NAMESTATE READ WRITE CKSUM rpool ONLINE 0 0 0 c3t0d0s0 ONLINE 0 0 0 errors: No known data errors pool: tank state: DEGRADED status: One or more devices has been taken offline by the administrator. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Online the device using 'zpool online' or replace the device with 'zpool replace'. scrub: none requested config: NAME STATE READ WRITE CKSUM tank DEGRADED 0 0 0 raidz1 DEGRADED 0 0 0 c5t0d0s0 OFFLINE 0 0 0 c5t1d0s0 ONLINE 0 0 0 c5t2d0s0 ONLINE 0 0 0 c5t3d0s0 ONLINE 0 0 0 errors: No known data errors pr1# zpool export tank --- I then disabled SNDR replication. pr1# sndradm -g zfs-tank -d Disable Remote Mirror? (Y/N) [N]: Y - Then I try to import the ZPOOL on the secondary. -- pr2# zpool import pool: tank id: 9795707198744908806 state: DEGRADED status: One or more devices are offlined. action: The pool can be imported despite missing or damaged devices. The fault tolerance of the pool may be compromised if imported. config: tank DEGRADED raidz1 DEGRADED c5t0d0s0 OFFLINE c5t1d0s0 ONLINE c5t2d0s0 ONLINE c5t3d0s0 ONLINE pr2# zpool import tank cannot import 'tank': one or more devices is currently unavailable pr2# zpool import -f tank cannot import 'tank': one or more devices is currently unavailable pr2# --- Importing on the primary gives the same error. Anyone have any ideas? Thanks Corey ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss