Re: [zfs-discuss] Zpool import Problem ZFS-8000-EY

2010-09-19 Thread rene
OK, the Pool is died and i had create a new one :-)

regards ré
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Zpool import Problem ZFS-8000-EY

2010-09-18 Thread rene
Hi Folks,

i had an ZFS Pool running on an Hard-Raid5 Controller. I had play with the 
Vendor Maintenance Tool an now my Zpool are in trouble: ZFS-8000-EY

r...@opensolaris:~# zpool import
  pool: data
id: 15260857908242801044
 state: FAULTED
status: The pool was last accessed by another system.
action: The pool cannot be imported due to damaged devices or data.
The pool may be active on another system, but can be imported using
the '-f' flag.
   see: http://www.sun.com/msg/ZFS-8000-EY
config:

dataFAULTED  corrupted data
  c4t0d0p0  ONLINE

r...@opensolaris:~# zpool import data
cannot import 'data': pool may be in use from other system, it was last 
accessed by file.menny.local (hostid: 0xc0d3b779) on Sun Sep  5 20:40:27 2010
use '-f' to import anyway


r...@opensolaris:~# zdb
rpool:
version: 22
name: 'rpool'
state: 0
txg: 4
pool_guid: 964227353010346491
hostid: 756419
hostname: 'opensolaris'
vdev_children: 1
vdev_tree:
type: 'root'
id: 0
guid: 964227353010346491
create_txg: 4
children[0]:
type: 'disk'
id: 0
guid: 5072264041238200753
path: '/dev/dsk/c5t0d0s0'
devid: 'id1,s...@n600508b10010413953535a3038470008/a'
phys_path: 
'/p...@0,0/pci8086,3...@2/pci8086,3...@0,2/pcie11,4...@2/s...@0,0:a'
whole_disk: 0
metaslab_array: 23
metaslab_shift: 28
ashift: 9
asize: 36375625728
is_log: 0
create_txg: 4

please can you give me a little bit help to solve this problem? 

Thanks ré
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Zpool import Problem ZFS-8000-EY

2010-09-18 Thread rene
so, i forgot the rest of information:

ZFS v3
Zpool v15
Hard-Raid5 with CPQary3

the Pool was create in an early release and update to v15 with FreeBSD 8.1 
AMD64, the FreeBSD Forum give me the tip to install OpenSolaris svn_134. Now 
its running ;-)

regards ré
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zpool import problem / missing label / corrupted data

2009-07-20 Thread Rainer Siedler
After a power outage due to a thunder storm my 3 disk raidz1 pool has become 
UNAVAILable.
It is a ZFV v13 pool using the whole 3 disks created on FreeBSD current 8 x64 
and worked well for over a month. Unfortunately I wasn't able to import the 
pool with neither a FreeBSD LiveCD or the current OpenSolaris LiveCD x86/x64.
When I tried to import the pool with FreeBSD the system just hangs (I waited 
for 2 days for a response, which didn't come, not even an error message).
Browsing the web informed me that this problem might be solved by using a newer 
ZFS revision. Therefore I gave opensolaris a try, resulting in the following 
zpool import -f output:

  pool: tank
  id: 582994605255991135
 state: UNAVAIL
status: The pool is formatted using an older on-disk version.
action: The pool cannot be imported due to damaged devices or data.
config:

tank   UNAVAILinsufficient replicas
   raidz1  UNAVAILcorrupted data
  c8t0d0p0   ONLINE
  c10d0p0ONLINE
  c11d0p0ONLINE

The OpenSolaris LiveCD (snv_111b) uses ZFS v14, which would explain the status 
message.

zdb -l shows full LABELs on c8t0d0p0, but the other disks are missing LABEL 2 
and LABEL 3.
c8t0d0p0 is connected via usb and the other disks via onboard SATA.

Some people with similar problems where able to solve them by exporting and 
importing their pools.
I can't export the pool, since I need to import it first in order to have a 
pool available to export, can't destroy it import it with the -D option for the 
same reason.

At least the disks themselves seem to be okay, since dd finished on all disks 
without errors.

Everything I found on the web so far was either not applicable or didn't lead 
to any results.
I hope that you can help me to get at least a fraction of the data back. The 
'important data' was the cifs share used by all in-house computers for central 
storage. There are partial backups of the data, on different disks, but those 
aren't up to date with the files on the pool. Snapshots have been made, but 
those are on the pool as well.

I planned on adding another disk for the OS and use the raidz for storage only. 
The new disk arrived a few days after that incident with the power outage. So
I am close to just setup the whole system from scratch with the new setup, but 
I want to give it a last try to recover it with your expertise. After all it 
would be pretty convenient to add the raidz pool to the fresh install without 
loosing any user data.

Regards, sid
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool import problem / missing label / corrupted data

2009-07-20 Thread raoul le crie qui desaoule
hello,

I cannot help for your problem but one of my drives wwas also destroyed by last 
week-end's storm. and what a storm it was!

good luck with your restore.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool import problem

2008-10-27 Thread Nigel Smith
Hi Terry
Please could you post back to this forum the output from

 # zdb -l /dev/rdsk/...

... for each of the 5 drives in your raidz2.
(maybe best as an attachment)
Are you seeing labels with the error  'failed to unpack'?
What is the reported 'status' of your zpool?
(You have not provided a 'zpool status')
Thanks
Nigel Smith
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool import problem

2008-10-27 Thread Eric Schrock
These are the symptoms of a shrinking device in a RAID-Z pool.  You can
try to run the attached script during the import to see if this the
case.  There's a bug filed on this, but I don't have it handy.

- Eric

On Sun, Oct 26, 2008 at 05:18:25PM -0700, Terry Heatlie wrote:
 Folks,
 I have a zpool with a raidz2 configuration which I've been switching between
 two machines - an old one with a hardware problem and a new one, which
 doesn't have hardware issues, but has a different configuration .   I've
 been trying to import the pool on the new machine, so I can back up the
 data, because the old (broken) machine resets (I don't think it's panicking,
 because there are no logged messages) every time I try to tar off the data
 from the ZFS.
 
  Unfortunately, the first time I tried to import the pool on the new
 machine, I didn't have the right five drives in it, so it didn't work.
  After I figured out that I was confused about which was the boot drive, I
 did get the five drives into the new machine and asked it to import the
 pool.  It said that the pool could not be imported due to damaged devices or
 data.   Which is slightly odd, since it had been mounting the pool fine on
 the broken machine before.
 
 I then moved the drives back into the old machine, figuring I'd at least
 copy some small stuff onto a USB stick (it only dies reading large files,
 apparently), but now the old machine can't mount the pool either, and asking
 it to import gives the same message.   It shows all five drives online, but
 says the pool is UNAVAIL due to insufficient replicas, and the  raidz2 is
 UNAVAIL due to corrupted data.
 
 Must I resign myself to having lost this pool due to the hardware problems
 I've had, and restore such backups as I have on the new machine, or is there
 something that can be done to get the pool back online at least in degraded
 mode?
 
 Thanks in advance,
 
 --Terry.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


--
Eric Schrock, Fishworkshttp://blogs.sun.com/eschrock
#!/sbin/dtrace -s

#pragma D option quiet

BEGIN
{
printf(run 'zpool import' to generate trace\n\n);
}

vdev_raidz_open:entry
{
printf(%d BEGIN RAIDZ OPEN\n, timestamp);
printf(%d config asize = %d\n, timestamp,
args[0]-vdev_asize);
printf(%d config ashift = %d\n, timestamp,
args[0]-vdev_top-vdev_ashift);

self-child = 1;
self-asize = args[1];
self-ashift = args[2];
}

vdev_disk_open:entry
/self-child/
{
self-disk_asize = args[1];
self-disk_ashift = args[2];
}

vdev_disk_open:return
/self-child/
{
printf(%d child[%d]: asize = %d, ashift = %d\n, timestamp,
   self-child - 1, *self-disk_asize, *self-disk_ashift);
self-disk_asize = 0;
self-disk_ashift = 0;
self-child++;
}

vdev_raidz_open:return
{
printf(%d asize = %d\n, timestamp, *self-asize);
printf(%d ashift = %d\n, timestamp, *self-ashift);
printf(%d END RAIDZ OPEN\n, timestamp);
self-child = 0;
self-asize = 0;
self-ashift = 0;
}
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool import problem

2008-10-27 Thread Victor Latushkin
Eric Schrock пишет:
 These are the symptoms of a shrinking device in a RAID-Z pool.  You can
 try to run the attached script during the import to see if this the
 case.  There's a bug filed on this, but I don't have it handy.

it's

6753869 labeling/shrinking a disk in raid-z vdev makes pool un-importable

victor

 
 - Eric
 
 On Sun, Oct 26, 2008 at 05:18:25PM -0700, Terry Heatlie wrote:
 Folks,
 I have a zpool with a raidz2 configuration which I've been switching between
 two machines - an old one with a hardware problem and a new one, which
 doesn't have hardware issues, but has a different configuration .   I've
 been trying to import the pool on the new machine, so I can back up the
 data, because the old (broken) machine resets (I don't think it's panicking,
 because there are no logged messages) every time I try to tar off the data
 from the ZFS.

  Unfortunately, the first time I tried to import the pool on the new
 machine, I didn't have the right five drives in it, so it didn't work.
  After I figured out that I was confused about which was the boot drive, I
 did get the five drives into the new machine and asked it to import the
 pool.  It said that the pool could not be imported due to damaged devices or
 data.   Which is slightly odd, since it had been mounting the pool fine on
 the broken machine before.

 I then moved the drives back into the old machine, figuring I'd at least
 copy some small stuff onto a USB stick (it only dies reading large files,
 apparently), but now the old machine can't mount the pool either, and asking
 it to import gives the same message.   It shows all five drives online, but
 says the pool is UNAVAIL due to insufficient replicas, and the  raidz2 is
 UNAVAIL due to corrupted data.

 Must I resign myself to having lost this pool due to the hardware problems
 I've had, and restore such backups as I have on the new machine, or is there
 something that can be done to get the pool back online at least in degraded
 mode?

 Thanks in advance,

 --Terry.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool import problem

2008-10-27 Thread MC
I recently tried to import a b97 pool into a b98 upgraded version of that os, 
and it failed because of some bug.  So maybe try eliminating that kind of 
problem by making sure to use the version that you know worked in the past.  
Maybe you already did this.

 div id=jive-html-wrapper-div
 Folks,divbr/divdivI have a zpool with a
 raidz2 configuration which I#39;ve been switching
 between two machines - an old one with a hardware
 problem and a new one, which doesn#39;t have
 hardware issues, but has a different configuration .
 nbsp; I#39;ve been trying to import the pool on the
 new machine, so I can back up the data, because the
 old (broken) machine resets (I don#39;t think
 it#39;s panicking, because there are no logged
 messages) every time I try to tar off the data from
 the ZFS. nbsp;/div
 divbr/divdivnbsp;Unfortunately, the first
 time I tried to import the pool on the new machine, I
 didn#39;t have the right five drives in it, so it
 didn#39;t work. nbsp;After I figured out that I was
 confused about which was the boot drive, I did get
 the five drives into the new machine and asked it to
 import the pool. nbsp;It said that the pool could
 not be imported due to damaged devices or data.
 nbsp; Which is slightly odd, since it had been
 mounting the pool fine on the broken machine before.
 nbsp;nbsp;/div
 divbr/divdivI then moved the drives back into
 the old machine, figuring I#39;d at least copy some
 small stuff onto a USB stick (it only dies reading
 large files, apparently), but now the old machine
 can#39;t mount the pool either, and asking it to
 import gives the same message. nbsp; It shows all
 five drives online, but says the pool is UNAVAIL due
 to insufficient replicas, and the nbsp;raidz2 is
 UNAVAIL due to corrupted data./div
 divbr/divdivMust I resign myself to having
 lost this pool due to the hardware problems I#39;ve
 had, and restore such backups as I have on the new
 machine, or is there something that can be done to
 get the pool back online at least in degraded
 mode?/div
 divbr/divdivThanks in
 advance,/divdivbr/divdiv--Terry./div
 
 /div___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discu
 ss
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zpool import problem

2008-10-26 Thread Terry Heatlie
Folks,
I have a zpool with a raidz2 configuration which I've been switching between
two machines - an old one with a hardware problem and a new one, which
doesn't have hardware issues, but has a different configuration .   I've
been trying to import the pool on the new machine, so I can back up the
data, because the old (broken) machine resets (I don't think it's panicking,
because there are no logged messages) every time I try to tar off the data
from the ZFS.

 Unfortunately, the first time I tried to import the pool on the new
machine, I didn't have the right five drives in it, so it didn't work.
 After I figured out that I was confused about which was the boot drive, I
did get the five drives into the new machine and asked it to import the
pool.  It said that the pool could not be imported due to damaged devices or
data.   Which is slightly odd, since it had been mounting the pool fine on
the broken machine before.

I then moved the drives back into the old machine, figuring I'd at least
copy some small stuff onto a USB stick (it only dies reading large files,
apparently), but now the old machine can't mount the pool either, and asking
it to import gives the same message.   It shows all five drives online, but
says the pool is UNAVAIL due to insufficient replicas, and the  raidz2 is
UNAVAIL due to corrupted data.

Must I resign myself to having lost this pool due to the hardware problems
I've had, and restore such backups as I have on the new machine, or is there
something that can be done to get the pool back online at least in degraded
mode?

Thanks in advance,

--Terry.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool import problem

2008-10-26 Thread Richard Elling
Terry Heatlie wrote:
 Folks,

 I have a zpool with a raidz2 configuration which I've been switching 
 between two machines - an old one with a hardware problem and a new 
 one, which doesn't have hardware issues, but has a different 
 configuration .   I've been trying to import the pool on the new 
 machine, so I can back up the data, because the old (broken) machine 
 resets (I don't think it's panicking, because there are no logged 
 messages) every time I try to tar off the data from the ZFS.  

  Unfortunately, the first time I tried to import the pool on the new 
 machine, I didn't have the right five drives in it, so it didn't work. 
  After I figured out that I was confused about which was the boot 
 drive, I did get the five drives into the new machine and asked it to 
 import the pool.  It said that the pool could not be imported due to 
 damaged devices or data.   Which is slightly odd, since it had been 
 mounting the pool fine on the broken machine before.   

 I then moved the drives back into the old machine, figuring I'd at 
 least copy some small stuff onto a USB stick (it only dies reading 
 large files, apparently), but now the old machine can't mount the pool 
 either, and asking it to import gives the same message.   It shows all 
 five drives online, but says the pool is UNAVAIL due to insufficient 
 replicas, and the  raidz2 is UNAVAIL due to corrupted data.

 Must I resign myself to having lost this pool due to the hardware 
 problems I've had, and restore such backups as I have on the new 
 machine, or is there something that can be done to get the pool back 
 online at least in degraded mode?

Note: we're also working on a troubleshooting wiki... need more days in the
hour...

You should try to read the labels from each device.
zdb -l /dev/rdsk/...

You should see 4 labels for each proper device.

Here is my hypothesis:

If you see a device which has only label 0 and 1, then it may
be the case that the label has overlapping partitions.  Why does
this matter?  Because under normal circumstances, the actual
devices used for creating or importing the pool are stored in the
/etc/zfs/zpool.cache file.  When the system boots, it looks there
first and will import the pools listed therein.

When you export the pool, the zpool.cache entries for the pool
are removed.

If the pool is not in zpool.cache, then zpool import scans all of
the devices found in /dev/dsk for valid pools.  If you have overlapping
partitions or slices, then a partially exposed vdev may be found.
But since it won't be complete, due to perhaps not being able to
see the end of the device which is where labels 2  3 are located,
then it will be marked as bad.  The solution would be to reconcile
the partions/slices using format.
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZPOOL Import Problem

2008-09-17 Thread Jim Dunham

On Sep 16, 2008, at 5:39 PM, Miles Nordin wrote:

 jd == Jim Dunham [EMAIL PROTECTED] writes:

jd If at the time the SNDR replica is deleted the set was
jd actively replicating, along with ZFS actively writing to the
jd ZFS storage pool, I/O consistency will be lost, leaving ZFS
jd storage pool in an indeterministic state on the remote node.

jd To address this issue, prior to deleting the replicas, the
jd replica should be placed into logging mode first.

 What if you stop the replication by breaking the network connection
 between primary and replica?  consistent or inconsistent?

Consistent.

 it sounds fishy, like ``we're always-consistent-on-disk with ZFS, but
 please use 'zpool offline' to avoid disastrous pool corruption.''

This is not the case at all.

Maintaining I/O consistency of all volumes in a single I/O consistency  
group, is an attribute of replication. The instant an SNDR replica is  
deleted, that volume is no longer being replicated, and it becomes  
inconsistent with all other write-order volumes. By placing all  
volumes in the I/O consistency group in logging mode, not 'zpool  
offline', and then deleting the replica there is no means for any of  
the remote volumes to become I/O inconsistent.

Yes, one will note that there is a group disable command sndradm -g  
group-name -d, but it was implemented for easy of administration,  
not for performing a write-order coordinated disable command.

jd ndr_ii. This is an automatic snapshot taken before
jd resynchronization starts,

 yeah that sounds fine, possibly better than DRBD in one way because it
 might allow the resync to go faster.

 From the PDF's it sounds like async replication isn't done the same
 way as the resync, it's done safely, and that it's even possible for
 async replication to accumulate hours of backlog in a ``disk queue''
 without losing write ordering so long as you use the ``blocking mode''
 variant of async.

Correct reading of the documentation.

 ii might also be good for debugging a corrupt ZFS, so you can tinker
 with it but still roll back to the original corrupt copy.  I'll read
 about it---I'm guessing I will need to prepare ahead of time if I want
 ii available in the toolbox after a disaster.

jd AVS has the concept of I/O consistency groups, where all disks
jd of a multi-volume filesystem (ZFS, QFS) or database (Oracle,
jd Sybase) are kept write-order consistent when using either sync
jd or async replication.

 Awesome, so long as people know to use it.  so I guess that's the
 answer for the OP: use consistency groups!

I use the name of the ZFS storage pool, as the name of the SNDR I/O  
consistency group.

 The one thing I worry about is, before, AVS was used between RAID and
 filesystem, which is impossible now because that inter-layer area n
 olonger exists.  If you put the individual device members of a
 redundant zpool vdev into an AVS consistency group, what will AVS do
 when one of the devices fails?

Nothing, as it is ZFS the reacts to the failed device

 Does it continue replicating the working devices and ignore the  
 failed one?

In this scenario ZFS knows he device failed, which means ZFS will stop  
writing to the disk, and thus the replica.


 This would sacrifice redundancy at the DR site.  UFS-AVS-RAID
 would not do that in the same situation.

 Or hide the failed device from ZFS and slow things down by sending all
 read/writes of the failed device to the remote mirror?  This would
 slwo down the primary site.  UFS-AVS-RAID would not do that in the
 same situation.

 The latter ZFS-AVS behavior might be rescueable, if ZFS had the
 statistical read-preference feature.  but writes would still be
 massively slowed with this scenario, while in UFS-AVS-RAID they would
 not be.  To get back the level of control one used to have for writes,
 you'd need a different zpool-level way to achieve the intent of the
 AVS sync/async option.  Maybe just a slog which is not AVS-replicated
 would be enough, modulo other ZFS fixes for hiding slow devices.

ZFS-AVS is not UFS-AVS-RAID, and although one can foresee some  
downside to replicating ZFS with AVS, there are some big wins.

Place SNDR in logging mode, and zpool scrub the secondary volumes for  
consistency, then resume replication.
Compressed ZFS Storage pools, result in compressed replication
Encrypted ZFS Storage pools, result in encrypted replication



 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Jim Dunham
Engineering Manager
Storage Platform Software Group
Sun Microsystems, Inc.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZPOOL Import Problem

2008-09-16 Thread Miles Nordin
 jd == Jim Dunham [EMAIL PROTECTED] writes:

jd If at the time the SNDR replica is deleted the set was
jd actively replicating, along with ZFS actively writing to the
jd ZFS storage pool, I/O consistency will be lost, leaving ZFS
jd storage pool in an indeterministic state on the remote node.

jd To address this issue, prior to deleting the replicas, the
jd replica should be placed into logging mode first.

What if you stop the replication by breaking the network connection
between primary and replica?  consistent or inconsistent?

it sounds fishy, like ``we're always-consistent-on-disk with ZFS, but
please use 'zpool offline' to avoid disastrous pool corruption.''

jd ndr_ii. This is an automatic snapshot taken before
jd resynchronization starts,

yeah that sounds fine, possibly better than DRBD in one way because it
might allow the resync to go faster.  

From the PDF's it sounds like async replication isn't done the same
way as the resync, it's done safely, and that it's even possible for
async replication to accumulate hours of backlog in a ``disk queue''
without losing write ordering so long as you use the ``blocking mode''
variant of async.

ii might also be good for debugging a corrupt ZFS, so you can tinker
with it but still roll back to the original corrupt copy.  I'll read
about it---I'm guessing I will need to prepare ahead of time if I want
ii available in the toolbox after a disaster.

jd AVS has the concept of I/O consistency groups, where all disks
jd of a multi-volume filesystem (ZFS, QFS) or database (Oracle,
jd Sybase) are kept write-order consistent when using either sync
jd or async replication.

Awesome, so long as people know to use it.  so I guess that's the
answer for the OP: use consistency groups!

The one thing I worry about is, before, AVS was used between RAID and
filesystem, which is impossible now because that inter-layer area n
olonger exists.  If you put the individual device members of a
redundant zpool vdev into an AVS consistency group, what will AVS do
when one of the devices fails?

Does it continue replicating the working devices and ignore the failed
one?  This would sacrifice redundancy at the DR site.  UFS-AVS-RAID
would not do that in the same situation.

Or hide the failed device from ZFS and slow things down by sending all
read/writes of the failed device to the remote mirror?  This would
slwo down the primary site.  UFS-AVS-RAID would not do that in the
same situation.

The latter ZFS-AVS behavior might be rescueable, if ZFS had the
statistical read-preference feature.  but writes would still be
massively slowed with this scenario, while in UFS-AVS-RAID they would
not be.  To get back the level of control one used to have for writes,
you'd need a different zpool-level way to achieve the intent of the
AVS sync/async option.  Maybe just a slog which is not AVS-replicated
would be enough, modulo other ZFS fixes for hiding slow devices.


pgpzm3T09CxRc.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZPOOL Import Problem

2008-09-13 Thread Jim Dunham
Corey

 -Original Message-
 From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED]
 Sent: Friday, September 12, 2008 4:34 PM
 To: Leopold, Corey
 Cc: zfs-discuss@opensolaris.org; [EMAIL PROTECTED]
 Subject: Re: [zfs-discuss] ZPOOL Import Problem

 Corey,

 I ran into an odd problem importing a zpool while testing avs.  I  
 was
 trying to simulate a drive failure, break SNDR replication, and then
 import the pool on the secondary.  To simulate the drive failure is
 just offlined one of the disks in the RAIDZ set.


 Are all constituent volumes of a single ZFS storage pool in the same
 SNDR I/O consistency group?


 Yes they were.

 One thing to note that shows some of my unfamiliarity with SNDR is  
 that
 I actually deleted the replication set config before trying to mount  
 on
 the secondary sndradm -g group -nd

 Since then I have been throwing them into logging mode sndradm -g  
 group
 -nl and haven't had a problem in similar tests. (i.e. offlining a  
 drive
 before making the secondary active).

 I don't believe that deleting the SNDR replication configuration  
 should
 have made the ZPOOL invalid though.  So there may still be a bug
 somewhere.

If at the time the SNDR replica is deleted the set was actively  
replicating, along with ZFS actively writing to the ZFS storage pool,  
I/O consistency will be lost, leaving ZFS storage pool in an  
indeterministic state on the remote node.  To address this issue,  
prior to deleting the replicas, the replica should be placed into  
logging mode first. Then ZFS will be left in I/O consistent after the  
disable is done.

 Corey

Jim Dunham
Engineering Manager
Storage Platform Software Group
Sun Microsystems, Inc.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZPOOL Import Problem

2008-09-12 Thread Jim Dunham
Corey,

 I ran into an odd problem importing a zpool while testing avs.  I was
 trying to simulate a drive failure, break SNDR replication, and then
 import the pool on the secondary.  To simulate the drive failure is  
 just
 offlined one of the disks in the RAIDZ set.


Are all constituent volumes of a single ZFS storage pool in the same  
SNDR I/O consistency group?



 --
 pr1# zpool status
  pool: rpool
 state: ONLINE
 scrub: none requested
 config:

NAMESTATE READ WRITE CKSUM
rpool   ONLINE   0 0 0
  c3t0d0s0  ONLINE   0 0 0

 errors: No known data errors

  pool: tank
 state: ONLINE
 scrub: none requested
 config:

NAME  STATE READ WRITE CKSUM
tank  ONLINE   0 0 0
  raidz1  ONLINE   0 0 0
c5t0d0s0  ONLINE   0 0 0
c5t1d0s0  ONLINE   0 0 0
c5t2d0s0  ONLINE   0 0 0
c5t3d0s0  ONLINE   0 0 0

 errors: No known data errors
 pr1# zpool offline
 missing pool name
 usage:
offline [-t] pool device ...
 pr1# zpool offline tank c5t0d0s0
 pr1# zpool status
  pool: rpool
 state: ONLINE
 scrub: none requested
 config:

NAMESTATE READ WRITE CKSUM
rpool   ONLINE   0 0 0
  c3t0d0s0  ONLINE   0 0 0

 errors: No known data errors

  pool: tank
 state: DEGRADED
 status: One or more devices has been taken offline by the  
 administrator.
Sufficient replicas exist for the pool to continue functioning
 in a
degraded state.
 action: Online the device using 'zpool online' or replace the device
 with
'zpool replace'.
 scrub: none requested
 config:

NAME  STATE READ WRITE CKSUM
tank  DEGRADED 0 0 0
  raidz1  DEGRADED 0 0 0
c5t0d0s0  OFFLINE  0 0 0
c5t1d0s0  ONLINE   0 0 0
c5t2d0s0  ONLINE   0 0 0
c5t3d0s0  ONLINE   0 0 0

 errors: No known data errors
 pr1# zpool export tank
 ---

 I then disabled SNDR replication.
 
 pr1# sndradm -g zfs-tank -d
 Disable Remote Mirror? (Y/N) [N]: Y
 -

 Then I try to import the ZPOOL on the secondary.

 --
 pr2# zpool import
  pool: tank
id: 9795707198744908806
 state: DEGRADED
 status: One or more devices are offlined.
 action: The pool can be imported despite missing or damaged devices.
 The
fault tolerance of the pool may be compromised if imported.
 config:

tank  DEGRADED
  raidz1  DEGRADED
c5t0d0s0  OFFLINE
c5t1d0s0  ONLINE
c5t2d0s0  ONLINE
c5t3d0s0  ONLINE
 pr2# zpool import tank
 cannot import 'tank': one or more devices is currently unavailable
 pr2# zpool import -f tank
 cannot import 'tank': one or more devices is currently unavailable
 pr2#
 ---

 Importing on the primary gives the same error.

 Anyone have any ideas?

 Thanks

 Corey
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Jim Dunham
Engineering Manager
Storage Platform Software Group
Sun Microsystems, Inc.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZPOOL Import Problem

2008-09-12 Thread Leopold, Corey


 -Original Message-
 From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED]
 Sent: Friday, September 12, 2008 4:34 PM
 To: Leopold, Corey
 Cc: zfs-discuss@opensolaris.org; [EMAIL PROTECTED]
 Subject: Re: [zfs-discuss] ZPOOL Import Problem
 
 Corey,
 
  I ran into an odd problem importing a zpool while testing avs.  I
was
  trying to simulate a drive failure, break SNDR replication, and then
  import the pool on the secondary.  To simulate the drive failure is
  just
  offlined one of the disks in the RAIDZ set.
 
 
 Are all constituent volumes of a single ZFS storage pool in the same
 SNDR I/O consistency group?
 

Yes they were.

One thing to note that shows some of my unfamiliarity with SNDR is that
I actually deleted the replication set config before trying to mount on
the secondary sndradm -g group -nd

Since then I have been throwing them into logging mode sndradm -g group
-nl and haven't had a problem in similar tests. (i.e. offlining a drive
before making the secondary active).

I don't believe that deleting the SNDR replication configuration should
have made the ZPOOL invalid though.  So there may still be a bug
somewhere.

Corey
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZPOOL Import Problem

2008-09-10 Thread Leopold, Corey
I ran into an odd problem importing a zpool while testing avs.  I was
trying to simulate a drive failure, break SNDR replication, and then
import the pool on the secondary.  To simulate the drive failure is just
offlined one of the disks in the RAIDZ set.



--
pr1# zpool status
  pool: rpool
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
rpool   ONLINE   0 0 0
  c3t0d0s0  ONLINE   0 0 0

errors: No known data errors

  pool: tank
 state: ONLINE
 scrub: none requested
config:

NAME  STATE READ WRITE CKSUM
tank  ONLINE   0 0 0
  raidz1  ONLINE   0 0 0
c5t0d0s0  ONLINE   0 0 0
c5t1d0s0  ONLINE   0 0 0
c5t2d0s0  ONLINE   0 0 0
c5t3d0s0  ONLINE   0 0 0

errors: No known data errors
pr1# zpool offline
missing pool name
usage:
offline [-t] pool device ...
pr1# zpool offline tank c5t0d0s0
pr1# zpool status
  pool: rpool
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
rpool   ONLINE   0 0 0
  c3t0d0s0  ONLINE   0 0 0

errors: No known data errors

  pool: tank
 state: DEGRADED
status: One or more devices has been taken offline by the administrator.
Sufficient replicas exist for the pool to continue functioning
in a
degraded state.
action: Online the device using 'zpool online' or replace the device
with
'zpool replace'.
 scrub: none requested
config:

NAME  STATE READ WRITE CKSUM
tank  DEGRADED 0 0 0
  raidz1  DEGRADED 0 0 0
c5t0d0s0  OFFLINE  0 0 0
c5t1d0s0  ONLINE   0 0 0
c5t2d0s0  ONLINE   0 0 0
c5t3d0s0  ONLINE   0 0 0

errors: No known data errors
pr1# zpool export tank
---

I then disabled SNDR replication.

pr1# sndradm -g zfs-tank -d
Disable Remote Mirror? (Y/N) [N]: Y
-

Then I try to import the ZPOOL on the secondary.

--
pr2# zpool import
  pool: tank
id: 9795707198744908806
 state: DEGRADED
status: One or more devices are offlined.
action: The pool can be imported despite missing or damaged devices.
The
fault tolerance of the pool may be compromised if imported.
config:

tank  DEGRADED
  raidz1  DEGRADED
c5t0d0s0  OFFLINE
c5t1d0s0  ONLINE
c5t2d0s0  ONLINE
c5t3d0s0  ONLINE
pr2# zpool import tank
cannot import 'tank': one or more devices is currently unavailable
pr2# zpool import -f tank
cannot import 'tank': one or more devices is currently unavailable
pr2#
---

Importing on the primary gives the same error.

Anyone have any ideas?

Thanks

Corey
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss