Re: [zfs-discuss] I.O error: zpool metadata corrupted after powercut

2008-01-31 Thread Richard Elling
kristof wrote:
> I don't have an exact copy of the error, but the following message was 
> reported by zpool status:
>
> Pool degraded. Meta data corrupted. Please restore pool from backup.
>
> All devices where online, but pool could not be imported. During import we 
> got I/O error.
>   

zpool would complain if it couldn't read the data, for example
because an I/O error occurred.  There are dozens of reasons for
why an I/O error can occur, but without more info, it is nearly
impossible to diagnose via a newsgroup.  An alternate file
system, such as UFS, might just go blissfully along until you
panic or lose data because of an I/O error.
 -- richard


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] I.O error: zpool metadata corrupted after powercut

2008-01-31 Thread kristof
I don't have an exact copy of the error, but the following message was reported 
by zpool status:

Pool degraded. Meta data corrupted. Please restore pool from backup.

All devices where online, but pool could not be imported. During import we got 
I/O error.

Krdoor
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] I.O error: zpool metadata corrupted after powercut

2008-01-30 Thread Richard Elling
kristof wrote:
> Last 2 weeks we had 2 zpools corrupted.
>
> Pool was visible via zpool import, but could not be imported anymore. During 
> import attempt we got I/O error,
>   

What exactly was the error message?
Also look at the fma messages, as they are often more precise.
 -- richard

> After a first powercut we lost our jumpstart/nfsroot zpool (another pool was 
> still OK). Luckaly jumpstart data was backed up and easely restored, nfsroot 
> Filesystems where not but those where just test machines.  We thought the 
> metadata corruption was caused because of the zfs no cache flush setting we 
> had configured in /etc/system (for perfomance reason) in combination with a 
> non battery backuppped NVRAM cache (areca raid controller).
>
> zpool was raidz with 10 local sata disks (JBOD mode)
>
>
> 2 days ago we had another powercut in our test labo :-(
>
> And again one pool was lost. This system was not configured with zfs no cache 
> flush. On the pool we had +/- 40 zvols used by running vm's (iscsi 
> boot/swap/data disks for xen & virtual box guests)
>
> The first failure was on a b68 system, the second on a b77 system.
>
> Last zpool was using iscsi disks: 
>
> setup:
>
> pool
>  mirror:
>iscsidisk1 san1
>iscsidisk1 san2
>  mirror:
>iscsidisk2 san1
>iscsidisk2 san2
>
> I thought zfs was always persistent on disk, but apparently a power cut has 
> can cause unrecoverable damage.
>
> I can accept the first failure (because of the dangerous setting), but 
> loosing that second pool was unacceptable for me.
>
> Since no fsck alike utility is available for zfs I was wondering if there are 
> any plans to create something like meta data repair tools?
>
> Using ZFS now for almost 1 year I was a big Fan, In one year I lost not 1 
> zpool till last week.
>
> At this time I'm concidering to say ZFS is not yet production ready
>
> any comment welcome...
>
> krdoor
>  
>  
> This message posted from opensolaris.org
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>   

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] I.O error: zpool metadata corrupted after powercut

2008-01-30 Thread kristof
Last 2 weeks we had 2 zpools corrupted.

Pool was visible via zpool import, but could not be imported anymore. During 
import attempt we got I/O error,

After a first powercut we lost our jumpstart/nfsroot zpool (another pool was 
still OK). Luckaly jumpstart data was backed up and easely restored, nfsroot 
Filesystems where not but those where just test machines.  We thought the 
metadata corruption was caused because of the zfs no cache flush setting we had 
configured in /etc/system (for perfomance reason) in combination with a non 
battery backuppped NVRAM cache (areca raid controller).

zpool was raidz with 10 local sata disks (JBOD mode)


2 days ago we had another powercut in our test labo :-(

And again one pool was lost. This system was not configured with zfs no cache 
flush. On the pool we had +/- 40 zvols used by running vm's (iscsi 
boot/swap/data disks for xen & virtual box guests)

The first failure was on a b68 system, the second on a b77 system.

Last zpool was using iscsi disks: 

setup:

pool
 mirror:
   iscsidisk1 san1
   iscsidisk1 san2
 mirror:
   iscsidisk2 san1
   iscsidisk2 san2

I thought zfs was always persistent on disk, but apparently a power cut has can 
cause unrecoverable damage.

I can accept the first failure (because of the dangerous setting), but loosing 
that second pool was unacceptable for me.

Since no fsck alike utility is available for zfs I was wondering if there are 
any plans to create something like meta data repair tools?

Using ZFS now for almost 1 year I was a big Fan, In one year I lost not 1 zpool 
till last week.

At this time I'm concidering to say ZFS is not yet production ready

any comment welcome...

krdoor
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss