Re: [zfs-discuss] Remove corrupt files from snapshot

2011-11-15 Thread Todd Urie
Use zpool status -v to see if any errors come up.  Then you can use zpool scrub 
to remove at least some of them.  I have had luck with this in the past.


---Todd



On Nov 14, 2011, at 04:25 , sbre...@hotmail.com sbre...@hotmail.com wrote:

 
 Back to this topic, since I cannot touch snapshots I thought I could simply 
 remove the corrupt files after the last snapshot, so the next incremental 
 backup will notice the difference (i.e. no file) and overwrite the 
 corrupt-and-removed files with valid ones. This was the plan.
 
 However, while checking for corrupt files, find stops at some directory 
 with fts_read: Not a directory:
 
 find . -exec md5 {} \;  /home/xxx/md5_out 2 /home/xxx/md5_err 
 
 tail /home/xxx/md5_err
 ...
 md5: ./.zfs/snapshot/20100323081201/Bazsi/Projects/Java Test 
 Client/java_test_client/lib/xxx/weblogic.jar: Input/output error
 md5: ./.zfs/snapshot/20100323081201/@Cache (Bazsi)/BMWi 
 SP/Publikationen/PDF-Broschâ–’ren/Nexxt.pdf: Input/output error
 find: fts_read: Not a directory
 
 What does this error mean? I cannot even scan the ZFS file system anymore? 
 Is there any fsck for ZFS?
 
 
 Cheers,
 B.
 
 
 From: zfsdisc...@orgdotuk.org.uk
 To: zfs-discuss@opensolaris.org
 Date: Mon, 7 Nov 2011 21:49:56 +
 Subject: Re: [zfs-discuss] Remove corrupt files from snapshot
 
 -Original Message-
 From: Edward Ned Harvey
 Sent: 04/11/2011 21:23
 
 You need to destroy the snapshot completely - But if you want
 to selectively
 delete from a snapshot, I think you can clone it, then
 promote the clone,
 then destroy the snapshot, then rm something from the clone and then
 snapshot the clone back to the original name, and then
 destroy the clone.
 
 Right?
 
 Not so fast! :-)
 
 If you promote this new clone, the current state / branch of your filesystem 
 becomes a clone instead, dependent on the snapshot.
 Then if you try to destroy the snapshot, you'll fail, because it has a 
 dependent clone (your current fs!!!). If you continue
 without realising the implications, and so try the 'destroy' again with 
 '-R', there goes the neighbourhood!
 
 I did this once, and was only saved by the fact that my cwd was in my 
 current filesystem, so couldn't be unmounted, and therefore
 couldn't be removed! Phew!! Nice to learn something and only get singed 
 eyebrows, instead of losing a leg!
 
 hth Andy
 
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss 
   
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Zpool with data errors

2011-06-22 Thread Todd Urie
I'll be doing this over the upcoming weekend so I'll see how it goes.

Thanks for all of the suggestions. 

Todd



On Jun 22, 2011, at 10:48 AM, Cindy Swearingen cindy.swearin...@oracle.com 
wrote:

 Hi Todd,
 
 Yes, I have seen zpool scrub do some miracles but I think it depends
 on the amount of corruption.
 
 A few suggestions are:
 
 1. Identify and resolve the corruption problems on the underlying
 hardware. No point in trying to clear the pool errors if this
 problem continues.
 
 The fmdump command and the fmdump -eV command output will
 tell you how long these errors have occurred.
 
 2. Run zpool scrub and zpool clear to attempt to clear the errors.
 
 3. If the errors below don't clear, then manually remove the corrupted
 files below, if possible, and restore from backup. Depending on what
 fmdump says, you might check your backups for corruption.
 
 4. Run zpool scrub and zpool clear again as needed.
 
 5. Consider replacing this configuration with a redundant ZFS storage
 pool. We can provide the recommended syntax.
 
 Let us know how this turns out.
 
 Thanks,
 
 Cindy
 
 On 06/20/11 23:36, Todd Urie wrote:
 I have a zpool that shows the following from a zpool status -v zpool name
 brsnnfs0104 [/var/spool/cron/scripts]# zpool status -v ABC0101
  pool:ABC0101
 state: ONLINE
 status: One or more devices has experienced an error resulting in data
corruption.  Applications may be affected.
 action: Restore the file in question if possible.  Otherwise restore the
entire pool from backup.
   see: http://www.sun.com/msg/ZFS-8000-8A
 scrub: none requested
 config:
NAME  STATE READ WRITE CKSUM
ABC0101   ONLINE   0 010
  /dev/vx/dsk/ABC01dg/ABC0101_01  ONLINE   0 0 2
  /dev/vx/dsk/ABC01dg/ABC0101_02  ONLINE   0 0 8
  /dev/vx/dsk/ABC01dg/ABC0101_03  ONLINE   0 010
 errors: Permanent errors have been detected in the following files:

 /clients/ABC0101/rep/local/bfm/web/htdocs/tmp/rscache/717b52282ea059452621587173561360

 /clients/ABC0101/rep/local/bfm/web/htdocs/tmp/rscache/6e6a9f37c4d13fdb3dcb8649272a2a49

 /clients/ABC0101/rep/d0/prod1/reports/ReutersCMOLoad/ReutersCMOLoad.ABCntss001.20110620.141330.26496.ROLLBACK_FOR_UPDATE_COUPONS.html

 /clients/ABC0101/rep/local/bfm/web/htdocs/tmp/G2_0.related_detail_loader.1308593666.54643.n5cpoli3355.data

 /clients/ABC0101/rep/d0/prod1/reports/gp_reports/ALLMNG/20110429/F_OLPO82_A.gp.ABCIM_GA.nlaf.xml.gz

 /clients/ABC0101/rep/d0/prod1/reports/gp_reports/ALLMNG/20110429/UNVLXCIAFI.gp.ABCIM_GA.nlaf.xml.gz

 /clients/ABC0101/rep/d0/prod1/reports/gp_reports/ALLMNG/20110429/UNIVLEXCIA.gp.BARCRATING_ABC.nlaf.xml.gz
 I think that a scrub at least has the possibility to clear this up.  A quick 
 search suggests that others have had some good experience with using scrub 
 in similar circumstances.  I was wondering if anyone could share some of 
 their experiences, good and bad, so that I can assess the risk and 
 probability of success with this approach.  Also, any other ideas would 
 certainly be appreciated.
 -RTU
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Zpool with data errors

2011-06-21 Thread Todd Urie
The volumes sit on HDS SAN.  The only reason for the volumes is to prevent
inadvertent import of the zpool on two nodes of a cluster simultaneously.
 Since we're on SAN with Raid internally, didn't seem to we would need zfs
to provide that redundancy also.

On Tue, Jun 21, 2011 at 4:17 AM, Remco Lengers re...@lengers.com wrote:

 **
 Todd,

 Is that ZFS on top of VxVM ?  Are those volumes okay? I wonder if this is
 really a sensible combination?

 ..Remco


 On 6/21/11 7:36 AM, Todd Urie wrote:

 I have a zpool that shows the following from a zpool status -v zpool name


  brsnnfs0104 [/var/spool/cron/scripts]# zpool status -v  ABC0101
   pool: ABC0101
  state: ONLINE
 status: One or more devices has experienced an error resulting in data
 corruption.  Applications may be affected.
 action: Restore the file in question if possible.  Otherwise restore the
 entire pool from backup.
see: http://www.sun.com/msg/ZFS-8000-8A
  scrub: none requested
 config:

  NAME  STATE READ WRITE CKSUM
  ABC0101   ONLINE   0 010
   /dev/vx/dsk/ ABC01dg/ ABC0101_01  ONLINE   0 0 2
   /dev/vx/dsk/ ABC01dg/ ABC0101_02  ONLINE   0 0 8
   /dev/vx/dsk/ ABC01dg/ ABC0101_03  ONLINE   0 010

  errors: Permanent errors have been detected in the following files:

 /clients/ABC0101/rep/local/bfm/web/htdocs/tmp/rscache/717b52282ea059452621587173561360
 /clients/
 ABC0101/rep/local/bfm/web/htdocs/tmp/rscache/6e6a9f37c4d13fdb3dcb8649272a2a49
 /clients/
 ABC0101/rep/d0/prod1/reports/ReutersCMOLoad/ReutersCMOLoad.
 ABCntss001.20110620.141330.26496.ROLLBACK_FOR_UPDATE_COUPONS.html
 /clients/
 ABC0101/rep/local/bfm/web/htdocs/tmp/G2_0.related_detail_loader.1308593666.54643.n5cpoli3355.data
 /clients/
 ABC0101/rep/d0/prod1/reports/gp_reports/ALLMNG/20110429/F_OLPO82_A.gp.
 ABCIM_GA.nlaf.xml.gz
 /clients/
 ABC0101/rep/d0/prod1/reports/gp_reports/ALLMNG/20110429/UNVLXCIAFI.gp.
 ABCIM_GA.nlaf.xml.gz
 /clients/
 ABC0101/rep/d0/prod1/reports/gp_reports/ALLMNG/20110429/UNIVLEXCIA.gp.BARCRATING_
 ABC.nlaf.xml.gz

  I think that a scrub at least has the possibility to clear this up.  A
 quick search suggests that others have had some good experience with using
 scrub in similar circumstances.  I was wondering if anyone could share some
 of their experiences, good and bad, so that I can assess the risk and
 probability of success with this approach.  Also, any other ideas would
 certainly be appreciated.


 -RTU


 ___
 zfs-discuss mailing 
 listzfs-discuss@opensolaris.orghttp://mail.opensolaris.org/mailman/listinfo/zfs-discuss




-- 
-RTU
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Zpool with data errors

2011-06-20 Thread Todd Urie
I have a zpool that shows the following from a zpool status -v zpool name

brsnnfs0104 [/var/spool/cron/scripts]# zpool status -v ABC0101
  pool:ABC0101
 state: ONLINE
status: One or more devices has experienced an error resulting in data
corruption.  Applications may be affected.
action: Restore the file in question if possible.  Otherwise restore the
entire pool from backup.
   see: http://www.sun.com/msg/ZFS-8000-8A
 scrub: none requested
config:

NAME  STATE READ WRITE CKSUM
ABC0101   ONLINE   0 010
  /dev/vx/dsk/ABC01dg/ABC0101_01  ONLINE   0 0 2
  /dev/vx/dsk/ABC01dg/ABC0101_02  ONLINE   0 0 8
  /dev/vx/dsk/ABC01dg/ABC0101_03  ONLINE   0 010

errors: Permanent errors have been detected in the following files:

/clients/ABC0101/rep/local/bfm/web/htdocs/tmp/rscache/717b52282ea059452621587173561360
/clients/
ABC0101/rep/local/bfm/web/htdocs/tmp/rscache/6e6a9f37c4d13fdb3dcb8649272a2a49
/clients/ABC0101/rep/d0/prod1/reports/ReutersCMOLoad/ReutersCMOLoad.
ABCntss001.20110620.141330.26496.ROLLBACK_FOR_UPDATE_COUPONS.html
/clients/
ABC0101/rep/local/bfm/web/htdocs/tmp/G2_0.related_detail_loader.1308593666.54643.n5cpoli3355.data
/clients/
ABC0101/rep/d0/prod1/reports/gp_reports/ALLMNG/20110429/F_OLPO82_A.gp.
ABCIM_GA.nlaf.xml.gz
/clients/
ABC0101/rep/d0/prod1/reports/gp_reports/ALLMNG/20110429/UNVLXCIAFI.gp.
ABCIM_GA.nlaf.xml.gz
/clients/
ABC0101/rep/d0/prod1/reports/gp_reports/ALLMNG/20110429/UNIVLEXCIA.gp.BARCRATING_
ABC.nlaf.xml.gz

I think that a scrub at least has the possibility to clear this up.  A quick
search suggests that others have had some good experience with using scrub
in similar circumstances.  I was wondering if anyone could share some of
their experiences, good and bad, so that I can assess the risk and
probability of success with this approach.  Also, any other ideas would
certainly be appreciated.


-RTU
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS file system without pool

2010-09-15 Thread Todd Urie
A ZFS filesystem without a zpool doesn't make much sense.  Unless I'm badly
mistaken, you have to have the pool to get the filesystem.

As far as using a Veritas volume for the zpools, that is easily done.  We do
that where I work for almost all of our ZFS filesystems as a way to
facilitate cluster failover of the filesystems and ensure that the zpool
can't be imported on two nodes simultaneously.   You just specify the VxVM
volumes as the block devices when you configure the zpool.  Of course, you
have to import the releveant disk group before you can import the zpool.

On Wed, Sep 15, 2010 at 8:25 AM, Ramesh Babu rama.b...@gmail.com wrote:

 I would like to know if  I can create ZFS file system without ZFS storage
 pool. Also I would like to know if I can create ZFS pool/ZFS pool on Veritas
 Volume.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss