I would like advise about how to replace a raid 0 lun.  The lun is basically a 
raid 0 lun which is from a single disk volume group / volume from our Flexline 
380 unit.  So every disk in the unit is a volume group/volume/lun mapped to the 
host.  We then let ZFS do the raid.  

We have a lun now which has been getting read errors and basically the 
underlying drive needs to be replaced.  I've done this in the past but as I 
remember it was cumbersome and didn't go that smoothly so I would like some 
advise about how to go about it.  As I recall if you just fail the disk via 
Santricity, the lun really doesn't go offline from a host point of via and 
hence ZFS still tries to write to the lun/disk.  I believe I either unmapped 
the lun to the host at which point ZFS kicked in a hot spare, or I offlined the 
disk via ZFS first.  Then at that point the drive can be replace on the storage 
and the volume re-initialized then remapped back to the host.  So part of my 
question... does the above sound reasonable, or should I be doing this 
differently?  Also I'm a little unsure about how to get the original lun back 
in operation and the spare back to a spare.

For example I have the following situation now:

          raidz2                                      ONLINE       0     0     0
            c10t600A0B800011399600007CF945E80E95d0    ONLINE       0     0     0
            c10t600A0B8000115EA20000FEEB45E8145Ed0    ONLINE       0     0     0
            spare                                     ONLINE       0     0     0
              c10t600A0B800011399600007D2345E81075d0  ONLINE       0     0     0
              c10t600A0B800011399600007CE145E80D4Dd0  ONLINE       0     0     0
            c10t600A0B800011399600007D3F45E81157d0    ONLINE       0     0     0
            c10t600A0B8000115EA20000FF1145E817BEd0    ONLINE       0     0     0
            c10t600A0B800011399600007D5D45E813EBd0    ONLINE       0     0     0
            c10t600A0B8000115EA20000FE7D45E80CDEd0    ONLINE       0     0     0
            c10t600A0B800011399600007C6145E808C7d0    ONLINE       0     0     0
            c10t600A0B8000115EA20000FE9945E80E6Ad0    ONLINE       0     0     0
            c10t600A0B800011399600007C8B45E80A59d0    ONLINE       0     0     0
            c10t600A0B800011399600007CA745E80B21d0    ONLINE       0     0     0
            c10t600A0B8000115EA20000FEB545E810D4d0    ONLINE       0     0     0
            c10t600A0B800011399600007CD145E80CD7d0    ONLINE       0     0     0
            c10t600A0B8000115EA20000FED145E8129Cd0    ONLINE       0     0     0
            c10t600A0B800011399600007CFB45E80EA5d0    ONLINE       0     0     0
            c10t600A0B8000115EA20000FEED45E8146Ed0    ONLINE       0     0     0

        spares
          c10t600A0B800011399600007CE145E80D4Dd0      INUSE     currently in use
          c10t600A0B8000115EA20000FEE145E81328d0      AVAIL   
          c10t600A0B800011399600007D0B45E80F21d0      AVAIL   
          c10t600A0B8000115EA20000FEFD45E81506d0      AVAIL   
          c10t600A0B800011399600007D3545E81107d0      AVAIL   
          c10t600A0B800011399600007D5345E81289d0      AVAIL   
          c10t600A0B8000115EA20000FF2345E81864d0      AVAIL   
          c10t600A0B800011399600007D6F45E8149Bd0      AVAIL   

I thought if I replace the original device with the same (but new) lun that 
should work, but I get
the following:

zpool replace tank c10t600A0B800011399600007D2345E81075d0 
c10t600A0B800011399600007D2345E81075d0
invalid vdev specification
use '-f' to override the following errors:
/dev/dsk/c10t600A0B800011399600007D2345E81075d0s0 is part of active ZFS pool 
tank. Please see zpool(1M).

The above lun is the same as before, but the underlying disk on the storage was 
replaced.  Do I need to do something to this lun to make ZFS think it is a new 
disk?  Or should I be doing something different?

I now have another disk which has gone back so I need to fix my above situation 
with the hot
spare first, and then go through the process again for my second failure.

The environment is S10U4, running on a x4600 with Flexline 380 storage units.

Tia,

David
 
 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to