Hi list,

I've a strange behaviour with autoreplace property. It is set to off by default, ok. I want to manually manage disk replacement so default "off" matches my need.

# zpool get autoreplace mypool
NAME   PROPERTY     VALUE    SOURCE
mypool  autoreplace  off      default

Then I added 2 spare disks.

        spares
          c1t18d0      AVAIL
          c1t19d0      AVAIL

Ok, fine.

Then I had failures with 1 disk of the pool and can see in logs the following :

------------------------
DESC: The number of I/O errors associated with a ZFS device exceeded
acceptable levels. Refer to http://sun.com/msg/ZFS-8000-FD for more information. AUTO-RESPONSE: The device has been offlined and marked as faulted. An attempt will be made to activate a hot spare if available.
-----------------------

This is where my problem occurs , zfs automatically replaced faulted disk by a spare ! even with autoreplace=off

# zpool status
  pool: mypool
 state: DEGRADED
status: One or more devices are faulted in response to persistent errors.
        Sufficient replicas exist for the pool to continue functioning in a
        degraded state.
action: Replace the faulted device, or use 'zpool clear' to mark the device
        repaired.
scrub: resilver completed after 0h0m with 0 errors on Thu Feb 4 00:10:25 2010
config:

        NAME           STATE     READ WRITE CKSUM
        mypool         DEGRADED     0     0     0
          mirror       ONLINE       0     0     0
            c0t2d0     ONLINE       0     0     0
            c0t3d0     ONLINE       0     0     0
            c0t4d0     ONLINE       0     0     0
            c0t5d0     ONLINE       0     0     0
          mirror       DEGRADED     0     0     0
            c0t6d0     ONLINE       0     0     0
            c0t7d0     ONLINE       0     0     0
            spare      DEGRADED     4     0     0
              c1t8d0   FAULTED      3    26     0  too many errors
              c1t18d0  ONLINE       0     0     4  56K resilvered
            c1t9d0     ONLINE       0     0     0
          mirror       ONLINE       0     0     0
            c1t10d0    ONLINE       0     0     0
            c1t11d0    ONLINE       0     0     0
            c1t12d0    ONLINE       0     0     0
            c1t13d0    ONLINE       0     0     0
          mirror       ONLINE       0     0     0
            c1t14d0    ONLINE       0     0     0
            c1t15d0    ONLINE       0     0     0
            c1t16d0    ONLINE       0     0     0
            c1t17d0    ONLINE       0     0     0
        cache
          c2d0         ONLINE       0     0     0
          c3d0         ONLINE       0     0     0
        spares
          c1t18d0      INUSE     currently in use
          c1t19d0      AVAIL

errors: No known data errors


Any idea why it has been done automatically ?

solaris 10U8 Generic_141445-09 - zpool version 15 - zfs version 4


Thx for your answers.

--
Francois
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to