Hi

strange problem on a x4440:

  pool: rpool
 state: ONLINE
 scrub: none requested
config:

        NAME                         STATE     READ WRITE CKSUM
        rpool                        ONLINE       0     0     0
          mirror-0                   ONLINE       0     0     0
            c1t5000CCA00A3E1D98d0s0  ONLINE       0     0     0
            c1t5000CCA00A3D1BE4d0s0  ONLINE       0     0     0

errors: No known data errors

root@x4440 # zpool iostat -v rpool 15 2
                                capacity     operations    bandwidth
pool                         alloc   free   read  write   read  write
---------------------------  -----  -----  -----  -----  -----  -----
rpool                        65.5G  70.5G      0    333  55.5K  35.4M
  mirror                     65.5G  70.5G      0    333  55.5K  35.4M
    c1t5000CCA00A3E1D98d0s0      -      -      0    300  28.3K  35.4M
    c1t5000CCA00A3D1BE4d0s0      -      -      0    300  28.3K  35.4M
---------------------------  -----  -----  -----  -----  -----  -----

                                capacity     operations    bandwidth
pool                         alloc   free   read  write   read  write
---------------------------  -----  -----  -----  -----  -----  -----
rpool                        65.5G  70.5G      0    643      0  65.8M
  mirror                     65.5G  70.5G      0    643      0  65.8M
    c1t5000CCA00A3E1D98d0s0      -      -      0    559      0  65.8M
    c1t5000CCA00A3D1BE4d0s0      -      -      0    557      0  65.7M
---------------------------  -----  -----  -----  -----  -----  -----

(2 10k SAS drives)

iosnoop does not really tell me something useful:

root@x4440 # ./iosnoop -d sd32 -d sd33 -A | head
TIME STIME DELTA DEVICE INS MAJ MIN UID PID PPID D BLOCK SIZE MOUNT FILE PATH 
COMM ARGS
1812818090420 1812818079998 10421 sd33 33 32 2112 0 5 0 W 183388660 131072 
<none> <none> <none> zpool-rpool zpool-rpool\0
1812818091454 1812818082162 9292 sd33 33 32 2112 0 5 0 W 183388916 131072 
<none> <none> <none> zpool-rpool zpool-rpool\0
1812818098863 1812818072105 26757 sd33 33 32 2112 0 5 0 W 183386612 131072 
<none> <none> <none> zpool-rpool zpool-rpool\0
1812818099938 1812818073155 26783 sd33 33 32 2112 0 5 0 W 183386868 131072 
<none> <none> <none> zpool-rpool zpool-rpool\0
1812818101468 1812818074220 27247 sd33 33 32 2112 0 5 0 W 183387124 131072 
<none> <none> <none> zpool-rpool zpool-rpool\0
1812818102540 1812818075278 27262 sd33 33 32 2112 0 5 0 W 183387380 131072 
<none> <none> <none> zpool-rpool zpool-rpool\0
1812818103615 1812818076822 26792 sd33 33 32 2112 0 5 0 W 183387636 131072 
<none> <none> <none> zpool-rpool zpool-rpool\0
1812818104689 1812818077894 26794 sd33 33 32 2112 0 5 0 W 183387892 131072 
<none> <none> <none> zpool-rpool zpool-rpool\0
1812818105738 1812818078971 26767 sd33 33 32 2112 0 5 0 W 183388148 131072 
<none> <none> <none> zpool-rpool zpool-rpool\0

PID 5 is just 
    root     5     0   0   Aug 25 ?        3102:27 zpool-rpool

to round it up:

root@x4440 # ./arc_summary.pl 
System Memory:
         Physical RAM:  131063 MB
         Free Memory :  13430 MB
         LotsFree:      1991 MB

ZFS Tunables (/etc/system):

ARC Size:
         Current Size:             89931 MB (arcsize)
         Target Size (Adaptive):   97854 MB (c)
         Min Size (Hard Limit):    16254 MB (zfs_arc_min)
         Max Size (Hard Limit):    130039 MB (zfs_arc_max)

ARC Size Breakdown:
         Most Recently Used Cache Size:          99%    97296 MB (p)
         Most Frequently Used Cache Size:         0%    557 MB (c-p)

ARC Efficency:
         Cache Access Total:             942093365
         Cache Hit Ratio:      99%       941013507      [Defined State for 
buffer]
         Cache Miss Ratio:      0%       1079858        [Undefined State for 
Buffer]
         REAL Hit Ratio:       96%       908987855      [MRU/MFU Hits Only]

         Data Demand   Efficiency:    99%
         Data Prefetch Efficiency:    73%

        CACHE HITS BY CACHE LIST:
          Anon:                        3%        31367306               [ New 
Customer, First Cache Hit ]
          Most Recently Used:         44%        417379490 (mru)        [ 
Return Customer ]
          Most Frequently Used:       52%        491608365 (mfu)        [ 
Frequent Customer ]
          Most Recently Used Ghost:    0%        635995 (mru_ghost)     [ 
Return Customer Evicted, Now Back ]
          Most Frequently Used Ghost:  0%        22351 (mfu_ghost)      [ 
Frequent Customer Evicted, Now Back ]
        CACHE HITS BY DATA TYPE:
          Demand Data:                91%        858469425 
          Prefetch Data:               0%        2843857 
          Demand Metadata:             5%        48740552 
          Prefetch Metadata:           3%        30959673 
        CACHE MISSES BY DATA TYPE:
          Demand Data:                 2%        24982 
          Prefetch Data:              94%        1023544 
          Demand Metadata:             1%        16042 
          Prefetch Metadata:           1%        15290 
---------------------------------------------


Has nayone any idea what's going on here?

Cheers

carsten
-- 
Dr. Carsten Aulbert - Max Planck Institute for Gravitational Physics
Callinstrasse 38, 30167 Hannover, Germany
Phone/Fax: +49 511 762-17185 / -17193
http://www.top500.org/system/9234 | http://www.top500.org/connfam/6
CaCert Assurer | Get free certificates from http://www.cacert.org/

Attachment: smime.p7s
Description: S/MIME cryptographic signature

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to