Hello, thanks for your suggestion. I tried settin zfs_arc_max to 0x30000000 
(768MB, out of 3GB). The system ran for almost 45 minutes before it froze. 

Here's an interesting piece of arcstat.pl, which I noticed just as it was 
pasing by:


    Time  read  miss  miss%  dmis  dm%  pmis  pm%  mmis  mm%  arcsz     c
15:17:41   152   152    100   152  100     0    0   152  100     2G  805M
15:17:42   139   139    100   139  100     0    0   139  100     2G  805M
<State Changed>
15:17:43   188   188    100   188  100     0    0   188  100     2G  805M
15:17:44   150   150    100   150  100     0    0   150  100     2G  805M
15:17:45   151   151    100   151  100     0    0   151  100     2G  805M
15:17:46   149   149    100   149  100     0    0   149  100     2G  805M
15:17:47   161   161    100   161  100     0    0   161  100     2G  805M
15:17:48   153   153    100   153  100     0    0   153  100     2G  219M
15:17:49   140   140    100   140  100     0    0   140  100     2G  100M
15:17:50   143   143    100   143  100     0    0   143  100     2G  100M
15:17:51   145   145    100   145  100     0    0   145  100     2G  100M

notice how it suddenly drops "c" from 805M to 100M in 2 seconds. Also "arcsz" 
is 2G, which is weird because it shouldn't grow beyond 0x30000000 (768M), 
right? And it's also weird to also get 100% MISS ratio

Here's "top" just before it froze:

last pid:  5253;  load avg:  0.47,  0.37,  0.33;       up 0+00:44:53     
15:20:14
77 processes: 75 sleeping, 1 running, 1 on cpu
CPU states: 57.5% idle,  1.0% user, 41.6% kernel,  0.0% iowait,  0.0% swap
Memory: 3072M phys mem, 28M free mem, 2055M swap, 1994M free swap

   PID USERNAME LWP PRI NICE  SIZE   RES STATE    TIME    CPU COMMAND
  1248 root       1  59    0 5940K 2736K sleep    0:14  0.82% arcstat.pl
  5206 root       9  59    0   47M 4892K sleep    0:01  0.35% java
   855 root       2  59    0 5076K 1588K sleep    0:09  0.33% apcupsd
  3134 root       1  59    0 5152K 1764K sleep    0:02  0.26% zpool
  1261 root       1  59    0 4104K  588K cpu      0:03  0.22% top
  3125 root       1  59    0 6352K 1536K sleep    0:00  0.06% sshd
  1151 root       1  59    0 6352K 1504K sleep    0:00  0.05% sshd
    62 root       1  59    0 1832K  540K sleep    0:01  0.05% powernowd
   849 root       1  59    0   11M 1100K sleep    0:00  0.05% snmpd
   465 proxy      1  59    0   15M 2196K run      0:00  0.04% squid
   271 daemon     1  59    0 6652K  264K sleep    0:00  0.03% rcapd
  1252 root       1  59    0 6352K 1292K sleep    0:00  0.02% sshd
     7 root      14  59    0   12M 5412K sleep    0:04  0.02% svc.startd
   880 root       1  59    0 6276K 2076K sleep    0:00  0.02% httpd
   847 root       1  59    0 2436K 1148K sleep    0:00  0.02% dhcpagent

and finally, zpool iostat 1:

tera        1.51T   312G    207      0  1.22M      0
tera        1.51T   312G    141      0   854K      0
tera        1.51T   312G     70      0   427K      0
tera        1.51T   312G    204      0  1.20M      0
tera        1.51T   312G    187      0  1.10M      0
tera        1.51T   312G    179      0  1.05M      0
tera        1.51T   312G    120      0   743K      0
tera        1.51T   312G     94      0   580K      0
tera        1.51T   312G     77      0   471K      0
tera        1.51T   312G    115      0   696K      0

Which shows a very poor read performance, for a 4xSATA2 (this array usually 
saturates my gigabit ethernet). And it's not that the kernel is processing that 
much data because the CPU is 57% idle and I THINK powernowd is making it run at 
900MHz.

HernĂ¡n
 
 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to