>>> Why does squid eat 100% of processor if the problem is in FS?
> How is your cache_dir defined? aufs (in general) is a better choice
> than ufs, diskd might still have some stability issues under load, and
> coss is a good supplement as a small object cache. Conceivably if Squid
> is set up with a ufs cache_dir mounted as NFS, it's spending a lot of
> time in a wait state, blocked while the I/O completes.
For 6 days uptime:
# vmstat
procs -----------memory---------- ---swap-- -----io---- --system--
-----cpu------
r b swpd free buff cache si so bi bo in cs us sy id wa st
2 0 92 104052 235704 2309956 0 0 3 43 24 33 10 16 73 1 0
As you can see system has spent only 1% of CPU time in I/O wait.
(cpu-wa column).
My cache dir directive looks like:
cache_dir ufs /var/spool/squid 16384 64 1024
# vmstat -d
vmstat -d
disk- ------------reads------------ ------------writes----------- -----IO------
total merged sectors ms total merged sectors ms cur sec
ram0 0 0 0 0 0 0 0 0 0 0
ram1 0 0 0 0 0 0 0 0 0 0
ram2 0 0 0 0 0 0 0 0 0 0
ram3 0 0 0 0 0 0 0 0 0 0
ram4 0 0 0 0 0 0 0 0 0 0
ram5 0 0 0 0 0 0 0 0 0 0
ram6 0 0 0 0 0 0 0 0 0 0
ram7 0 0 0 0 0 0 0 0 0 0
ram8 0 0 0 0 0 0 0 0 0 0
ram9 0 0 0 0 0 0 0 0 0 0
ram10 0 0 0 0 0 0 0 0 0 0
ram11 0 0 0 0 0 0 0 0 0 0
ram12 0 0 0 0 0 0 0 0 0 0
ram13 0 0 0 0 0 0 0 0 0 0
ram14 0 0 0 0 0 0 0 0 0 0
ram15 0 0 0 0 0 0 0 0 0 0
sda 50114 8198 1197972 239114 771044 986524 13061742 1616345 0 1239
sdb 125 1430 2383 100 3 20 184 43 0 0
sdc 547181 13909 6116481 6209599 2893943 6771249 77505040 42580590 0 8027
dm-0 6659 0 143594 45401 528574 0 4228592 1248409 0 269
dm-1 13604 0 408122 82828 883993 0 7071944 3118925 0 677
dm-2 150 0 1132 387 2 0 10 2 0 0
dm-3 36240 0 639146 173982 178529 0 1428232 540632 0 229
dm-4 164 0 1136 610 35 0 76 155 0 0
dm-5 216 0 1240 817 166439 0 332884 262910 0 185
hda 0 0 0 0 0 0 0 0 0 0
fd0 0 0 0 0 0 0 0 0 0 0
md0 0 0 0 0 0 0 0 0 0 0
If it's not an I/O wait problem then what can cause squid to use 100%
of CPU core? I tried to clear cache but after an hour or so squid
began to use as much CPU as usual (~100%).
I'm not sure but maybe it started after we enlarged our outer link
from 2Mbps to 4Mbps.
I will try to move squid cache to local disk but squid works in VMware
Virtual Infrastructure. So if I move any of virtual machine partitions
from shared to local storage I wouldn't have an ability to move squid
VM from one HA cluster node to the other ('cause local partitions on
cluster nodes are different from each other).
Regards,
Nikita.