On what OS?
and also what is the output of the ulimit -Ha and ulimit -Sa

Eliezer

On 6/11/2013 6:32 PM, Mike Mitchell wrote:
I dropped the cache size to 150 GB instead of 300 GB.  Cached object count 
dropped
from ~7 million to ~3.5 million.  After a week I saw one occurrence of the same 
problem.
CPU usage climbed steadily over 4 hours from <10% to 100%, then squid became
unresponsive for 20 minutes.  After that it picked up as if nothing had 
happened -- no
error messages in any logs, no restarts, no core dumps.

I'm now testing again using version 3.3.5-20130607-r12573 instead of 
3.2.11-20130524-r11822.
I've left everything else the same, with the cache size still at 150 GB.

Mike Mitchell

On 30/05/2013 08:43:24 -0700, Ron Wheeler wrote:

Some ideas here.
http://www.freeproxies.org/blog/2007/10/03/squid-cache-disk-io-performance-enhancements/
http://www.gcsdstaff.org/roodhouse/?p=2784


You might try dropping your disk cache to 50Gb and see what happens.

I am not sure that caching 7 Million pages gives you much of an advantage over 1 
million. The 1,000,001th most > popular page probably does not come up that 
often and by the time you get down to a page that is 7,000,000 in the list of most 
accessed pages, you are not seeing much demand for that page.

Probably most of the cache is just accessed once.

Your cache_mem looks low but is not related to your problem but would improve 
performance a lot. Getting a few > thousand of the most active pages in memory 
is worth a lot more than 6 million of the least active pages sitting on a disk.


I am not a big squid expert but have run squid for a long time.

Ron


Reply via email to