Hi,

We're running pound with a lot of backends on a lot of different
ListenHTTP/HTTPS directives (different certificates). A grep of BackEnd and
wc -l on pound.cfg reveals 732(!).
We also use extremely high TimeOut values, it's really something our
application need.

Pound version: pound-2.4.4

When we for example patch the servers, it take ages (15-20 minutes?) for
some sites to get back up. This only seem to happen in production
environment. (FreeBSD)

I tried mimic it locally (Linux 2.6) by using the production config file,
but binding to a lot of 127.0.0.0 IP's, but the backends are detected within
a few seconds.

1) placed an iptables rule to block all backends of a specific ListenHTTP
(and ListenHTTPS) directive
2) did a GET / and of course got a 503 thrown
3) removed the iptable rules and it takes about 5-10 seconds for pound to
detect them as up and deliver the "/"

The only difference is the OS and the lack of the thousands of requests per
second that the prod env has.

Could it be that the detection of dead backends is serialized, and our
massive number in conjunction with the high timeout value has this unwanted
effect?


--
To unsubscribe send an email with subject unsubscribe to [email protected].
Please contact [email protected] for questions.

Reply via email to