On 2026-02-18 05:59, ליאור בראון wrote:

I am currently working on a research project involving request dispatching and peer selection within the Squid-cache core.

During my development, I have encountered several mechanisms and logic blocks within the source code that deliberately prevent or queue parallel HTTP requests for the same URL, effectively serializing them.

I do not think Squid does that by default. Either you are misinterpreting Squid code or you are looking at code related to collapsed_forwarding feature (see squid.conf.documented). That feature is off by default (and it queues but does not serialize parallel requests).

I recommend providing at least one specific code example to illustrate/substantiate your claim.


HTH,

Alex.


I would like to understand the fundamental design rationale behind these restrictions. Specifically: Are these blocks in place due to specific architectural constraints (such as memory management or Store Entry state transitions)? Are there known side effects or risks I should be aware of if I attempt to implement parallel requesting for identical objects in a research environment? I want to ensure I fully understand the system's design philosophy before proceeding with any modifications.
Thank you for your time and insights.

_______________________________________________
squid-dev mailing list
[email protected]
https://lists.squid-cache.org/listinfo/squid-dev

Reply via email to