On 24/01/2012 11:10 p.m., Pieter De Wit wrote:
Sorry - my mail client messes with the layout, I will double space
from now on :)
Spotted a few mistakes with my suggestion:
On 24/01/2012 23:02, Pieter De Wit wrote:
<snip>
Perhaps a 9) Implement dual IO queues - I *think* the IO has been
moved into it's own thread, if not, the queuing can still be
applied. Any form of checking the cache is going to effect squid,
so how do we ensure we are idle, dual queues :) Queue 1 holds the
requests for squid, queue 2 holds the admin/clean up requests. The
IO "thread" (if not threaded), before handling an admin/clean up
request checks Queue 1 for requests, empties is *totally before*
heading into Queue 2. This will allow you to have the same caching
as now, relieving the start-up problems ? Might lead to the same
double cache of objects as above (if you make the cache writable
before the scan is done)
I wonder about priority queues every now and then. It is an
interesting idea. The I/O is currently done with pluggable modules
for various forms. DiskThreads and AIO sort of do this but are FIFO
queued in N parallel queues. Prioritised queues could be an
interesting additional DiskIO module.
Hard to implement given the current "leg work" is already done ? How
well does the current version of squid handle multicores and can this
take advantage of cores ?
Should be easy. We have not exactly checked and cocumented the DiskIO
library API. But
The current AIO handles SMP exactly as well as the system AIO library
can, same for pthreads library behind DiskThreads.
What I'm looking for is a little bit more abstracted towards the
architecture level across cache type and implementation. At that
scale we can't use any form of "totally empty" queue condition
because on caches that receive much traffic the queue would be quite
full, maybe never actually empty. Several of the problems we have
now are waiting on the cache load completed (ie the load action
queue empty) before a cache is even considered for use.
Amos
At that scale, no matter what you do, you will impact
performance/your "wanted" outcome. It's about reaching an acceptable
balance which I think, you, as a dev, will have a hard time
predicting for any real life usage out there. Perhaps "we" (in "
since I am yet to contrib a single line of code :) ) can make it
"Weighted Priority" and as such, have squid.conf options to tune it.
The Admin has to decide how aggresive squid must be at rebuilding
(makes me think of the raid rebuild options in HP RAID controllers)
the cache. I am thinking of:
cache_rebuild_weight <0-"max int"> ?
Can't be zero since we won't rebuild then, but what if we want have
more than 1 per 1, maybe we should have 2 options:
cache_rebuild_weight <1-max int>
cache_request_weight <1-max int>
?
Each cache maintains its own order and block loading algorithm(s). It
would need to be a per-cache_dir setting. It should be relatively easy
to add there.
The cache_dir can report this up to the top layer via their loading
factor when they are not servicing requests. I was considering it to
prioritise CLEAN builds before DIRTY ones or cache_dir by the speed of
its storage type and loading factor.
For every x requests, action an "admin/clean up" request, unless
"Queue 1" is empty, then drain "Queue 2"
I am also thinking of a "third" queue, something like:
Queue 1 - Write requests (depends on cache state, but has the most
impact - writes are slow)
Queue 2 - Read requests (as above, but less of an impact)
Queue 3 - Admin/Clean up
The only problem I have so far is Queue 1 is above Queue 2.....they
might be swapped since you are reading more than writing ? Perhaps
another config option.....
cache_dir /var/dir1 128G 128 128 Q1=read Q2=write (cache_dir syntax
wrong....)
cache_dir /var/dir2 32G 128 128 Q1=write Q2=read (as above, but this
might be on ssd)
I think this might be going too far ?
Cheers,
Pieter
Also, if we have the "squid.state" loaded, what stops us from writing
objects in free space, if there is ? We know how big the cache is/was
and how big it's allowed to be ? As before, this will lead to the
double storage of objects, but, this can be free'd
This would be (5) and (6). Permitting the 'done' areas of
partially-loaded caches to be used while the rest is still loading.
Amos