On Mon, Feb 16, 2009 at 11:02 AM, Thomas Dalton <[email protected]> wrote:
> Alternatively, the number of jobs processed per request could be made
> a function of the length of the backlog (in terms of time) - the
> longer the backlog is, the faster we process jobs. Then if the job
> queue get to being months behind we would all notice it because
> everything would start running really slowly.

Jobs are not processed on requests.  They're processed by a cron job.
You can't just automatically run them at a crazy rate, because that
will cause slave lag and other bad stuff.  If too many are
accumulating, it's probably due to a programming error that needs to
be found and fixed by human inspection.  (Tim just made several
commits fixing things that were spewing out too many jobs.)

> (Obviously, the length
> of the job queue needs to be added to whatever diagnostic screen the
> devs first check when the site slows down, otherwise it won't help
> much.)

#wikimedia-tech has enough people that regular warnings posted there
would probably get noticed.

_______________________________________________
Wikitech-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to