using mssql, the code itself is in gluon scheduler.py - this happens with no interaction from the app
On Thursday, January 26, 2017 at 12:03:41 PM UTC-5, Dave S wrote: > > > > On Thursday, January 26, 2017 at 8:44:25 AM UTC-8, Jason Solack wrote: >> >> So the issue is we run 6 workers on a machine and it works. If we do 3 >> workers on 2 machines we get deadlocks. That is no exaggeration - 6 >> records in our worker table and we're getting dealocks. >> >> > Which DB are you using? Can you show your relevant code? > > /dps > > >> On Wednesday, January 25, 2017 at 3:05:37 AM UTC-5, Niphlod wrote: >>> >>> you *should* have one different db for each environment. Each scheduler >>> tied to the same db will process incoming tasks, and it doesn't matter what >>> app effectively pushes them. >>> This is good if you want to have a single scheduler (which can be >>> composed by several workers) serving many apps, but *generally* you don't >>> want to *merge* prod and beta apps. >>> >>> The is_ticker bit is fine: only one worker tied to a db is elegible to >>> be a ticker, which is the one process than manages asssigning tasks (to >>> itself AND to other available workers). >>> Locking, once in a while, can happen and is self-healed. Continuous >>> locking is not good: either you have too many workers tied to the db OR >>> your db isn't processing concurrency at the rate that it needs. >>> SQLite can handle at most 2 or 3 workers. All the other "solid" backends >>> can manage up to 10, 15 at most. >>> If you wanna go higher, you need to turn to the redis-backed scheduler. >>> >>> On Tuesday, January 24, 2017 at 10:59:31 PM UTC+1, Jason Solack wrote: >>>> >>>> Hello all, >>>> >>>> I'm having some re-occurring issue with the scheduler. We are >>>> currently running multiple environments (production, beta) and have >>>> several >>>> nodes in each environment. If we have scheduler services running on all >>>> machines on each node we get a lot of deadlock errors. If we drop each >>>> environment down to one node we get no deadlock errors. I am noticing the >>>> field "is_ticker" in the worker table will only have one ticker across all >>>> the workers (spanning environments). Is that the expected behavior? I >>>> don't see any documentation about the ticker field so i'm not sure what to >>>> expect from that. >>>> >>>> Also is there any best practices about running the scheduler in an >>>> environment that i've described? >>>> >>>> Thanks in advance >>>> >>>> Jason >>>> >>> -- Resources: - http://web2py.com - http://web2py.com/book (Documentation) - http://github.com/web2py/web2py (Source code) - https://code.google.com/p/web2py/issues/list (Report Issues) --- You received this message because you are subscribed to the Google Groups "web2py-users" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. For more options, visit https://groups.google.com/d/optout.

