Would a work stealing approach work better for you here? Then the only signalling overhead would be when a core runs out of work
On Thu, 19 Aug 2021, 05:36 Stephen J. Turnbull, < turnbull.stephen...@u.tsukuba.ac.jp> wrote: > Christopher Barker writes: > > > The worker pool approach is probably the way to go, but there is a fair > bit > > of overhead to creating a multiprocessing job. So fewer, larger jobs are > > faster than many small jobs. > > True, but processing those rows would have to be awfully fast for the > increase in overhead from 16 chunks x 10^6 rows/chunk to 64 chunks x > 250,000 rows/chunk to matter, and that would be plenty granular to > give a good approximation to his 2 chunks by fast core : 1 chunk by > slow core nominal goal with a single queue, multiple workers > approach. (Of course, it almost certainly will do a lot better, since > 2 : 1 was itself a very rough approximation, but the single queue > approach adjusts to speed differences automatically.) > > And if it's that fast, he could do it on a single core, and still done > by the time he's finished savoring a sip of coffee. ;-) > > Steve > _______________________________________________ > Python-ideas mailing list -- python-ideas@python.org > To unsubscribe send an email to python-ideas-le...@python.org > https://mail.python.org/mailman3/lists/python-ideas.python.org/ > Message archived at > https://mail.python.org/archives/list/python-ideas@python.org/message/TCC7ZZLP7YMOCWSKIC2KXQQVBKT3UIMZ/ > Code of Conduct: http://python.org/psf/codeofconduct/ >
_______________________________________________ Python-ideas mailing list -- python-ideas@python.org To unsubscribe send an email to python-ideas-le...@python.org https://mail.python.org/mailman3/lists/python-ideas.python.org/ Message archived at https://mail.python.org/archives/list/python-ideas@python.org/message/XURUCC3WOCZ3UT6W2FHTJ6NGLCBR5WDO/ Code of Conduct: http://python.org/psf/codeofconduct/