Freddie Witherden <[email protected]> writes: > -------- Original Message -------- > Subject: Re: [PyCUDA] Contexts and Threading > Date: Sat, 29 Sep 2012 19:08:09 +0200 > From: Eelco Hoogendoorn <[email protected]> > To: Freddie Witherden <[email protected]> > > > > Actually; Seems I should RTFM; see the pycuda FAQ > Combining threads and streams does not seem to work at all (or I am doing > something really stupid). Seems like you need to init the context in the > thread, and can not share it between them. > > At least for the thing I have in mind, creating a context per thread > wouldn’t > really make sense; a context has a huge overhead, and trying to get > multiple > contexts to play nicely on the same device at the same time has so far > eluded me as well. > > That is rather disappointing, as it seems there is no way around the hacky > state machine stream nonsense, if you want to run a lot of small kernels in > parallel (I am thinking millions of calls, each of which would be lucky to > saturate a single SMP) > > Am I missing something?
Does CUDA allow multiple threads to be attached to a context nowadays? Note all of this sharing between threads is a non-issue (i.e. works without a problem) in OpenCL. Currently, each thread needs its own kernel, though... (see message on PyOpenCL list that I'll send in a bit). Andreas _______________________________________________ PyCUDA mailing list [email protected] http://lists.tiker.net/listinfo/pycuda
