I've seen this error a few times, but it's not reproducible. Can
anyone give any insight into what might be going wrong?
Traceback (most recent call last):
File /home/elis/edit/work/dev/mms/common/util/threads.py, line 219, in run
mod = cudahelper.compileSourceModule(kernel.code_str,
I recall the pci bus ID behavior changing between... 4.0 and 4.1
maybe? It swapped from being a property-like thing to being a
callable, with only changing the underlying cuda version (no PyCUDA
update).
What version of cuda are you using?
Eli
On Tue, May 1, 2012 at 11:50 AM, Serge Rey
Slightly off-topic: what's the benefit of having a dedicated stack
exchange, vs. using Stack Overflow? I can only think of downsides
related to fracturing the existing set of answers and responders, esp.
for ones that aren't directly related to PyCUDA.
On topic: I don't see a reason to not just
Cool, thanks.
As a followup, the threads are working fine, but we have an off-by-one
error on some array accesses, which is probably related to why
sometimes things terminate and sometimes don't
(uninitialized/arbitrary memory). Thanks for the help!
Eli
On Fri, Mar 9, 2012 at 5:50 PM, Andreas
Thanks for all of the pointers. I'm going to hold off on MPI for now;
I'm hesitant to add additional dependencies unless they're really
needed.
The threading solution as outlined here:
http://stackoverflow.com/questions/5904872/python-multiprocessing-with-pycuda
Seems to be working well,
, Eli Stevens (Gmail)
wickedg...@gmail.com wrote:
Thanks for all of the pointers. I'm going to hold off on MPI for now;
I'm hesitant to add additional dependencies unless they're really
needed.
The threading solution as outlined here:
http://stackoverflow.com/questions/5904872/python
Hello,
I was wondering if the following will work:
- Main thread spins up thread B.
- Thread B creates a context, invokes a kernel, and creates an event.
- Event is saved.
- Thread B pops the context (kernel is still running at this point)
and finishes.
- Main thread join()s B and grabs the
...@informa.tiker.net wrote:
#part sign=pgpmime
Hi Eli,
On Thu, 8 Mar 2012 09:33:21 -0800, Eli Stevens (Gmail)
wickedg...@gmail.com wrote:
I was wondering if the following will work:
- Main thread spins up thread B.
- Thread B creates a context, invokes a kernel, and creates an event
On Thu, Aug 25, 2011 at 10:33 PM, Francis fccaba...@gmail.com wrote:
I could make use of the tens of thousands of threads in
CUDA to get the length of each substring/subarray.
The python list structure stores the length of the list already (it
increments / decrements with appends / pops, etc.),
On Thu, Aug 25, 2011 at 2:21 AM, Francis fccaba...@gmail.com wrote:
Thanks for the replies @Eli and @David. I suppose given a 'small' enough
list of sub-lists doing what I need to do in the host is good enough instead
of moving and doing the task in the device. I am just looking out for
On Wed, Aug 24, 2011 at 8:20 AM, David Mertens dcmertens.p...@gmail.com wrote:
Ah, I see. I think you might be trying to fit a round peg into a
square hole, so to speak.
I agree.
At the very least, you're going to need a loop over each list, calling
len() on it (to then stick it into an array
This is a sorta-cross-post from stack overflow:
http://stackoverflow.com/questions/6892280/how-do-i-diagnose-a-cuda-launch-failure-due-to-being-out-of-resources
I'm getting an out-of-resources error when trying to launch a CUDA
kernel (through PyCUDA), and I'm wondering if it's possible to get
I'm trying to turn a 3d numpy array with float32 data into a texture
that I can read via tex3D inside of kernel code. I have this in the
kernel:
texturefloat, cudaTextureType3D, cudaReadModeElementType my_tex;
And I have tried both:
my_texref = cuda_module.get_texref(my_tex)
my_gpu =
13 matches
Mail list logo