Hi everyone,

When trying to speed up my code, I noticed that simply by reordering my
data I could get more than twice as fast for the simplest operations:
```
import numpy
a = numpy.random.rand(50, 50, 50)

%timeit a[0] + a[1]
1000000 loops, best of 3: 1.7 µs per loop

%timeit a[:, 0] + a[:, 1]
100000 loops, best of 3: 4.42 µs per loop

%timeit a[..., 0] + a[..., 1]
100000 loops, best of 3: 5.99 µs per loop
```
This makes sense considering the fact that, by default, NumPy features
C-style memory allocation: the last index runs fastest. The blocks that are
added with `a[0] + a[1]` are contiguous in memory, so cache is nicely made
use of. As opposed to that, the blocks that are added with `a[:, 0] + a[:,
1]` are not contiguous, even more so with `a[..., 0] + a[..., 1]`; hence
the slowdown. Would that be the correct explanation?

If yes, I'm wondering why most numpy.linalg methods, when vectorized, put
the vector index up front. E.g., to mass-compute determinants, one has to do
```
a = numpy.random.rand(777, 3, 3)
numpy.linalg.det(a)
```
This way, all 3x3 matrices form a memory block, so if you do `det` block by
block, that'll be fine. However, vectorized operations (like `+` above)
will be slower than necessary.
Any background on this?

(I came across this when having to rearrange my data (swapaxes, rollaxis)
from shape (3, 3, 777) (which allows for fast vectorized operations in the
rest of the code) to (777, 3, 3) for using numpy's svd.)

Cheers,
Nico
_______________________________________________
NumPy-Discussion mailing list
NumPy-Discussion@scipy.org
https://mail.scipy.org/mailman/listinfo/numpy-discussion

Reply via email to