As nice as numba is, it’s a very heavy dependency that many high profile libraries want to avoid.

Best
Christian 

Am 18.08.2025 um 14:25 schrieb Kevin Sheppard <kevin.k.shepp...@gmail.com>:


This is a good place for numba if performance and memory use are a concern.

Good NumPy
125 ms ± 2.42 ms per loop (mean ± std. dev. of 7 runs, 10 loops each)

Naive numba (memory efficient)
386 ms ± 1.83 ms per loop (mean ± std. dev. of 7 runs, 1 loop each)

Parallel numba (memory efficient, cpu bounded)
46.8 ms ± 1.12 ms per loop (mean ± std. dev. of 7 runs, 10 loops each)


All were on a 12 core Ryzen.

Kevin

On Mon, Aug 18, 2025 at 12:35 PM Christian Lorentzen via NumPy-Discussion <numpy-discussion@python.org> wrote:

If W is just a vector, i.e. my main use case, than what you write is what I wanted to imply and actually how I implemented it in several cases.
The point is that X.T * weights[None, :] still doubles memory (same size as X) and this solution doesn't exploit the symmetry.

Best
Christian

On 17.08.2025 22:13, Kevin Jacobs wrote:
Why not:

XTWX = (X.T * weights[None, :]) @ X

if `weights` is a vector?  Allocating `diag(weights)` is the big memory and CPU hog at least for my generally long and skinny X matrices.

-Kevin


On Fri, Aug 15, 2025 at 10:08 AM Christian Lorentzen via NumPy-Discussion <numpy-discussion@python.org> wrote:

Dear numpy community

I would like to propose a mew feature in np.linalg to compute weighted gram matrix (sandwich product), see https://github.com/numpy/numpy/issues/29559.

Proposed new feature or change:

The solvers for many important statistical models like Generalized Linear Models often need to compute X.T @ W @ X where X is a 2-dimensional array of features (features in columns, observations in rows) and W is very often a diagonal array of (current) weights. This computation is usually the main computational bottleneck.

It would be great if numpy could provide an efficient implementation of it, e.g. np.linalg.sandwicht_product(X, weight=w).

Why numpy? It seems even more unrealistic to me, to get it into BLAS implementations. Numpy has support for SIMD via Highways.

Computational alternatives

Drawbacks of (X.T * diag(W)) @ X:

  • Additional memory allocation to compute X.T @ diag(W) , same size as (usually large) X.
  • The result is symmetric, but this fact is not used. So at least a factor of 2 is possible.
    Note that without weights, X.T @ X uses BLAS syrk. But the weights are crucial.

Drawback of Z = np.sqrt(diag(W))[:, None] * X and then Z @ Z:

  • Additional memory allocation for Z, same size as (usually large) X.
  • Taking square roots for possibly large diag(W)

Additional information

https://github.com/Quantco/tabmat has an implementation of it with XSIMD.


Best
Christian Lorentzen

_______________________________________________
NumPy-Discussion mailing list -- numpy-discussion@python.org
To unsubscribe send an email to numpy-discussion-le...@python.org
https://mail.python.org/mailman3//lists/numpy-discussion.python.org
Member address: jac...@bioinformed.com

_______________________________________________
NumPy-Discussion mailing list -- numpy-discussion@python.org
To unsubscribe send an email to numpy-discussion-le...@python.org
https://mail.python.org/mailman3//lists/numpy-discussion.python.org
Member address: lorentzen...@gmail.com

_______________________________________________
NumPy-Discussion mailing list -- numpy-discussion@python.org
To unsubscribe send an email to numpy-discussion-le...@python.org
https://mail.python.org/mailman3//lists/numpy-discussion.python.org
Member address: kevin.k.shepp...@gmail.com
_______________________________________________
NumPy-Discussion mailing list -- numpy-discussion@python.org
To unsubscribe send an email to numpy-discussion-le...@python.org
https://mail.python.org/mailman3//lists/numpy-discussion.python.org
Member address: arch...@mail-archive.com

Reply via email to