walterddr commented on issue #10919:
URL: https://github.com/apache/pinot/issues/10919#issuecomment-1608653404
CPU solutions only make sense in certain scenarios IMO and I am not sure if
those are fit.
- Q: can it perform significantly better in specific use cases, for example
ANNS use cases that the setup/GPU I/O overhead outweighs the batch performance
on the GPU.
- Q: can we use an algorithm that doesn't depend on product quantization (or
any that specifically designed to leverage the large parallelism of GPU size)
for example graph search algo that performs.
- this also echoes back to Q1 b/c most likely these branching algorithm
are not good for batching
- Q: would we perform significantly cheaper while still maintain the equal
amount of performance? and is there a use cases similar to that (for example
ad-hoc exploration of the dataset before massively scaled up when GPU is
justify)
specifically Pinot, i knew that most of the vector databases leverage
"inverted index" mechanism to speed up the ANNS algorithm. i don't think that's
identical to the inverted index we have in Pinot but we should see if the
indexing framework after index-spi is introduce can be used.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]