Sounds fair and reasonable to me. Thanks, Atita
On Thu, 12 Oct 2023, 21:50 Richard Zowalla, <r...@apache.org> wrote: > I am fine with the suggested approach. > > Maybe, we should do a minor release after this is merged in order to > unblock the SOLR folks? > > Gruß > Richard > > Am Donnerstag, dem 12.10.2023 um 12:41 -0400 schrieb Jeff Zemerick: > > Hi all, > > > > I created OPENNLP-1515 to change the ONNX Runtime dependency from > > onnxruntime-gpu to onnxruntime. This change will remove GPU support > > and > > cause OpenNLP to always use CPU for inference. The reason for this > > change > > is the onnxruntime dependency supports Linux, Windows, and Mac x64, > > and the > > onnxruntime-gpu dependency only supports Linux and Windows. ( > > https://onnxruntime.ai/docs/get-started/with-java.html) I think > > OpenNLP > > should support the most operating systems out of the box instead of > > favoring GPU. Please take a look at the pull request: > > https://github.com/apache/opennlp/pull/551 > > > > This change is partially proposed to support OpenNLP's ONNX > > integration in > > Apache Solr: https://github.com/apache/solr/pull/1999 > > > > I think GPU support in OpenNLP should be easily accessible to users, > > so I > > wrote OPENNLP-1516 to capture that with a link to one possible > > method. If > > the above PR is merged, a user can still enable GPU in OpenNLP by > > manually > > replacing the onnxruntime.jar with onnxruntime-gpu.jar on their > > classpath > > until OPENNLP-1516 is resolved. > > > > All comments/suggestions are welcome! > > > > Thanks, > > Jeff > >