I think it would be good to do a minor release to support Apache Solr as
long as it doesn't put too much burden on any of us at the moment.

Thanks,
Jeff

On Fri, Oct 13, 2023 at 1:55 AM Atita Arora <atitaar...@gmail.com> wrote:

> Sounds fair  and reasonable to me.
>
> Thanks,
> Atita
>
> On Thu, 12 Oct 2023, 21:50 Richard Zowalla, <r...@apache.org> wrote:
>
> > I am fine with the suggested approach.
> >
> > Maybe, we should do a minor release after this is merged in order to
> > unblock the SOLR folks?
> >
> > Gruß
> > Richard
> >
> > Am Donnerstag, dem 12.10.2023 um 12:41 -0400 schrieb Jeff Zemerick:
> > > Hi all,
> > >
> > > I created OPENNLP-1515 to change the ONNX Runtime dependency from
> > > onnxruntime-gpu to onnxruntime. This change will remove GPU support
> > > and
> > > cause OpenNLP to always use CPU for inference. The reason for this
> > > change
> > > is the onnxruntime dependency supports Linux, Windows, and Mac x64,
> > > and the
> > > onnxruntime-gpu dependency only supports Linux and Windows. (
> > > https://onnxruntime.ai/docs/get-started/with-java.html) I think
> > > OpenNLP
> > > should support the most operating systems out of the box instead of
> > > favoring GPU. Please take a look at the pull request:
> > > https://github.com/apache/opennlp/pull/551
> > >
> > > This change is partially proposed to support OpenNLP's ONNX
> > > integration in
> > > Apache Solr: https://github.com/apache/solr/pull/1999
> > >
> > > I think GPU support in OpenNLP should be easily accessible to users,
> > > so I
> > > wrote OPENNLP-1516 to capture that with a link to one possible
> > > method. If
> > > the above PR is merged, a user can still enable GPU in OpenNLP by
> > > manually
> > > replacing the onnxruntime.jar with onnxruntime-gpu.jar on their
> > > classpath
> > > until OPENNLP-1516 is resolved.
> > >
> > > All comments/suggestions are welcome!
> > >
> > > Thanks,
> > > Jeff
> >
> >
>

Reply via email to