I've started the discussion here
http://apache-ignite-developers.2346864.n4.nabble.com/Proposed-Release-2-7-7-2-7-6-ML-extra-fixed-bugs-td43537.html


пт, 13 сент. 2019 г. в 22:12, Alexey Zinoviev <zaleslaw....@gmail.com>:

> The reason was that the last year there is no significant releases of
> Ignite between 2.7 and 2.8, only minor releases with long story of renaming.
> I am and another ML guys are ready in 1-2 months prepare ML module for 2.8
> or for the minor release 2.7.7 = 2.7.6 + updated ML + new fixed bugs
>
> Let's discuss it in separate thread next week
>
>
>
> пт, 13 сент. 2019 г. в 21:55, Denis Magda <dma...@apache.org>:
>
>> Alexey, I'm wondering,
>>
>> Are there any dependencies on Ignite Core that make us put off the ML
>> changes release until 2.8? I know that you do not support the idea of ML
>> as
>> a separate Ignite module but this concept would allow us to release ML as
>> frequently as we want not being blocked by Ignite core releases.
>>
>>
>> -
>> Denis
>>
>>
>> On Fri, Sep 13, 2019 at 11:45 AM Alexey Zinoviev <zaleslaw....@gmail.com>
>> wrote:
>>
>> > I could answer as one of developers of ML module.
>> > Currently is available the ML in version 2.7.5, it supports a lot of
>> > algorithms and could be used in production, but the API is not stable
>> and
>> > will be changed in 2.8
>> >
>> > The ML module will be stable since next release 2.8, also we have no
>> > performance report to compare for example with Spark ML
>> > Based on my exploration the performance of in terms of Big O notation is
>> > the same like in Spark ML (real numbers says that Ignite ML is more
>> faster
>> > than Spark ML due to Ignite in-memory nature and so on)
>> >
>> > Since 2.8 it will have good integration with TensorFlow, Spark ML,
>> XGBoost
>> > via model inference.
>> >
>> > You as a user have no ability to run, for-example scikit-learn or R
>> > packages in distributed mode over Ignite, but you could run the
>> TensorFlow,
>> > using Ignite as a distributed back-end instead of Horovod.
>> >
>> > If you have any questions, please let me know
>> >
>> >
>> >
>> > пт, 13 сент. 2019 г. в 21:28, Denis Magda <dma...@apache.org>:
>> >
>> >> David,
>> >>
>> >> Let me loop in Ignite dev list that has Ignite ML experts subscribed.
>> >> Please, could you share more details in regards to your performance
>> >> testing
>> >> and objectives for Ignite ML overall?
>> >>
>> >> The module is ready for production and we're ready to help solve any
>> >> cornerstones.
>> >>
>> >> -
>> >> Denis
>> >>
>> >>
>> >> On Fri, Sep 6, 2019 at 4:50 AM David Williams <leeon2...@gmail.com>
>> >> wrote:
>> >>
>> >> > Python is 25 times slower than Java for ML at runtimes, which I found
>> >> out
>> >> > online. But I don't know that statement is true or not. I need
>> insiders'
>> >> > opinion.  Which ml other packages are best options for Ignite?
>> >> >
>> >> > On Fri, Sep 6, 2019 at 7:28 PM Mikael <mikael-arons...@telia.com>
>> >> wrote:
>> >> >
>> >> >> Hi!
>> >> >>
>> >> >> I have never used it myself but it's been there for long time and I
>> >> >> would expect it to be stable, and yes it will run distributed, I
>> can't
>> >> >> say anything about performance as I have never used it.
>> >> >>
>> >> >> You will find a lot of more information at:
>> >> >>
>> >> >> https://apacheignite.readme.io/docs/machine-learning
>> >> >>
>> >> >> Mikael
>> >> >>
>> >> >>
>> >> >> Den 2019-09-06 kl. 11:50, skrev David Williams:
>> >> >> >
>> >> >> >
>> >> >> > I am evaluating ML framework for Java platform. I knew Ignite has
>> ML
>> >> >> > package.
>> >> >> > But I like to know its stability and performance for production.
>> Can
>> >> >> > Ignite
>> >> >> > ML code run in distribute way?
>> >> >> >
>> >> >> > Except its own ML package, which ml packages are best options for
>> >> >> Ignite?
>> >> >>
>> >> >
>> >>
>> >
>>
>

Reply via email to