[ https://issues.apache.org/jira/browse/IGNITE-10286?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Alexey Zinoviev updated IGNITE-10286: ------------------------------------- Fix Version/s: (was: 2.8) 2.9 > [ML] Umbrella: Model serving > ---------------------------- > > Key: IGNITE-10286 > URL: https://issues.apache.org/jira/browse/IGNITE-10286 > Project: Ignite > Issue Type: New Feature > Components: ml > Reporter: Yury Babak > Assignee: Alexey Zinoviev > Priority: Major > Fix For: 2.9 > > > We want to have convenient API for model serving. It means that we need a > mechanism for storing models and infer them inside Apache Ignite. > For now, I see 2 important features - distributed storage for any models and > inference. > From my point of view, we could use some built-in(predefined) cache as model > storage. And use service grid for model inference. We could implement some > "ModelService" for access to our storage, receive the list of all suitable > model(including model metrics and some other information about a model), > choose one(or several) and infer it from this service. > Model from TF should also use the same mechanisms for storing and inference. -- This message was sent by Atlassian Jira (v8.3.4#803005)