aaronmarkham commented on a change in pull request #11037: Website landing page 
for MMS
URL: https://github.com/apache/incubator-mxnet/pull/11037#discussion_r190747027
 
 

 ##########
 File path: docs/mms/index.md
 ##########
 @@ -0,0 +1,91 @@
+# Model Server for Apache MXNet (incubating)
+
+Model Server for Apache MXNet (incubating), otherwise known as MXNet Model 
Server (MMS), is an open source project aimed at providing a simple yet 
scalable solution for model inference. It is a set of command line tools for 
packaging model archives and serving them. The tools are written in Python, and 
have been extended to support containers for easy deployment and scaling. MMS 
also supports basic logging and advanced logging with AWS CloudWatch 
integration.
+
+
+## Multi-Framework Model Support with ONNX
+
+While the name implies that MMS is just for MXNet, it is in fact much more 
flexible, as it can support models in the [ONNX](https://onnx.ai) format. This 
means that models created and trained in PyTorch, Caffe2, or other 
ONNX-supporting frameworks can be served with MMS.
+
+To find out more about MXNet's support for ONNX models and using ONNX with 
MMS, refer to the following resources:
+
+* [MXNet-ONNX Docs](../api/python/contrib/onnx.md)
+* [Export an ONNX Model to Serve with 
MMS](https://github.com/awslabs/mxnet-model-server/docs/export_from_onnx.md)
+
+## Getting Started
+
+To install MMS with ONNX support, make sure you have Python installed, then 
for Ubuntu run:
+
+```bash
+sudo apt-get install protobuf-compiler libprotoc-dev
+pip install mxnet-model-server
+```
+
+Or for Mac run:
+
+```bash
+conda install -c conda-forge protobuf
+pip install mxnet-model-server
+```
+
+
+## Serving a Model
+
+To serve a model you must first create or download a model archive. Visit the 
[model zoo](https://github.com/awslabs/mxnet-model-server/docs/model_zoo.md) to 
browse the free models. MMS options can be explored as follows:
+
+```bash
+mxnet-model-server --help
+```
+
+Here is an easy example for serving an object classification model. You can 
use any URI and the model will be downloaded first, then served from that 
location:
+
+```bash
+mxnet-model-server \
+  --models 
squeezenet=https://s3.amazonaws.com/model-server/models/squeezenet_v1.1/squeezenet_v1.1.model
+```
+
 
 Review comment:
   added output sample

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

Reply via email to