This is an automated email from the ASF dual-hosted git repository.

zhasheng pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-mxnet.git


The following commit(s) were added to refs/heads/master by this push:
     new a3115d5  [DOC] update contributors and readme (#19041)
a3115d5 is described below

commit a3115d552933bf50f46a06f9c56695168b4a42cd
Author: Sheng Zha <[email protected]>
AuthorDate: Sun Aug 30 13:17:21 2020 -0700

    [DOC] update contributors and readme (#19041)
    
    * update contributors and readme
    
    * update
---
 CONTRIBUTORS.md | 159 +++++++++++++++++++++++++++++++++-----------------------
 README.md       | 158 +++++++++++++++++++++++++++++++------------------------
 2 files changed, 184 insertions(+), 133 deletions(-)

diff --git a/CONTRIBUTORS.md b/CONTRIBUTORS.md
index 4146d45..3bf9ece 100644
--- a/CONTRIBUTORS.md
+++ b/CONTRIBUTORS.md
@@ -17,76 +17,122 @@
 
 Contributors of Apache MXNet (incubating)
 =========================================
-MXNet has been developed by a community of people who are interested in 
large-scale machine learning and deep learning.
-Everyone is more than welcomed to is a great way to make the project better 
and more accessible to more users.
+Apache MXNet adopts the Apache way and governs by merit. We believe that it is 
important to create
+an inclusive community where everyone can use, contribute to, and influence 
the direction of
+the project. We actively invite contributors who have earned the merit to be 
part of the
+development community. See [MXNet Community 
Guide][https://mxnet.apache.org/community/community].
 
-Committers
-----------
-Committers are people who have made substantial contribution to the project 
and being active.
-The committers are the granted write access to the project.
 
+PMC
+---
+The Project Management Committee(PMC) consists group of active committers that 
moderate the
+discussion, manage the project release, and proposes new committer/PMC 
members. Here's the list of
+PMC members in alphabetical order by first name.
+
+* [Anirudh Subramanian](https://github.com/anirudh2290)
 * [Bing Xu](https://github.com/antinucleon)
   - Bing is the initiator and major contributor of operators and ndarray 
modules of MXNet.
-* [Tianjun Xiao](https://github.com/sneakerkg)
-  - Tianqjun is the master behind the fast data loading and preprocessing.
-* [Yutian Li](https://github.com/hotpxl)
-  - Yutian is the ninja behind the dependency and storage engine of MXNet.
-* [Mu Li](https://github.com/mli)
-  - Mu is the contributor of the distributed key-value store in MXNet.
-* [Tianqi Chen](https://github.com/tqchen)
-  - Tianqi is one of the initiator of the MXNet project.
-* [Min Lin](https://github.com/mavenlin)
-  - Min is the guy behind the symbolic magics of MXNet.
-* [Naiyan Wang](https://github.com/winstywang)
-  - Naiyan is the creator of static symbolic graph module of MXNet.
-* [Mingjie Wang](https://github.com/jermainewang)
-  - Mingjie is the initiator, and contributes the design of the dependency 
engine.
-* [Chuntao Hong](https://github.com/hjk41)
-  - Chuntao is the initiator and provides the initial design of engine.
+* [Carin Meier](https://github.com/gigasquid)
+  - Carin created and is the current maintainer for the Clojure interface.
 * [Chiyuan Zhang](https://github.com/pluskid)
   - Chiyuan is the creator of MXNet Julia package.
+* [Chris Olivier](https://github.com/cjolivier01)
+* [Dick Carter](https://github.com/DickJC123)
 * [Junyuan Xie](https://github.com/piiswrong)
 * [Haibin Lin](https://github.com/eric-haibin-lin)
+* [Henri Yandell](https://github.com/hen)
+* [Hongliang Liu](https://github.com/phunterlau)
+* [Indhu Bharathi](https://github.com/indhub)
+* [Jian Zhang](https://github.com/jzhang-zju)
+* [Joe Spisak](https://github.com/jspisak)
+* [Jun Wu](https://github.com/reminisce)
+* [Leonard Lausen](https://github.com/leezu)
+* [Liang Depeng](https://github.com/Ldpe2G)
+* [Ly Nguyen](https://github.com/lxn2)
+* [Madan Jampani](https://github.com/madjam)
+* [Marco de Abreu](https://github.com/marcoabreu)
+ - Marco is the creator of the current MXNet CI.
+* [Mu Li](https://github.com/mli)
+  - Mu is the contributor of the distributed key-value store in MXNet.
+* [Nan Zhu](https://github.com/CodingCat)
+* [Naveen Swamy](https://github.com/nswamy)
 * [Qiang Kou](https://github.com/thirdwing)
   - KK is a R ninja, he makes mxnet available for R users.
+* [Qing Lan](https://github.com/lanking520)
+* [Sandeep Krishnamurthy](https://github.com/sandeep-krishnamurthy)
+* [Sergey Kolychev](https://github.com/sergeykolychev)
+  - Sergey is original author and current maintainer of Perl5 interface.
+* [Sheng Zha](https://github.com/szha)
+* [Shiwen Hu](https://github.com/yajiedesign)
+* [Tao Lv](https://github.com/TaoLv)
+  - Tao is a major contributor to the MXNet MKL-DNN backend and performance on 
CPU.
+* [Terry Chen](https://github.com/terrychenism)
+* [Thomas Delteil](https://github.com/ThomasDelteil)
+* [Tianqi Chen](https://github.com/tqchen)
+  - Tianqi is one of the initiator of the MXNet project.
 * [Tong He](https://github.com/hetong007)
   - Tong is the major maintainer of MXNet R package, he designs the MXNet 
interface and wrote many of the tutorials on R.
+* [Tsuyoshi Ozawa](https://github.com/oza)
+* [Xingjian Shi](https://github.com/sxjscience)
+* [Yifeng Geng](https://github.com/gengyifeng)
 * [Yizhi Liu](https://github.com/yzhliu)
   - Yizhi is the main creator on mxnet scala project to make deep learning 
available for JVM stacks.
-* [Zixuan Huang](https://github.com/yanqingmen)
-  - Zixuan is one of major maintainers of MXNet Scala package.
+* [Yu Zhang](https://github.com/yzhang87)
 * [Yuan Tang](https://github.com/terrytangyuan)
   - Yuan is one of major maintainers of MXNet Scala package.
-* [Chris Olivier](https://github.com/cjolivier01)
-* [Sergey Kolychev](https://github.com/sergeykolychev)
-  - Sergey is original author and current maintainer of Perl5 interface.
-* [Naveen Swamy](https://github.com/nswamy)
-* [Marco de Abreu](https://github.com/marcoabreu)
- - Marco is the creator of the current MXNet CI.
-* [Carin Meier](https://github.com/gigasquid)
-  - Carin created and is the current maintainer for the Clojure interface.
-* [Patric Zhao](https://github.com/pengzhao-intel)
-  - Patric is a parallel computing expert and a major contributor to the MXNet 
MKL-DNN backend.
-* [Tao Lv](https://github.com/TaoLv)
-  - Tao is a major contributor to the MXNet MKL-DNN backend and performance on 
CPU.
-* [Zach Kimberg](https://github.com/zachgk)
-  - Zach is one of the major maintainers of the MXNet Scala package.
-* [Lin Yuan](https://github.com/apeforest)
-  - Lin supports MXNet distributed training using Horovod and is also a major 
contributor to higher order gradients.
+* [Yutian Li](https://github.com/hotpxl)
+  - Yutian is the ninja behind the dependency and storage engine of MXNet.
+* [Zhi Zhang](https://github.com/zhreshold)
+* [Zihao Zheng](https://github.com/zihaolucky)
+* [Ziheng Jiang](https://github.com/zihengjiang)
+* [Ziyue Huang](https://github.com/ZiyueHuang)
 
 
-### Become a Committer
-MXNet is a opensource project and we are actively looking for new committers
-who are willing to help maintaining and leading the project. Committers come 
from contributors who:
-* Made substantial contribution to the project.
-* Willing to actively spend time on maintaining and leading the project.
+Committers
+----------
+Committers are individuals who are granted the write access to the project. A 
committer is usually
+responsible for a certain area or several areas of the code where they oversee 
the code review
+process. The area of contribution can take all forms, including code 
contributions and code
+reviews, documents, education, and outreach. Committers are essential for a 
high quality and
+healthy project. The community actively look for new committers from 
contributors.
 
-New committers will be proposed by current committers, with support from more 
than two of current committers.
+* [Aaron Markham](https://github.com/aaronmarkham)
+* [Alex Zai](https://github.com/azai91)
+* [Anirudh Acharya](https://github.com/anirudhacharya)
+* [Aston Zhang](https://github.com/astonzhang)
+* [Da Zheng](https://github.com/zheng-da)
+* [Ding Kuo](https://github.com/chinakook)
+* [Hao Jin](https://github.com/haojin2)
+* [Haozheng Fan](https://github.com/hzfan)
+* [Iblis Lin](https://github.com/iblis17)
+* [Jackie Wu](https://github.com/wkcn)
+* [Jeremie Desgagne-Bouchard](https://github.com/jeremiedb)
+* [Jiajun Wang](https://github.com/arcadiaphy)
+* [Junru Shao](https://github.com/junrushao1994)
+* [Kedar Bellare](https://github.com/kedarbellare)
+* [Kellen Sunderland](https://github.com/KellenSunderland)
+* [Kevin Qin](https://github.com/ZhennanQin)
+* [Lai Wei](https://github.com/roywei)
+* [Lin Yuan](https://github.com/apeforest)
+  - Lin supports MXNet distributed training using Horovod and is also a major 
contributor to higher order gradients.
+* [Nicolas Modrzyk](https://github.com/hellonico)
+* [Patric Zhao](https://github.com/pengzhao-intel)
+* - Patric is a parallel computing expert and a major contributor to the MXNet 
MKL-DNN backend.
+* [Przemysław Trędak](https://github.com/ptrendx)
+* [Rahul Huilgol](https://github.com/rahul003)
+* [Roshani Nagmote](https://github.com/roshrini)
+* [Sam Skalicky](https://github.com/samskalicky)
+* [Steffen Rochel](https://github.com/srochel)
+* [Xi Wang](https://github.com/xidulu)
+* [Yang Shi](https://github.com/ys2843)
+* [Yuxi Hu](https://github.com/yuxihu)
+* [Zach Kimberg](https://github.com/zachgk)
+  - Zach is one of the major maintainers of the MXNet Scala package.
 
 
 List of Contributors
 --------------------
-* [Full List of 
Contributors](https://github.com/apache/incubator-mxnet/graphs/contributors)
+* [Top-100 
Contributors](https://github.com/apache/incubator-mxnet/graphs/contributors)
   - To contributors: please add your name to the list when you submit a patch 
to the project:)
 * [Feng Wang](https://github.com/happynear)
   - Feng makes MXNet compatible with Windows Visual Studio.
@@ -94,7 +140,6 @@ List of Contributors
   - Jack created the amalgamation script and Go bind for MXNet.
 * [Li Dong](https://github.com/donglixp)
 * [Piji Li](https://github.com/lipiji)
-* [Hu Shiwen](https://github.com/yajiedesign)
 * [Boyuan Deng](https://github.com/bryandeng)
 * [Junran He](https://github.com/junranhe)
   - Junran makes device kvstore allocation strategy smarter
@@ -122,14 +167,11 @@ List of Contributors
 * [Mathis](https://github.com/sveitser)
 * [sennendoko](https://github.com/sennendoko)
 * [srand99](https://github.com/srand99)
-* [Yizhi Liu](https://github.com/yzhliu)
 * [Taiyun](https://github.com/taiyun)
 * [Yanghao Li](https://github.com/lyttonhao)
-* [Nan Zhu](https://github.com/CodingCat)
 * [Ye Zhou](https://github.com/zhouye)
 * [Zhang Chen](https://github.com/zhangchen-qinyinghua)
 * [Xianliang Wang](https://github.com/wangxianliang)
-* [Junru Shao](https://github.com/yzgysjr)
 * [Xiao Liu](https://github.com/skylook)
 * [Lowik CHANUSSOT](https://github.com/Nzeuwik)
 * [Alexander Skidanov](https://github.com/SkidanovAlex)
@@ -141,13 +183,11 @@ List of Contributors
 * [Tao Wei](https://github.com/taoari)
 * [Max Kuhn](https://github.com/topepo)
 * [Yuqi Li](https://github.com/ziyeqinghan)
-* [Depeng Liang](https://github.com/Ldpe2G)
 * [Kiko Qiu](https://github.com/kikoqiu)
 * [Yang Bo](https://github.com/Atry)
 * [Jonas Amaro](https://github.com/jonasrla)
 * [Yan Li](https://github.com/Godricly)
 * [Yuance Li](https://github.com/liyuance)
-* [Sandeep Krishnamurthy](https://github.com/sandeep-krishnamurthy)
 * [Andre Moeller](https://github.com/andremoeller)
 * [Miguel Gonzalez-Fierro](https://github.com/miguelgfierro)
 * [Mingjie Xing](https://github.com/EricFisher)
@@ -162,14 +202,11 @@ List of Contributors
 * [Piyush Singh](https://github.com/Piyush3dB)
 * [Freddy Chua](https://github.com/freddycct)
 * [Jie Zhang](https://github.com/luoyetx)
-* [Leonard Lausen](https://github.com/leezu)
 * [Robert Stone](https://github.com/tlby)
 * [Pedro Larroy](https://github.com/larroy)
-* [Jun Wu](https://github.com/reminisce)
 * [Dom Divakaruni](https://github.com/domdivakaruni)
 * [David Salinas](https://github.com/geoalgo)
 * [Asmus Hetzel](https://github.com/asmushetzel)
-* [Roshani Nagmote](https://github.com/Roshrini)
 * [Chetan Khatri](https://github.com/chetkhatri/)
 * [James Liu](https://github.com/jamesliu/)
 * [Nir Ben-Zvi](https://github.com/nirbenz/)
@@ -192,19 +229,14 @@ List of Contributors
 * [David Braude](https://github.com/dabraude/)
 * [Nick Robinson](https://github.com/nickrobinson)
 * [Kan Wu](https://github.com/wkcn)
-* [Rahul Huilgol](https://github.com/rahul003)
-* [Anirudh Subramanian](https://github.com/anirudh2290/)
 * [Zheqin Wang](https://github.com/rasefon)
 * [Thom Lane](https://github.com/thomelane)
 * [Sina Afrooze](https://github.com/safrooze)
 * [Sergey Sokolov](https://github.com/Ishitori)
-* [Thomas Delteil](https://github.com/ThomasDelteil)
 * [Jesse Brizzi](https://github.com/jessebrizzi)
 * [Hang Zhang](http://hangzh.com)
 * [Kou Ding](https://github.com/chinakook)
 * [Istvan Fehervari](https://github.com/ifeherva)
-* [Aaron Markham](https://github.com/aaronmarkham)
-* [Sam Skalicky](https://github.com/samskalicky)
 * [Per Goncalves da Silva](https://github.com/perdasilva)
 * [Zhijingcheng Yu](https://github.com/jasonyu1996)
 * [Cheng-Che Lee](https://github.com/stu1130)
@@ -212,9 +244,7 @@ List of Contributors
 * [LuckyPigeon](https://github.com/LuckyPigeon)
 * [Anton Chernov](https://github.com/lebeg)
 * [Denisa Roberts](https://github.com/D-Roberts)
-* [Dick Carter](https://github.com/DickJC123)
 * [Rahul Padmanabhan](https://github.com/rahul3)
-* [Yuxi Hu](https://github.com/yuxihu)
 * [Harsh Patel](https://github.com/harshp8l)
 * [Xiao Wang](https://github.com/BeyonderXX)
 * [Piyush Ghai](https://github.com/piyushghai)
@@ -253,7 +283,6 @@ List of Contributors
 * [Oliver Kowalke](https://github.com/olk)
 * [Connor Goggins](https://github.com/connorgoggins)
 * [Wei Chu](https://github.com/waytrue17)
-* [Yang Shi](https://github.com/ys2843)
 * [Joe Evans](https://github.com/josephevans)
 
 Label Bot
@@ -263,7 +292,7 @@ Label Bot
   - To use me, comment:
     - @mxnet-label-bot add [specify comma separated labels here]
     - @mxnet-label-bot remove [specify comma separated labels here]
-    - @mxnet-label-bot update [specify comma separated labels here]  
+    - @mxnet-label-bot update [specify comma separated labels here]
       (i.e. @mxnet-label-bot update [Bug, Python])
 
   - Available label names which are supported: 
[Labels](https://github.com/apache/incubator-mxnet/labels)
diff --git a/README.md b/README.md
index 60bfe5a..f0a38b9 100644
--- a/README.md
+++ b/README.md
@@ -19,98 +19,120 @@
   <a href="https://mxnet.apache.org/";><img 
src="https://raw.githubusercontent.com/dmlc/web-data/master/mxnet/image/mxnet_logo_2.png";></a><br>
 </div>
 
-Apache MXNet (incubating) for Deep Learning
-=====
-| Master         | Docs          | License  |
-| :-------------:|:-------------:|:--------:|
-[![CentOS CPU Build 
Status](http://jenkins.mxnet-ci.amazon-ml.com/job/mxnet-validation/job/centos-cpu/job/master/badge/icon?subject=build%20centos%20cpu)](http://jenkins.mxnet-ci.amazon-ml.com/job/mxnet-validation/job/centos-cpu/job/master/)
 [![CentOS GPU Build 
Status](http://jenkins.mxnet-ci.amazon-ml.com/job/mxnet-validation/job/centos-gpu/job/master/badge/icon?subject=build%20centos%20gpu)](http://jenkins.mxnet-ci.amazon-ml.com/job/mxnet-validation/job/centos-gpu/job/master/)
 [![Clang [...]
+[![banner](https://raw.githubusercontent.com/dmlc/web-data/master/mxnet/image/banner.png)](https://mxnet.apache.org)
 
-![banner](https://raw.githubusercontent.com/dmlc/web-data/master/mxnet/image/banner.png)
+Apache MXNet (incubating) for Deep Learning
+===========================================
+[![GitHub release (latest 
SemVer)](https://img.shields.io/github/v/release/apache/incubator-mxnet)](https://github.com/apache/incubator-mxnet/releases)
 [![GitHub 
stars](https://img.shields.io/github/stars/apache/incubator-mxnet)](https://github.com/apache/incubator-mxnet/stargazers)
 [![GitHub 
forks](https://img.shields.io/github/forks/apache/incubator-mxnet)](https://github.com/apache/incubator-mxnet/network)
 [![GitHub 
contributors](https://img.shields.io/github/contributors-anon/apache/ [...]
 
-Apache MXNet (incubating) is a deep learning framework designed for both 
*efficiency* and *flexibility*.
+Apache MXNet is a deep learning framework designed for both *efficiency* and 
*flexibility*.
 It allows you to ***mix*** [symbolic and imperative 
programming](https://mxnet.apache.org/api/architecture/program_model)
 to ***maximize*** efficiency and productivity.
 At its core, MXNet contains a dynamic dependency scheduler that automatically 
parallelizes both symbolic and imperative operations on the fly.
 A graph optimization layer on top of that makes symbolic execution fast and 
memory efficient.
-MXNet is portable and lightweight, scaling effectively to multiple GPUs and 
multiple machines.
+MXNet is portable and lightweight, scalable to many GPUs and machines.
 
-MXNet is more than a deep learning project. It is a collection of
-[blue prints and 
guidelines](https://mxnet.apache.org/api/architecture/overview) for building
-deep learning systems, and interesting insights of DL systems for hackers.
+MXNet is more than a deep learning project. It is a 
[community](https://mxnet.apache.org/community)
+on a mission of democratizing AI. It is a collection of [blue prints and 
guidelines](https://mxnet.apache.org/api/architecture/overview)
+for building deep learning systems, and interesting insights of DL systems for 
hackers.
 
-Ask Questions
--------------
-* Please use [discuss.d2l.ai](https://discuss.d2l.ai/c/d2l-en/mxnet/) or [old 
version:discuss.mxnet.io](https://discuss.mxnet.io/) for asking questions. 
-* Please use [mxnet/issues](https://github.com/apache/incubator-mxnet/issues) 
for reporting bugs.
-* [Frequent Asked Questions](https://mxnet.apache.org/faq/faq.html)
+Licensed under an 
[Apache-2.0](https://github.com/apache/incubator-mxnet/blob/master/LICENSE) 
license.
 
-How to Contribute
------------------
-* [Contribute to MXNet](https://mxnet.apache.org/community/contribute.html)
+| Branch  | Build Status  |
+|:-------:|:-------------:|
+| [master](https://github.com/apache/incubator-mxnet/tree/master) | [![CentOS 
CPU Build 
Status](http://jenkins.mxnet-ci.amazon-ml.com/job/mxnet-validation/job/centos-cpu/job/master/badge/icon?subject=build%20centos%20cpu)](http://jenkins.mxnet-ci.amazon-ml.com/job/mxnet-validation/job/centos-cpu/job/master/)
 [![CentOS GPU Build 
Status](http://jenkins.mxnet-ci.amazon-ml.com/job/mxnet-validation/job/centos-gpu/job/master/badge/icon?subject=build%20centos%20gpu)](http://jenkins.mxnet-ci.ama
 [...]
+| [v1.x](https://github.com/apache/incubator-mxnet/tree/v1.x) | [![CentOS CPU 
Build 
Status](http://jenkins.mxnet-ci.amazon-ml.com/job/mxnet-validation/job/centos-cpu/job/v1.x/badge/icon?subject=build%20centos%20cpu)](http://jenkins.mxnet-ci.amazon-ml.com/job/mxnet-validation/job/centos-cpu/job/v1.x/)
 [![CentOS GPU Build 
Status](http://jenkins.mxnet-ci.amazon-ml.com/job/mxnet-validation/job/centos-gpu/job/v1.x/badge/icon?subject=build%20centos%20gpu)](http://jenkins.mxnet-ci.amazon-ml.com
 [...]
+
+Features
+--------
+* NumPy-like programming interface, and is integrated with the new, 
easy-to-use Gluon 2.0 interface. NumPy users can easily adopt MXNet and start 
in deep learning.
+* Automatic hybridization provides imperative programming with the performance 
of traditional symbolic programming.
+* Lightweight, memory-efficient, and portable to smart devices through native 
cross-compilation support on ARM, and through ecosystem projects such as 
[TVM](https://tvm.ai), 
[TensorRT](https://docs.nvidia.com/deeplearning/tensorrt/developer-guide/index.html),
 
[OpenVINO](https://software.intel.com/content/www/us/en/develop/tools/openvino-toolkit.html).
+* Scales up to multi GPUs and distributed setting with auto parallelism 
through [ps-lite](https://github.com/dmlc/ps-lite), 
[Horovod](https://github.com/horovod/horovod), and 
[BytePS](https://github.com/bytedance/byteps).
+* Extensible backend that supports full customization, allowing integration 
with custom accelerator libraries and in-house hardware without the need to 
maintain a fork.
+* Support for [Python](https://mxnet.apache.org/api/python), 
[Java](https://mxnet.apache.org/api/java), 
[C++](https://mxnet.apache.org/api/cpp), [R](https://mxnet.apache.org/api/r), 
[Scala](https://mxnet.apache.org/api/scala), 
[Clojure](https://mxnet.apache.org/api/clojure), 
[Go](https://github.com/jdeng/gomxnet/), 
[Javascript](https://github.com/dmlc/mxnet.js/), 
[Perl](https://mxnet.apache.org/api/perl), and 
[Julia](https://mxnet.apache.org/api/julia)
+* Cloud-friendly and directly compatible with AWS and Azure.
+
+Contents
+--------
+* [Installation](https://mxnet.apache.org/get_started)
+* [Tutorials](https://mxnet.apache.org/api/python/docs/tutorials/)
+* [Ecosystem](https://mxnet.apache.org/ecosystem)
+* [API Documentation](https://mxnet.apache.org/api)
+* [Examples](https://github.com/apache/incubator-mxnet-examples)
+* [Social Media](https://mxnet.apache.org/community#social-media)
+* [Contribute to MXNet](https://mxnet.apache.org/community/)
 
 What's New
 ----------
-* [Version 1.6.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.6.0) - MXNet 
1.6.0 Release.
-* [Version 1.5.1 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.5.1) - MXNet 
1.5.1 Patch Release.
-* [Version 1.5.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.5.0) - MXNet 
1.5.0 Release.
-* [Version 1.4.1 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.4.1) - MXNet 
1.4.1 Patch Release.
-* [Version 1.4.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.4.0) - MXNet 
1.4.0 Release.
-* [Version 1.3.1 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.3.1) - MXNet 
1.3.1 Patch Release.
-* [Version 1.3.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.3.0) - MXNet 
1.3.0 Release.
-* [Version 1.2.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.2.0) - MXNet 
1.2.0 Release.
-* [Version 1.1.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.1.0) - MXNet 
1.1.0 Release.
-* [Version 1.0.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.0.0) - MXNet 
1.0.0 Release.
-* [Version 0.12.1 
Release](https://github.com/apache/incubator-mxnet/releases/tag/0.12.1) - MXNet 
0.12.1 Patch Release.
-* [Version 0.12.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/0.12.0) - MXNet 
0.12.0 Release.
-* [Version 0.11.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/0.11.0) - MXNet 
0.11.0 Release.
+* [1.7.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.7.0) - MXNet 
1.7.0 Release.
+* [1.6.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.6.0) - MXNet 
1.6.0 Release.
+* [1.5.1 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.5.1) - MXNet 
1.5.1 Patch Release.
+* [1.5.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.5.0) - MXNet 
1.5.0 Release.
+* [1.4.1 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.4.1) - MXNet 
1.4.1 Patch Release.
+* [1.4.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.4.0) - MXNet 
1.4.0 Release.
+* [1.3.1 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.3.1) - MXNet 
1.3.1 Patch Release.
+* [1.3.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.3.0) - MXNet 
1.3.0 Release.
+* [1.2.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.2.0) - MXNet 
1.2.0 Release.
+* [1.1.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.1.0) - MXNet 
1.1.0 Release.
+* [1.0.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/1.0.0) - MXNet 
1.0.0 Release.
+* [0.12.1 
Release](https://github.com/apache/incubator-mxnet/releases/tag/0.12.1) - MXNet 
0.12.1 Patch Release.
+* [0.12.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/0.12.0) - MXNet 
0.12.0 Release.
+* [0.11.0 
Release](https://github.com/apache/incubator-mxnet/releases/tag/0.11.0) - MXNet 
0.11.0 Release.
 * [Apache Incubator](http://incubator.apache.org/projects/mxnet.html) - We are 
now an Apache Incubator project.
-* [Version 0.10.0 Release](https://github.com/dmlc/mxnet/releases/tag/v0.10.0) 
- MXNet 0.10.0 Release.
-* [Version 0.9.3 Release](./docs/architecture/release_note_0_9.md) - First 0.9 
official release.
-* [Version 0.9.1 Release (NNVM 
refactor)](./docs/architecture/release_note_0_9.md) - NNVM branch is merged 
into master now. An official release will be made soon.
-* [Version 0.8.0 Release](https://github.com/dmlc/mxnet/releases/tag/v0.8.0)
-* [Updated Image Classification with new Pre-trained 
Models](./example/image-classification)
-* [Notebooks How to Use MXNet](https://github.com/d2l-ai/d2l-en)
+* [0.10.0 Release](https://github.com/dmlc/mxnet/releases/tag/v0.10.0) - MXNet 
0.10.0 Release.
+* [0.9.3 Release](./docs/architecture/release_note_0_9.md) - First 0.9 
official release.
+* [0.9.1 Release (NNVM refactor)](./docs/architecture/release_note_0_9.md) - 
NNVM branch is merged into master now. An official release will be made soon.
+* [0.8.0 Release](https://github.com/dmlc/mxnet/releases/tag/v0.8.0)
+
+### Ecosystem News
+
 * [MKLDNN for Faster CPU 
Performance](docs/python_docs/python/tutorials/performance/backend/mkldnn/mkldnn_readme.md)
 * [MXNet Memory Monger, Training Deeper Nets with Sublinear Memory 
Cost](https://github.com/dmlc/mxnet-memonger)
 * [Tutorial for NVidia GTC 2016](https://github.com/dmlc/mxnet-gtc-tutorial)
 * [MXNet.js: Javascript Package for Deep Learning in Browser (without 
server)](https://github.com/dmlc/mxnet.js/)
 * [Guide to Creating New Operators 
(Layers)](https://mxnet.apache.org/api/faq/new_op)
 * [Go binding for inference](https://github.com/songtianyi/go-mxnet-predictor)
-* [Large Scale Image 
Classification](https://github.com/apache/incubator-mxnet/tree/master/example/image-classification)
 
-Contents
---------
-* [Book](https://d2l.ai)
-* [Website](https://mxnet.apache.org)
-* [Documentation](https://mxnet.apache.org/api)
-* [Blog](https://mxnet.apache.org/blog)
-* [Code 
Examples](https://github.com/apache/incubator-mxnet/tree/master/example)
-* [Installation](https://mxnet.apache.org/get_started)
-* [Features](https://mxnet.apache.org/features)
-* [Ecosystem](https://mxnet.apache.org/ecosystem)
+Stay Connected
+--------------
 
-Features
---------
-* Design notes providing useful insights that can re-used by other DL projects
-* Flexible configuration for arbitrary computation graph
-* Mix and match imperative and symbolic programming to maximize flexibility 
and efficiency
-* Lightweight, memory efficient and portable to smart devices
-* Scales up to multi GPUs and distributed setting with auto parallelism
-* Support for [Python](https://mxnet.apache.org/api/python), 
[Scala](https://mxnet.apache.org/api/scala), 
[C++](https://mxnet.apache.org/api/cpp), 
[Java](https://mxnet.apache.org/api/java), 
[Clojure](https://mxnet.apache.org/api/clojure), 
[R](https://mxnet.apache.org/api/r), [Go](https://github.com/jdeng/gomxnet/), 
[Javascript](https://github.com/dmlc/mxnet.js/), 
[Perl](https://mxnet.apache.org/api/perl), 
[Matlab](https://github.com/apache/incubator-mxnet/tree/master/matlab), and 
[Julia] [...]
-* Cloud-friendly and directly compatible with AWS S3, AWS Deep Learning AMI, 
AWS SageMaker, HDFS, and Azure
-
-License
--------
-Licensed under an 
[Apache-2.0](https://github.com/apache/incubator-mxnet/blob/master/LICENSE) 
license.
+| Channel | Purpose |
+|---|---|
+| [Follow MXNet Development on 
Github](https://github.com/apache/incubator-mxnet/issues) | See what's going on 
in the MXNet project. |
+| [MXNet Confluence Wiki for 
Developers](https://cwiki.apache.org/confluence/display/MXNET/Apache+MXNet+Home)
 <i class="fas fa-external-link-alt"> | MXNet developer wiki for information 
related to project development, maintained by contributors and developers. To 
request write access, send an email to [send request to the dev 
list](mailto:[email protected]?subject=Requesting%20CWiki%20write%20access) 
<i class="far fa-envelope"></i>. |
+| [[email protected] mailing 
list](https://lists.apache.org/[email protected]) | The "dev 
list". Discussions about the development of MXNet. To subscribe, send an email 
to [[email protected]](mailto:[email protected]) <i 
class="far fa-envelope"></i>. |
+| [discuss.mxnet.io](https://discuss.mxnet.io) <i class="fas 
fa-external-link-alt"></i> | Asking & answering MXNet usage questions. |
+| [Apache Slack #mxnet Channel](https://the-asf.slack.com/archives/C7FN4FCP9) 
<i class="fas fa-external-link-alt"> | Connect with MXNet and other Apache 
developers. To join the MXNet slack channel [send request to the dev 
list](mailto:[email protected]?subject=Requesting%20slack%20access) <i 
class="far fa-envelope"></i>. |
+| [Follow MXNet on Social Media](#social-media) | Get updates about new 
features and events. |
 
-Reference Paper
----------------
 
-Tianqi Chen, Mu Li, Yutian Li, Min Lin, Naiyan Wang, Minjie Wang, Tianjun Xiao,
-Bing Xu, Chiyuan Zhang, and Zheng Zhang.
-[MXNet: A Flexible and Efficient Machine Learning Library for Heterogeneous 
Distributed 
Systems](https://github.com/dmlc/web-data/raw/master/mxnet/paper/mxnet-learningsys.pdf).
-In Neural Information Processing Systems, Workshop on Machine Learning 
Systems, 2015
+### Social Media
+
+Keep connected with the latest MXNet news and updates.
+
+<p>
+<a href="https://twitter.com/apachemxnet";><img 
src="https://raw.githubusercontent.com/dmlc/web-data/master/mxnet/social/twitter.svg?sanitize=true";
 height="30px"/> Apache MXNet on Twitter</a>
+</p>
+<p>
+<a href="https://medium.com/apache-mxnet";><img 
src="https://raw.githubusercontent.com/dmlc/web-data/master/mxnet/social/medium_black.svg?sanitize=true";
 height="30px"/> Contributor and user blogs about MXNet</a>
+</p>
+<p>
+<a href="https://reddit.com/r/mxnet";><img 
src="https://raw.githubusercontent.com/dmlc/web-data/master/mxnet/social/reddit_blue.svg?sanitize=true";
 height="30px" alt="reddit"/> Discuss MXNet on r/mxnet</a>
+</p>
+<p>
+<a href="https://www.youtube.com/apachemxnet";><img 
src="https://raw.githubusercontent.com/dmlc/web-data/master/mxnet/social/youtube_red.svg?sanitize=true";
 height="30px"/> Apache MXNet YouTube channel</a>
+</p>
+<p>
+<a href="https://www.linkedin.com/company/apache-mxnet";><img 
src="https://raw.githubusercontent.com/dmlc/web-data/master/mxnet/social/linkedin.svg?sanitize=true";
 height="30px"/> Apache MXNet on LinkedIn</a>
+</p>
+
 
 History
 -------
 MXNet emerged from a collaboration by the authors of 
[cxxnet](https://github.com/dmlc/cxxnet), 
[minerva](https://github.com/dmlc/minerva), and 
[purine2](https://github.com/purine/purine2). The project reflects what we have 
learned from the past projects. MXNet combines aspects of each of these 
projects to achieve flexibility, speed, and memory efficiency.
+
+Tianqi Chen, Mu Li, Yutian Li, Min Lin, Naiyan Wang, Minjie Wang, Tianjun Xiao,
+Bing Xu, Chiyuan Zhang, and Zheng Zhang.
+[MXNet: A Flexible and Efficient Machine Learning Library for Heterogeneous 
Distributed 
Systems](https://github.com/dmlc/web-data/raw/master/mxnet/paper/mxnet-learningsys.pdf).
+In Neural Information Processing Systems, Workshop on Machine Learning 
Systems, 2015

Reply via email to