This is an automated email from the ASF dual-hosted git repository.
haibin pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-mxnet.git
The following commit(s) were added to refs/heads/master by this push:
new 50850af replaced how_to with faq (#9575)
50850af is described below
commit 50850af1ddc39cb9bbd3535ab626a53bbe9802b3
Author: thinksanky <[email protected]>
AuthorDate: Fri Jan 26 19:49:15 2018 -0800
replaced how_to with faq (#9575)
* replaced how_to with faq
* fixed broken links from 197 report
---
R-package/README.md | 2 +-
README.md | 12 ++++-----
docs/architecture/release_note_0_9.md | 2 +-
docs/community/index.md | 4 +--
docs/faq/env_var.md | 2 +-
docs/faq/faq.md | 4 +--
docs/faq/finetune.md | 2 +-
docs/faq/gradient_compression.md | 2 +-
docs/faq/multi_devices.md | 12 ++++-----
docs/faq/nnpack.md | 2 +-
docs/faq/perf.md | 4 +--
docs/faq/s3_integration.md | 2 +-
docs/faq/visualize_graph.md | 4 +--
docs/install/amazonlinux_setup.md | 4 +--
docs/install/build_from_source.md | 2 +-
docs/install/centos_setup.md | 4 +--
docs/install/osx_setup.md | 4 +--
docs/install/raspbian_setup.md | 4 +--
docs/install/tx2_setup.md | 4 +--
docs/install/ubuntu_setup.md | 4 +--
docs/install/windows_setup.md | 2 +-
docs/tutorials/basic/data.md | 30 +++++++++++-----------
docs/tutorials/basic/module.md | 8 +++---
docs/tutorials/basic/ndarray.md | 2 +-
docs/tutorials/basic/symbol.md | 8 +++---
docs/tutorials/embedded/wine_detector.md | 4 +--
docs/tutorials/gluon/mnist.md | 2 +-
docs/tutorials/python/linear-regression.md | 6 ++---
docs/tutorials/python/mnist.md | 4 +--
docs/tutorials/python/predict_image.md | 2 +-
docs/tutorials/r/ndarray.md | 2 +-
docs/tutorials/r/symbol.md | 2 +-
docs/tutorials/scala/char_lstm.md | 4 +--
docs/tutorials/scala/mnist.md | 2 +-
docs/tutorials/scala/mxnet_scala_on_intellij.md | 2 +-
docs/tutorials/sparse/csr.md | 2 +-
docs/tutorials/sparse/row_sparse.md | 2 +-
docs/tutorials/sparse/train.md | 6 ++---
.../tutorials/vision/large_scale_classification.md | 4 +--
example/caffe/README.md | 2 +-
example/image-classification/README.md | 6 ++---
example/recommenders/crossentropy.py | 2 +-
example/recommenders/randomproj.py | 2 +-
example/rnn/bucketing/README.md | 2 +-
example/rnn/old/README.md | 2 +-
example/sparse/linear_classification/README.md | 2 +-
example/ssd/tools/caffe_converter/README.md | 2 +-
.../AI-MXNet/lib/AI/MXNet/Gluon/Trainer.pm | 2 +-
plugin/caffe/README.md | 2 +-
python/mxnet/context.py | 2 +-
scala-package/README.md | 2 +-
setup-utils/install-mxnet-osx-python.sh | 2 +-
src/operator/custom/custom.cc | 2 +-
tools/caffe_converter/README.md | 2 +-
tools/caffe_translator/README.md | 4 +--
tools/caffe_translator/faq.md | 4 +--
56 files changed, 107 insertions(+), 109 deletions(-)
diff --git a/R-package/README.md b/R-package/README.md
index e21d6b1..78a6214 100644
--- a/R-package/README.md
+++ b/R-package/README.md
@@ -24,7 +24,7 @@ options(repos = cran)
install.packages("mxnet")
```
-To use the GPU version or to use it on Linux, please follow [Installation
Guide](http://mxnet.io/get_started/install.html)
+To use the GPU version or to use it on Linux, please follow [Installation
Guide](http://mxnet.io/install/index.html)
License
-------
diff --git a/README.md b/README.md
index 5dd5f02..feff029 100644
--- a/README.md
+++ b/README.md
@@ -36,13 +36,13 @@ What's New
* [MKLDNN for Faster CPU Performance](./MKL_README.md)
* [MXNet Memory Monger, Training Deeper Nets with Sublinear Memory
Cost](https://github.com/dmlc/mxnet-memonger)
* [Tutorial for NVidia GTC 2016](https://github.com/dmlc/mxnet-gtc-tutorial)
-* [Embedding Torch layers and functions in
MXNet](https://mxnet.incubator.apache.org/how_to/torch.html)
+* [Embedding Torch layers and functions in
MXNet](https://mxnet.incubator.apache.org/faq/torch.html)
* [MXNet.js: Javascript Package for Deep Learning in Browser (without server)
](https://github.com/dmlc/mxnet.js/)
* [Design Note: Design Efficient Deep Learning Data Loading
Module](https://mxnet.incubator.apache.org/architecture/note_data_loading.html)
-* [MXNet on Mobile
Device](https://mxnet.incubator.apache.org/how_to/smart_device.html)
-* [Distributed
Training](https://mxnet.incubator.apache.org/how_to/multi_devices.html)
-* [Guide to Creating New Operators
(Layers)](https://mxnet.incubator.apache.org/how_to/new_op.html)
+* [MXNet on Mobile
Device](https://mxnet.incubator.apache.org/faq/smart_device.html)
+* [Distributed
Training](https://mxnet.incubator.apache.org/faq/multi_devices.html)
+* [Guide to Creating New Operators
(Layers)](https://mxnet.incubator.apache.org/faq/new_op.html)
* [Go binding for inference](https://github.com/songtianyi/go-mxnet-predictor)
* [Amalgamation and Go Binding for
Predictors](https://github.com/jdeng/gomxnet/) - Outdated
* [Large Scale Image
Classification](https://github.com/apache/incubator-mxnet/tree/master/example/image-classification)
@@ -52,10 +52,10 @@ Contents
* [Documentation](https://mxnet.incubator.apache.org/) and
[Tutorials](https://mxnet.incubator.apache.org/tutorials/)
* [Design Notes](https://mxnet.incubator.apache.org/architecture/index.html)
* [Code Examples](https://github.com/dmlc/mxnet/tree/master/example)
-* [Installation](https://mxnet.incubator.apache.org/get_started/install.html)
+* [Installation](https://mxnet.incubator.apache.org/install/index.html)
* [Pretrained Models](https://github.com/dmlc/mxnet-model-gallery)
* [Contribute to
MXNet](https://mxnet.incubator.apache.org/community/contribute.html)
-* [Frequent Asked
Questions](https://mxnet.incubator.apache.org/how_to/faq.html)
+* [Frequent Asked Questions](https://mxnet.incubator.apache.org/faq/faq.html)
Features
--------
diff --git a/docs/architecture/release_note_0_9.md
b/docs/architecture/release_note_0_9.md
index 61bad50..afcc091 100644
--- a/docs/architecture/release_note_0_9.md
+++ b/docs/architecture/release_note_0_9.md
@@ -4,7 +4,7 @@ Version 0.9 brings a number of important features and changes,
including a back-
## NNVM Refactor
-NNVM is a library for neural network graph construction, optimization, and
operator registration. It serves as an intermediary layer between the front-end
(MXNet user API) and the back-end (computation on the device). After version
0.9, MXNet fully adopts the NNVM framework. Now it's easier to create
operators. You can also register "pass"es that process and optimizes the graph
when `bind` is called on the symbol. For more discussion on how to create
operators with NNVM, please refer to [...]
+NNVM is a library for neural network graph construction, optimization, and
operator registration. It serves as an intermediary layer between the front-end
(MXNet user API) and the back-end (computation on the device). After version
0.9, MXNet fully adopts the NNVM framework. Now it's easier to create
operators. You can also register "pass"es that process and optimizes the graph
when `bind` is called on the symbol. For more discussion on how to create
operators with NNVM, please refer to [...]
Other changes brought by NNVM include:
- Backward shape inference is now supported
diff --git a/docs/community/index.md b/docs/community/index.md
index 6d3f345..ab98856 100644
--- a/docs/community/index.md
+++ b/docs/community/index.md
@@ -8,9 +8,9 @@ We track bugs and new feature requests in the MXNet Github repo
in the issues fo
## Contributors
MXNet has been developed and is used by a group of active community members.
Contribute to improving it! For more information, see
[contributions](http://mxnet.io/community/contribute.html).
-Please join the contributor mailing list.
[subscribe]('mailto:[email protected]')
[archive](https://lists.apache.org/[email protected])
+Please join the contributor mailing list.
[subscribe](mailto://[email protected])
[archive](https://lists.apache.org/[email protected])
-To join the MXNet slack channel send request to the contributor mailing list.
[subscribe]('mailto:[email protected]')
[archive](https://the-asf.slackarchive.io/mxnet)
+To join the MXNet slack channel send request to the contributor mailing list.
[subscribe](mailto://[email protected])
[archive](https://the-asf.slackarchive.io/mxnet)
## Roadmap
diff --git a/docs/faq/env_var.md b/docs/faq/env_var.md
index 7a4f8d5..41b8bca 100644
--- a/docs/faq/env_var.md
+++ b/docs/faq/env_var.md
@@ -24,7 +24,7 @@ export MXNET_GPU_WORKER_NTHREADS=3
- The number of threads given to prioritized CPU jobs.
* MXNET_CPU_NNPACK_NTHREADS
- Values: Int ```(default=4)```
- - The number of threads used for NNPACK. NNPACK package aims to provide
high-performance implementations of some layers for multi-core CPUs. Checkout
[NNPACK](http://mxnet.io/how_to/nnpack.html) to know more about it.
+ - The number of threads used for NNPACK. NNPACK package aims to provide
high-performance implementations of some layers for multi-core CPUs. Checkout
[NNPACK](http://mxnet.io/faq/nnpack.html) to know more about it.
## Memory Options
diff --git a/docs/faq/faq.md b/docs/faq/faq.md
index 0569963..668587e 100644
--- a/docs/faq/faq.md
+++ b/docs/faq/faq.md
@@ -48,10 +48,10 @@ copied_model = mx.model.FeedForward(ctx=mx.gpu(),
symbol=new_symbol,
arg_params=old_arg_params,
aux_params=old_aux_params,
allow_extra_params=True);
```
-For information about copying model parameters from an existing
```old_arg_params```, see this
[notebook](https://github.com/dmlc/mxnet-notebooks/blob/master/python/how_to/predict.ipynb).
More notebooks please refer to
[dmlc/mxnet-notebooks](https://github.com/dmlc/mxnet-notebooks).
+For information about copying model parameters from an existing
```old_arg_params```, see this
[notebook](https://github.com/dmlc/mxnet-notebooks/blob/master/python/faq/predict.ipynb).
More notebooks please refer to
[dmlc/mxnet-notebooks](https://github.com/dmlc/mxnet-notebooks).
#### How to Extract the Feature Map of a Certain Layer
-See this
[notebook](https://github.com/dmlc/mxnet-notebooks/blob/master/python/how_to/predict.ipynb).
More notebooks please refer to
[dmlc/mxnet-notebooks](https://github.com/dmlc/mxnet-notebooks).
+See this
[notebook](https://github.com/dmlc/mxnet-notebooks/blob/master/python/faq/predict.ipynb).
More notebooks please refer to
[dmlc/mxnet-notebooks](https://github.com/dmlc/mxnet-notebooks).
#### What Is the Relationship Between MXNet and CXXNet, Minerva, and Purine2?
diff --git a/docs/faq/finetune.md b/docs/faq/finetune.md
index 533c3ca..2c6c7e3 100644
--- a/docs/faq/finetune.md
+++ b/docs/faq/finetune.md
@@ -15,7 +15,7 @@ with these pretrained weights when training on our new task.
This process is
commonly called _fine-tuning_. There are a number of variations of fine-tuning.
Sometimes, the initial neural network is used only as a _feature extractor_.
That means that we freeze every layer prior to the output layer and simply
learn
-a new output layer. In [another
document](https://github.com/dmlc/mxnet-notebooks/blob/master/python/how_to/predict.ipynb),
we explained how to
+a new output layer. In [another
document](https://github.com/dmlc/mxnet-notebooks/blob/master/python/faq/predict.ipynb),
we explained how to
do this kind of feature extraction. Another approach is to update all of
the network's weights for the new task, and that's the approach we demonstrate
in
this document.
diff --git a/docs/faq/gradient_compression.md b/docs/faq/gradient_compression.md
index 4cd58f0..e2dbd32 100644
--- a/docs/faq/gradient_compression.md
+++ b/docs/faq/gradient_compression.md
@@ -85,7 +85,7 @@ A reference `gluon` implementation with a gradient
compression option can be fou
mod = mx.mod.Module(..., compression_params={'type’:'2bit', 'threshold':0.5})
```
-A `module` example is provided with [this guide for setting up MXNet with
distributed
training](https://mxnet.incubator.apache.org/versions/master/how_to/multi_devices.html#distributed-training-with-multiple-machines).
It comes with the option of turning on gradient compression as an argument to
the [train_mnist.py
script](https://github.com/apache/incubator-mxnet/blob/master/example/image-classification/train_mnist.py).
+A `module` example is provided with [this guide for setting up MXNet with
distributed
training](https://mxnet.incubator.apache.org/versions/master/faq/multi_devices.html#distributed-training-with-multiple-machines).
It comes with the option of turning on gradient compression as an argument to
the [train_mnist.py
script](https://github.com/apache/incubator-mxnet/blob/master/example/image-classification/train_mnist.py).
### Configuration Details
diff --git a/docs/faq/multi_devices.md b/docs/faq/multi_devices.md
index 9bd582c..5d538bc 100644
--- a/docs/faq/multi_devices.md
+++ b/docs/faq/multi_devices.md
@@ -32,7 +32,7 @@ gradients are then summed over all GPUs before updating the
model.
> To use GPUs, we need to compile MXNet with GPU support. For
> example, set `USE_CUDA=1` in `config.mk` before `make`. (see
-> [MXNet installation guide](http://mxnet.io/get_started/install.html) for
more options).
+> [MXNet installation guide](http://mxnet.io/install/index.html) for more
options).
If a machine has one or more GPU cards installed,
then each card is labeled by a number starting from 0.
@@ -57,17 +57,17 @@ If the available GPUs are not all equally powerful,
we can partition the workload accordingly.
For example, if GPU 0 is 3 times faster than GPU 2,
then we might use the workload option `work_load_list=[3, 1]`,
-see [Module](../api/python/module.html#mxnet.module.Module)
+see [Module](http://mxnet.io/api/python/module/module.html#mxnet.module.Module)
for more details.
Training with multiple GPUs should yield the same results
-as training on a single GPU if all other hyper-parameters are the same.
+as training on a single GPU if all other hyper-parameters are the same.f
In practice, the results may exhibit small differences,
owing to the randomness of I/O (random order or other augmentations),
weight initialization with different seeds, and CUDNN.
We can control on which devices the gradient is aggregated
-and on which device the model is updated via
[`KVStore`](http://mxnet.io/api/python/kvstore.html),
+and on which device the model is updated via
[`KVStore`](http://mxnet.io/api/python/kvstore/kvstore.html),
the _MXNet_ module that supports data communication.
One can either use `mx.kvstore.create(type)` to get an instance
or use the program flag `--kv-store type`.
@@ -101,7 +101,7 @@ When using a large number of GPUs, e.g. >=4, we suggest
using `device` for bette
### How to Launch a Job
> To use distributed training, we need to compile with `USE_DIST_KVSTORE=1`
-> (see [MXNet installation guide](http://mxnet.io/get_started/install.html)
for more options).
+> (see [MXNet installation guide](http://mxnet.io/install/index.html) for more
options).
Launching a distributed job is a bit different from running on a single
machine. MXNet provides
@@ -210,4 +210,4 @@ export PS_VERBOSE=1; python ../../tools/launch.py ...
### More
- See more launch options by `python ../../tools/launch.py -h`
-- See more options of
[ps-lite](http://ps-lite.readthedocs.org/en/latest/how_to.html)
+- See more options of
[ps-lite](http://ps-lite.readthedocs.org/en/latest/faq.html)
diff --git a/docs/faq/nnpack.md b/docs/faq/nnpack.md
index b17c6ee..ed38cb0 100644
--- a/docs/faq/nnpack.md
+++ b/docs/faq/nnpack.md
@@ -69,7 +69,7 @@ $ cd ~
* Set lib path of NNPACK as the environment variable, e.g. `export
LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$YOUR_NNPACK_INSTALL_PATH/lib`
* Add the include file of NNPACK and its third-party to `ADD_CFLAGS` in
config.mk, e.g. `ADD_CFLAGS = -I$(YOUR_NNPACK_INSTALL_PATH)/include/
-I$(YOUR_NNPACK_INSTALL_PATH)/third-party/pthreadpool/include/`
* Set `USE_NNPACK = 1` in config.mk.
-* Build MXNet from source following the [install
guide](http://mxnet.io/get_started/install.html).
+* Build MXNet from source following the [install
guide](http://mxnet.io/install/index.html).
### NNPACK Performance
diff --git a/docs/faq/perf.md b/docs/faq/perf.md
index 8899ecc..5199598 100644
--- a/docs/faq/perf.md
+++ b/docs/faq/perf.md
@@ -191,7 +191,7 @@ where the batch size for Alexnet is increased by 8x.
If more than one GPU or machine are used, MXNet uses `kvstore` to communicate
data.
It's critical to use the proper type of `kvstore` to get the best performance.
-Refer to [multi_device.md](http://mxnet.io/how_to/multi_devices.html) for more
+Refer to [multi_device.md](http://mxnet.io/faq/multi_devices.html) for more
details.
Besides, we can use
[tools/bandwidth](https://github.com/dmlc/mxnet/tree/master/tools/bandwidth)
@@ -225,7 +225,7 @@ by summarizing at the operator level, instead of a
function, kernel, or instruct
In order to be able to use the profiler, you must compile _MXNet_ with the
`USE_PROFILER=1` flag in `config.mk`.
-The profiler can then be turned on with an [environment
variable](http://mxnet.io/how_to/env_var.html#control-the-profiler)
+The profiler can then be turned on with an [environment
variable](http://mxnet.io/faq/env_var.html#control-the-profiler)
for an entire program run, or programmatically for just part of a run.
See
[example/profiler](https://github.com/dmlc/mxnet/tree/master/example/profiler)
for complete examples of how to use the profiler in code, but briefly, the
Python code looks like:
diff --git a/docs/faq/s3_integration.md b/docs/faq/s3_integration.md
index 4e6e965..0243567 100644
--- a/docs/faq/s3_integration.md
+++ b/docs/faq/s3_integration.md
@@ -15,7 +15,7 @@ Following are detailed instructions on how to use data from
S3 for training.
## Step 1: Build MXNet with S3 integration enabled
-Follow instructions [here](http://mxnet.io/get_started/install.html) to
install MXNet from source with the following additional steps to enable S3
integration.
+Follow instructions [here](http://mxnet.io/install/index.html) to install
MXNet from source with the following additional steps to enable S3 integration.
1. Install `libcurl4-openssl-dev` and `libssl-dev` before building MXNet.
These packages are required to read/write from AWS S3.
2. Append `USE_S3=1` to `config.mk` before building MXNet.
diff --git a/docs/faq/visualize_graph.md b/docs/faq/visualize_graph.md
index 21ab36f..0601021 100644
--- a/docs/faq/visualize_graph.md
+++ b/docs/faq/visualize_graph.md
@@ -11,12 +11,12 @@ from which the result can be read.
## Prerequisites
You need the [Jupyter Notebook](http://jupyter.readthedocs.io/en/latest/)
and [Graphviz](http://www.graphviz.org/) libraries to visualize the network.
-Please make sure you have followed [installation
instructions](http://mxnet.io/get_started/install.html)
+Please make sure you have followed [installation
instructions](http://mxnet.io/install/index.html)
in setting up above dependencies along with setting up MXNet.
## Visualize the sample Neural Network
-```mx.viz.plot_network``` takes
[Symbol](http://mxnet.io/api/python/symbol.html), with your Network definition,
and optional node_attrs, parameters for the shape of the node in the graph, as
input and generates a computation graph.
+```mx.viz.plot_network``` takes
[Symbol](http://mxnet.io/api/python/symbol/symbol.html), with your Network
definition, and optional node_attrs, parameters for the shape of the node in
the graph, as input and generates a computation graph.
We will now try to visualize a sample Neural Network for linear matrix
factorization:
- Start Jupyter notebook server
diff --git a/docs/install/amazonlinux_setup.md
b/docs/install/amazonlinux_setup.md
index 054e030..42a4fcb 100644
--- a/docs/install/amazonlinux_setup.md
+++ b/docs/install/amazonlinux_setup.md
@@ -1,8 +1,8 @@
<!-- This page should be deleted after sometime (Allowing search engines
to update links) -->
-<meta http-equiv="refresh" content="3;
url=http://mxnet.io/get_started/install.html" />
+<meta http-equiv="refresh" content="3; url=http://mxnet.io/install/index.html"
/>
<!-- Just in case redirection does not work -->
<p>
- <a href="http://mxnet.io/get_started/install.html">
+ <a href="http://mxnet.io/install/index.html">
This content is moved to a new MXNet install page. Redirecting... </a>
</p>
diff --git a/docs/install/build_from_source.md
b/docs/install/build_from_source.md
index 4f7083a..5c558a9 100644
--- a/docs/install/build_from_source.md
+++ b/docs/install/build_from_source.md
@@ -1,6 +1,6 @@
# Build MXNet from Source
-**NOTE:** For MXNet with Python installation, please refer to the [new install
guide](http://mxnet.io/get_started/install.html).
+**NOTE:** For MXNet with Python installation, please refer to the [new install
guide](http://mxnet.io/install/index.html).
This document explains how to build MXNet from sources. Building MXNet from
sources is a 2 step process.
diff --git a/docs/install/centos_setup.md b/docs/install/centos_setup.md
index 054e030..42a4fcb 100644
--- a/docs/install/centos_setup.md
+++ b/docs/install/centos_setup.md
@@ -1,8 +1,8 @@
<!-- This page should be deleted after sometime (Allowing search engines
to update links) -->
-<meta http-equiv="refresh" content="3;
url=http://mxnet.io/get_started/install.html" />
+<meta http-equiv="refresh" content="3; url=http://mxnet.io/install/index.html"
/>
<!-- Just in case redirection does not work -->
<p>
- <a href="http://mxnet.io/get_started/install.html">
+ <a href="http://mxnet.io/install/index.html">
This content is moved to a new MXNet install page. Redirecting... </a>
</p>
diff --git a/docs/install/osx_setup.md b/docs/install/osx_setup.md
index a009123..8980de5 100644
--- a/docs/install/osx_setup.md
+++ b/docs/install/osx_setup.md
@@ -1,6 +1,6 @@
# Installing MXNet froum source on OS X (Mac)
-**NOTE:** For prebuild MXNet with Python installation, please refer to the
[new install guide](http://mxnet.io/get_started/install.html).
+**NOTE:** For prebuild MXNet with Python installation, please refer to the
[new install guide](http://mxnet.io/install/index.html).
Installing MXNet is a two-step process:
@@ -217,5 +217,5 @@ After you build the shared library, run the following
command from the MXNet sou
## Next Steps
* [Tutorials](http://mxnet.io/tutorials/index.html)
-* [How To](http://mxnet.io/how_to/index.html)
+* [How To](http://mxnet.io/faq/index.html)
* [Architecture](http://mxnet.io/architecture/index.html)
diff --git a/docs/install/raspbian_setup.md b/docs/install/raspbian_setup.md
index 054e030..42a4fcb 100644
--- a/docs/install/raspbian_setup.md
+++ b/docs/install/raspbian_setup.md
@@ -1,8 +1,8 @@
<!-- This page should be deleted after sometime (Allowing search engines
to update links) -->
-<meta http-equiv="refresh" content="3;
url=http://mxnet.io/get_started/install.html" />
+<meta http-equiv="refresh" content="3; url=http://mxnet.io/install/index.html"
/>
<!-- Just in case redirection does not work -->
<p>
- <a href="http://mxnet.io/get_started/install.html">
+ <a href="http://mxnet.io/install/index.html">
This content is moved to a new MXNet install page. Redirecting... </a>
</p>
diff --git a/docs/install/tx2_setup.md b/docs/install/tx2_setup.md
index 054e030..42a4fcb 100644
--- a/docs/install/tx2_setup.md
+++ b/docs/install/tx2_setup.md
@@ -1,8 +1,8 @@
<!-- This page should be deleted after sometime (Allowing search engines
to update links) -->
-<meta http-equiv="refresh" content="3;
url=http://mxnet.io/get_started/install.html" />
+<meta http-equiv="refresh" content="3; url=http://mxnet.io/install/index.html"
/>
<!-- Just in case redirection does not work -->
<p>
- <a href="http://mxnet.io/get_started/install.html">
+ <a href="http://mxnet.io/install/index.html">
This content is moved to a new MXNet install page. Redirecting... </a>
</p>
diff --git a/docs/install/ubuntu_setup.md b/docs/install/ubuntu_setup.md
index 15d06fc..d33c042 100644
--- a/docs/install/ubuntu_setup.md
+++ b/docs/install/ubuntu_setup.md
@@ -1,6 +1,6 @@
# Installing MXNet on Ubuntu
-**NOTE:** For MXNet with Python installation, please refer to the [new install
guide](http://mxnet.io/get_started/install.html).
+**NOTE:** For MXNet with Python installation, please refer to the [new install
guide](http://mxnet.io/install/index.html).
MXNet currently supports Python, R, Julia, Scala, and Perl. For users of R on
Ubuntu operating systems, MXNet provides a set of Git Bash scripts that
installs all of the required MXNet dependencies and the MXNet library.
@@ -262,5 +262,5 @@ Before you build MXNet for Perl from source code, you must
complete [building th
## Next Steps
* [Tutorials](http://mxnet.io/tutorials/index.html)
-* [How To](http://mxnet.io/how_to/index.html)
+* [How To](http://mxnet.io/faq/index.html)
* [Architecture](http://mxnet.io/architecture/index.html)
diff --git a/docs/install/windows_setup.md b/docs/install/windows_setup.md
index e5e92a7..598a12f 100755
--- a/docs/install/windows_setup.md
+++ b/docs/install/windows_setup.md
@@ -296,5 +296,5 @@ To install the MXNet Scala package into your local Maven
repository, run the fol
## Next Steps
* [Tutorials](http://mxnet.io/tutorials/index.html)
-* [How To](http://mxnet.io/how_to/index.html)
+* [How To](http://mxnet.io/faq/index.html)
* [Architecture](http://mxnet.io/architecture/index.html)
diff --git a/docs/tutorials/basic/data.md b/docs/tutorials/basic/data.md
index b60626a..66479d5 100644
--- a/docs/tutorials/basic/data.md
+++ b/docs/tutorials/basic/data.md
@@ -8,7 +8,7 @@ Here we discuss the API conventions and several provided
iterators.
To complete this tutorial, we need:
-- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html).
+- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html).
- [OpenCV Python library](http://opencv.org/opencv-3-2.html), [Python
Requests](http://docs.python-requests.org/en/master/),
[Matplotlib](https://matplotlib.org/) and [Jupyter
Notebook](http://jupyter.org/index.html).
@@ -31,10 +31,10 @@ Iterators provide an abstract interface for traversing
various types of iterable
In MXNet, data iterators return a batch of data as `DataBatch` on each call to
`next`.
A `DataBatch` often contains *n* training examples and their corresponding
labels. Here *n* is the `batch_size` of the iterator. At the end of the data
stream when there is no more data to read, the iterator raises
``StopIteration`` exception like Python `iter`.
-The structure of `DataBatch` is defined
[here](http://mxnet.io/api/python/io.html#mxnet.io.DataBatch).
+The structure of `DataBatch` is defined
[here](http://mxnet.io/api/python/io/io.html#mxnet.io.DataBatch).
Information such as name, shape, type and layout on each training example and
their corresponding label can be provided as `DataDesc` data descriptor objects
via the `provide_data` and `provide_label` properties in `DataBatch`.
-The structure of `DataDesc` is defined
[here](http://mxnet.io/api/python/io.html#mxnet.io.DataDesc).
+The structure of `DataDesc` is defined
[here](http://mxnet.io/api/python/io/io.html#mxnet.io.DataDesc).
All IO in MXNet is handled via `mx.io.DataIter` and its subclasses. In this
tutorial, we'll discuss a few commonly used iterators provided by MXNet.
@@ -56,7 +56,7 @@ warnings.filterwarnings("ignore", category=DeprecationWarning)
## Reading data in memory
When data is stored in memory, backed by either an `NDArray` or ``numpy``
`ndarray`,
-we can use the
[__`NDArrayIter`__](http://mxnet.io/api/python/io.html#mxnet.io.NDArrayIter) to
read data as below:
+we can use the
[__`NDArrayIter`__](http://mxnet.io/api/python/io/io.html#mxnet.io.NDArrayIter)
to read data as below:
```python
@@ -69,7 +69,7 @@ for batch in data_iter:
```
## Reading data from CSV files
-MXNet provides [`CSVIter`](http://mxnet.io/api/python/io.html#mxnet.io.CSVIter)
+MXNet provides
[`CSVIter`](http://mxnet.io/api/python/io/io.html#mxnet.io.CSVIter)
to read from CSV files and can be used as below:
```python
@@ -88,7 +88,7 @@ An iterator in _MXNet_ should
1. Implement `next()` in ``Python2`` or `__next()__` in ``Python3``,
returning a `DataBatch` or raising a `StopIteration` exception if at the
end of the data stream.
2. Implement the `reset()` method to restart reading from the beginning.
-3. Have a `provide_data` attribute, consisting of a list of `DataDesc` objects
that store the name, shape, type and layout information of the data (more info
[here](http://mxnet.io/api/python/io.html#mxnet.io.DataBatch)).
+3. Have a `provide_data` attribute, consisting of a list of `DataDesc` objects
that store the name, shape, type and layout information of the data (more info
[here](http://mxnet.io/api/python/io/io.html#mxnet.io.DataBatch)).
4. Have a `provide_label` attribute consisting of a list of `DataDesc` objects
that store the name, shape, type and layout information of the label.
When creating a new iterator, you can either start from scratch and define an
iterator or reuse one of the existing iterators.
@@ -209,8 +209,8 @@ Record IO is a file format used by MXNet for data IO.
It compactly packs the data for efficient read and writes from distributed
file system like Hadoop HDFS and AWS S3.
You can learn more about the design of `RecordIO`
[here](http://mxnet.io/architecture/note_data_loading.html).
-MXNet provides
[__`MXRecordIO`__](http://mxnet.io/api/python/io.html#mxnet.recordio.MXRecordIO)
-and
[__`MXIndexedRecordIO`__](http://mxnet.io/api/python/io.html#mxnet.recordio.MXIndexedRecordIO)
+MXNet provides
[__`MXRecordIO`__](http://mxnet.io/api/python/io/io.html#mxnet.recordio.MXRecordIO)
+and
[__`MXIndexedRecordIO`__](http://mxnet.io/api/python/io/io.html#mxnet.recordio.MXIndexedRecordIO)
for sequential access of data and random access of the data.
### MXRecordIO
@@ -273,7 +273,7 @@ The `mx.recordio` package provides a few utility functions
for such operations,
#### Packing/Unpacking Binary Data
-[__`pack`__](http://mxnet.io/api/python/io.html#mxnet.recordio.pack) and
[__`unpack`__](http://mxnet.io/api/python/io.html#mxnet.recordio.unpack) are
used for storing float (or 1d array of float) label and binary data. The data
is packed along with a header. The header structure is defined
[here](http://mxnet.io/api/python/io.html#mxnet.recordio.IRHeader).
+[__`pack`__](http://mxnet.io/api/python/io/io.html#mxnet.recordio.pack) and
[__`unpack`__](http://mxnet.io/api/python/io/io.html#mxnet.recordio.unpack) are
used for storing float (or 1d array of float) label and binary data. The data
is packed along with a header. The header structure is defined
[here](http://mxnet.io/api/python/io/io.html#mxnet.recordio.IRHeader).
```python
@@ -296,7 +296,7 @@ print(mx.recordio.unpack(s2))
#### Packing/Unpacking Image Data
-MXNet provides
[__`pack_img`__](http://mxnet.io/api/python/io.html#mxnet.recordio.pack_img)
and
[__`unpack_img`__](http://mxnet.io/api/python/io.html#mxnet.recordio.unpack_img)
to pack/unpack image data.
+MXNet provides
[__`pack_img`__](http://mxnet.io/api/python/io/io.html#mxnet.recordio.pack_img)
and
[__`unpack_img`__](http://mxnet.io/api/python/io/io.html#mxnet.recordio.unpack_img)
to pack/unpack image data.
Records packed by `pack_img` can be loaded by `mx.io.ImageRecordIter`.
@@ -321,9 +321,9 @@ An example of how to use the script for converting to
*RecordIO* format is shown
In this section, we will learn how to preprocess and load image data in MXNet.
There are 4 ways of loading image data in MXNet.
- 1. Using
[__mx.image.imdecode__](http://mxnet.io/api/python/io.html#mxnet.image.imdecode)
to load raw image files.
- 2. Using
[__`mx.img.ImageIter`__](http://mxnet.io/api/python/io.html#mxnet.image.ImageIter)
implemented in Python which is very flexible to customization. It can read
from .rec(`RecordIO`) files and raw image files.
- 3. Using
[__`mx.io.ImageRecordIter`__](http://mxnet.io/api/python/io.html#mxnet.io.ImageRecordIter)
implemented on the MXNet backend in C++. This is less flexible to
customization but provides various language bindings.
+ 1. Using
[__mx.image.imdecode__](http://mxnet.io/api/python/io/io.html#mxnet.image.imdecode)
to load raw image files.
+ 2. Using
[__`mx.img.ImageIter`__](http://mxnet.io/api/python/io/io.html#mxnet.image.ImageIter)
implemented in Python which is very flexible to customization. It can read
from .rec(`RecordIO`) files and raw image files.
+ 3. Using
[__`mx.io.ImageRecordIter`__](http://mxnet.io/api/python/io/io.html#mxnet.io.ImageRecordIter)
implemented on the MXNet backend in C++. This is less flexible to
customization but provides various language bindings.
4. Creating a Custom iterator inheriting `mx.io.DataIter`
@@ -407,7 +407,7 @@ os.system("python %s/tools/im2rec.py --num-thread=4
--pass-through=1 data/caltec
The record io files are now saved at here (./data)
#### Using ImageRecordIter
-[__`ImageRecordIter`__](http://mxnet.io/api/python/io.html#mxnet.io.ImageRecordIter)
can be used for loading image data saved in record io format. To use
ImageRecordIter, simply create an instance by loading your record file:
+[__`ImageRecordIter`__](http://mxnet.io/api/python/io/io.html#mxnet.io.ImageRecordIter)
can be used for loading image data saved in record io format. To use
ImageRecordIter, simply create an instance by loading your record file:
```python
@@ -428,7 +428,7 @@ plt.show()
```
#### Using ImageIter
-[__ImageIter__](http://mxnet.io/api/python/io.html#mxnet.io.ImageIter) is a
flexible interface that supports loading of images in both RecordIO and Raw
format.
+[__ImageIter__](http://mxnet.io/api/python/io/io.html#mxnet.io.ImageIter) is a
flexible interface that supports loading of images in both RecordIO and Raw
format.
```python
diff --git a/docs/tutorials/basic/module.md b/docs/tutorials/basic/module.md
index 6141f3e..2d44951 100644
--- a/docs/tutorials/basic/module.md
+++ b/docs/tutorials/basic/module.md
@@ -18,7 +18,7 @@ this tutorial.
To complete this tutorial, we need:
-- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html).
+- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html).
- [Jupyter Notebook](http://jupyter.org/index.html) and [Python
Requests](http://docs.python-requests.org/en/master/) packages.
```
@@ -141,7 +141,7 @@ for epoch in range(5):
Epoch 4, Training ('accuracy', 0.764375)
-To learn more about these APIs, visit [Module
API](http://mxnet.io/api/python/module.html).
+To learn more about these APIs, visit [Module
API](http://mxnet.io/api/python/module/module.html).
## High-level Interface
@@ -149,7 +149,7 @@ To learn more about these APIs, visit [Module
API](http://mxnet.io/api/python/mo
Module also provides high-level APIs for training, predicting and evaluating
for
user convenience. Instead of doing all the steps mentioned in the above
section,
-one can simply call [fit
API](http://mxnet.io/api/python/module.html#mxnet.module.BaseModule.fit)
+one can simply call [fit
API](http://mxnet.io/api/python/module/module.html#mxnet.module.BaseModule.fit)
and it internally executes the same steps.
To fit a module, call the `fit` function as follows:
@@ -232,7 +232,7 @@ assert score[0][1] > 0.77, "Achieved accuracy (%f) is less
than expected (0.77)"
Some of the other metrics which can be used are `top_k_acc`(top-k-accuracy),
`F1`, `RMSE`, `MSE`, `MAE`, `ce`(CrossEntropy). To learn more about the
metrics,
-visit [Evaluation metric](http://mxnet.io/api/python/metric.html).
+visit [Evaluation metric](http://mxnet.io/api/python/metric/metric.html).
One can vary number of epochs, learning_rate, optimizer parameters to change
the score
and tune these parameters to get best score.
diff --git a/docs/tutorials/basic/ndarray.md b/docs/tutorials/basic/ndarray.md
index bc5ce89..2c171f2 100644
--- a/docs/tutorials/basic/ndarray.md
+++ b/docs/tutorials/basic/ndarray.md
@@ -42,7 +42,7 @@ Each NDArray supports some important attributes that you'll
often want to query:
To complete this tutorial, we need:
-- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html)
+- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html)
- [Jupyter](http://jupyter.org/)
```
pip install jupyter
diff --git a/docs/tutorials/basic/symbol.md b/docs/tutorials/basic/symbol.md
index dc7daae..3a40e59 100644
--- a/docs/tutorials/basic/symbol.md
+++ b/docs/tutorials/basic/symbol.md
@@ -26,7 +26,7 @@ which values will be needed later on.
But with symbolic programming, we declare the required outputs in advance.
This means that we can recycle memory allocated in intermediate steps,
as by performing operations in place. Symbolic API also uses less memory for
the
-same network. Refer to [How To](http://mxnet.io/how_to/index.html) and
+same network. Refer to [How To](http://mxnet.io/faq/index.html) and
[Architecture](http://mxnet.io/architecture/index.html) section to know more.
In our design notes, we present [a more thorough discussion on the comparative
strengths
@@ -40,7 +40,7 @@ can produce multiple output symbols
and can maintain internal state symbols.
For a visual explanation of these concepts, see
-[Symbolic Configuration and Execution in
Pictures](http://mxnet.io/api/python/symbol_in_pictures.html).
+[Symbolic Configuration and Execution in
Pictures](http://mxnet.io/api/python/symbol_in_pictures/symbol_in_pictures.html).
To make things concrete, let's take a hands-on look at the Symbol API.
There are a few different ways to compose a `Symbol`.
@@ -49,7 +49,7 @@ There are a few different ways to compose a `Symbol`.
To complete this tutorial, we need:
-- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html)
+- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html)
- [Jupyter](http://jupyter.org/)
```
pip install jupyter
@@ -383,7 +383,7 @@ Most operators such as `mx.sym.Convolution` and
`mx.sym.Reshape` are implemented
in C++ for better performance. MXNet also allows users to write new operators
using any front-end language such as Python. It often makes the developing and
debugging much easier. To implement an operator in Python, refer to
-[How to create new operators](http://mxnet.io/how_to/new_op.html).
+[How to create new operators](http://mxnet.io/faq/new_op.html).
## Advanced Usages
diff --git a/docs/tutorials/embedded/wine_detector.md
b/docs/tutorials/embedded/wine_detector.md
index f2f7a4e..605b657 100644
--- a/docs/tutorials/embedded/wine_detector.md
+++ b/docs/tutorials/embedded/wine_detector.md
@@ -37,9 +37,7 @@ To complete this tutorial, you need:
## Building MXNet for The Pi
-The first step will be to get MXNet with the Python bindings running on your
Raspberry Pi 3. There is a tutorial for that provided on
[here](http://mxnet.io/get_started/raspbian_setup.html). In short you will have
to download the dependencies, and build the full MXNet library for the Pi with
the ARM specific compile flags. Be sure to build the library with open CV as we
will be using a model that requires it to process images. Then you will finally
the Python bindings. Once this is done [...]
-
-The first step is to get MXNet with the Python bindings running on your
Raspberry Pi 3. There is a tutorial for that provided
[here](http://mxnet.io/get_started/raspbian_setup.html). The linked tutorial
walks you through downloading the dependencies, and building the full MXNet
library for the Pi with the ARM specific compile flags. Be sure to build the
library with open CV as we will be using a model that requires it to process
images. Then you will register the Python bindings to MXNet [...]
+The first step is to get MXNet with the Python bindings running on your
Raspberry Pi 3. There is a tutorial for that provided
[here](http://mxnet.io/insstall/index.html). The linked tutorial walks you
through downloading the dependencies, and building the full MXNet library for
the Pi with the ARM specific compile flags. Be sure to build the library with
open CV as we will be using a model that requires it to process images. Then
you will register the Python bindings to MXNet. After this [...]
```bash
diff --git a/docs/tutorials/gluon/mnist.md b/docs/tutorials/gluon/mnist.md
index ce23f1f..0bd616c 100644
--- a/docs/tutorials/gluon/mnist.md
+++ b/docs/tutorials/gluon/mnist.md
@@ -16,7 +16,7 @@ This is based on the Mnist tutorial with symbolic approach.
You can find it [her
## Prerequisites
To complete this tutorial, we need:
-- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html).
+- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html).
- [Python Requests](http://docs.python-requests.org/en/master/) and [Jupyter
Notebook](http://jupyter.org/index.html).
diff --git a/docs/tutorials/python/linear-regression.md
b/docs/tutorials/python/linear-regression.md
index fc3e713..9dfcf07 100644
--- a/docs/tutorials/python/linear-regression.md
+++ b/docs/tutorials/python/linear-regression.md
@@ -8,7 +8,7 @@ The function we are trying to learn is: *y = x<sub>1</sub> +
2x<sub>2</sub>*,
To complete this tutorial, we need:
-- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html).
+- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html).
- [Jupyter Notebook](http://jupyter.org/index.html).
@@ -56,7 +56,7 @@ eval_iter = mx.io.NDArrayIter(eval_data, eval_label,
batch_size, shuffle=False)
In the above example, we have made use of `NDArrayIter`, which is useful for
iterating
over both numpy ndarrays and MXNet NDArrays. In general, there are different
types of iterators in
MXNet and you can use one based on the type of data you are processing.
-Documentation for iterators can be found
[here](http://mxnet.io/api/python/io.html).
+Documentation for iterators can be found
[here](http://mxnet.io/api/python/io/io.html).
## MXNet Classes
@@ -94,7 +94,7 @@ and make up various components of the model. Symbols are used
to define:
The ones described above and other symbols are chained together with the
output of
one symbol serving as input to the next to build the network topology. More
information
-about the different types of symbols can be found
[here](http://mxnet.io/api/python/symbol.html).
+about the different types of symbols can be found
[here](http://mxnet.io/api/python/symbol/symbol.html).
```python
X = mx.sym.Variable('data')
diff --git a/docs/tutorials/python/mnist.md b/docs/tutorials/python/mnist.md
index 8e33409..067ded9 100644
--- a/docs/tutorials/python/mnist.md
+++ b/docs/tutorials/python/mnist.md
@@ -11,7 +11,7 @@ MNIST is a widely used dataset for the hand-written digit
classification task. I
## Prerequisites
To complete this tutorial, we need:
-- MXNet version 0.10 or later. See the installation instructions for your
operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html).
+- MXNet version 0.10 or later. See the installation instructions for your
operating system in [Setup and
Installation](http://mxnet.io/install/index.html).
- [Python Requests](http://docs.python-requests.org/en/master/) and [Jupyter
Notebook](http://jupyter.org/index.html).
@@ -57,7 +57,7 @@ data = mx.sym.flatten(data=data)
```
One might wonder if we are discarding valuable information by flattening. That
is indeed true and we'll cover this more when we talk about convolutional
neural networks where we preserve the input shape. For now, we'll go ahead and
work with flattened images.
-MLPs contains several fully connected layers. A fully connected layer or FC
layer for short, is one where each neuron in the layer is connected to every
neuron in its preceding layer. From a linear algebra perspective, an FC layer
applies an [affine
transform](https://en.wikipedia.org/wiki/Affine_transformation) to the *n x m*
input matrix *X* and outputs a matrix *Y* of size *n x k*, where *k* is the
number of neurons in the FC layer. *k* is also referred to as the hidden size.
The outp [...]
+MLPs contains several fully connected layers. A fully connected layer or FC
layer for short, is one where each neuron in the layer is connected to every
neuron in its preceding layer. From a linear algebra perspective, an FC layer
applies an [affine
transform](https://en.wikipedia.org/wiki/Affine_transformation) to the *n x m*
input matrix *X* and outputs a matrix *Y* of size *n x k*, where *k* is the
number of neurons in the FC layer. *k* is also referred to as the hidden size.
The outp [...]
In an MLP, the outputs of most FC layers are fed into an activation function,
which applies an element-wise non-linearity. This step is critical and it gives
neural networks the ability to classify inputs that are not linearly separable.
Common choices for activation functions are sigmoid, tanh, and [rectified
linear unit](https://en.wikipedia.org/wiki/Rectifier_%28neural_networks%29)
(ReLU). In this example, we'll use the ReLU activation function which has
several desirable properties a [...]
diff --git a/docs/tutorials/python/predict_image.md
b/docs/tutorials/python/predict_image.md
index 9a62e67..afd2bd7 100644
--- a/docs/tutorials/python/predict_image.md
+++ b/docs/tutorials/python/predict_image.md
@@ -7,7 +7,7 @@ pre-trained model, and how to perform feature extraction.
To complete this tutorial, we need:
-- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html)
+- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html)
- [Python Requests](http://docs.python-requests.org/en/master/),
[Matplotlib](https://matplotlib.org/) and [Jupyter
Notebook](http://jupyter.org/index.html).
diff --git a/docs/tutorials/r/ndarray.md b/docs/tutorials/r/ndarray.md
index e00f947..cb7639a 100644
--- a/docs/tutorials/r/ndarray.md
+++ b/docs/tutorials/r/ndarray.md
@@ -199,7 +199,7 @@ the results.
## Next Steps
* [Symbol](http://mxnet.io/tutorials/r/symbol.html)
-* [Write and use callback
functions](http://mxnet.io/tutorials/r/CallbackFunctionTutorial.html)
+* [Write and use callback
functions](http://mxnet.io/tutorials/r/CallbackFunction.html)
* [Neural Networks with MXNet in Five
Minutes](http://mxnet.io/tutorials/r/fiveMinutesNeuralNetwork.html)
* [Classify Real-World Images with Pre-trained
Model](http://mxnet.io/tutorials/r/classifyRealImageWithPretrainedModel.html)
* [Handwritten Digits Classification
Competition](http://mxnet.io/tutorials/r/mnistCompetition.html)
diff --git a/docs/tutorials/r/symbol.md b/docs/tutorials/r/symbol.md
index 6ab4dc2..4a87643 100644
--- a/docs/tutorials/r/symbol.md
+++ b/docs/tutorials/r/symbol.md
@@ -123,7 +123,7 @@ be more memory efficient than CXXNet and gets to the same
runtime with
greater flexibility.
## Next Steps
-* [Write and use callback
functions](http://mxnet.io/tutorials/r/CallbackFunctionTutorial.html)
+* [Write and use callback
functions](http://mxnet.io/tutorials/r/CallbackFunction.html)
* [Neural Networks with MXNet in Five
Minutes](http://mxnet.io/tutorials/r/fiveMinutesNeuralNetwork.html)
* [Classify Real-World Images with Pre-trained
Model](http://mxnet.io/tutorials/r/classifyRealImageWithPretrainedModel.html)
* [Handwritten Digits Classification
Competition](http://mxnet.io/tutorials/r/mnistCompetition.html)
diff --git a/docs/tutorials/scala/char_lstm.md
b/docs/tutorials/scala/char_lstm.md
index 466d827..5ec303e 100644
--- a/docs/tutorials/scala/char_lstm.md
+++ b/docs/tutorials/scala/char_lstm.md
@@ -6,7 +6,7 @@ There are many documents that explain LSTM concepts. If you
aren't familiar with
- Christopher Olah's [Understanding LSTM blog
post](http://colah.github.io/posts/2015-08-Understanding-LSTMs/)
- [Training a LSTM char-rnn in Julia to Generate Random
Sentences](http://dmlc.ml/mxnet/2015/11/15/char-lstm-in-julia.html)
- [Bucketing in MXNet in
Python](https://github.com/dmlc/mxnet-notebooks/blob/master/python/tutorials/char_lstm.ipynb)
-- [Bucketing in MXNet](http://mxnet.io/how_to/bucketing.html)
+- [Bucketing in MXNet](http://mxnet.io/faq/bucketing.html)
## How to Use This Tutorial
@@ -56,7 +56,7 @@ In this tutorial, you will accomplish the following:
To complete this tutorial, you need:
-- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html)
+- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html)
- [Scala 2.11.8](https://www.scala-lang.org/download/2.11.8.html)
- [Maven 3](https://maven.apache.org/install.html)
diff --git a/docs/tutorials/scala/mnist.md b/docs/tutorials/scala/mnist.md
index ad55ee4..6df9175 100644
--- a/docs/tutorials/scala/mnist.md
+++ b/docs/tutorials/scala/mnist.md
@@ -7,7 +7,7 @@ Let's train a 3-layer network (i.e multilayer perceptron
network) on the MNIST d
## Prerequisites
To complete this tutorial, we need:
-- to compile the latest MXNet version. See the MXNet installation instructions
for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html).
+- to compile the latest MXNet version. See the MXNet installation instructions
for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html).
- to compile the Scala API. See Scala API build instructions in
[Build](https://github.com/dmlc/mxnet/tree/master/scala-package).
## Define the Network
diff --git a/docs/tutorials/scala/mxnet_scala_on_intellij.md
b/docs/tutorials/scala/mxnet_scala_on_intellij.md
index eb667e9..dd2ac63 100644
--- a/docs/tutorials/scala/mxnet_scala_on_intellij.md
+++ b/docs/tutorials/scala/mxnet_scala_on_intellij.md
@@ -7,7 +7,7 @@ To use this tutorial, you need:
- [Maven 3](https://maven.apache.org/install.html).
- [Scala 2.11.8](https://www.scala-lang.org/download/2.11.8.html).
-- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html).
+- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html).
- The MXNet package for Scala. For installation instructions, see [this
procedure](http://mxnet.io/get_started/osx_setup.html#install-the-mxnet-package-for-scala).
- [IntelliJ IDE](https://www.jetbrains.com/idea/).
diff --git a/docs/tutorials/sparse/csr.md b/docs/tutorials/sparse/csr.md
index f4d7b7d..bbe71ff 100644
--- a/docs/tutorials/sparse/csr.md
+++ b/docs/tutorials/sparse/csr.md
@@ -21,7 +21,7 @@ The introduction of `CSRNDArray` also brings a new attribute,
`stype` as a holde
To complete this tutorial, you will need:
-- MXNet. See the instructions for your operating system in [Setup and
Installation](https://mxnet.io/get_started/install.html)
+- MXNet. See the instructions for your operating system in [Setup and
Installation](https://mxnet.io/install/index.html)
- [Jupyter](http://jupyter.org/)
```
pip install jupyter
diff --git a/docs/tutorials/sparse/row_sparse.md
b/docs/tutorials/sparse/row_sparse.md
index 70ca6b8..d4f6884 100644
--- a/docs/tutorials/sparse/row_sparse.md
+++ b/docs/tutorials/sparse/row_sparse.md
@@ -80,7 +80,7 @@ In this tutorial, we will describe what the row sparse format
is and how to use
To complete this tutorial, we need:
-- MXNet. See the instructions for your operating system in [Setup and
Installation](https://mxnet.io/get_started/install.html)
+- MXNet. See the instructions for your operating system in [Setup and
Installation](https://mxnet.io/install/index.html)
- [Jupyter](http://jupyter.org/)
```
pip install jupyter
diff --git a/docs/tutorials/sparse/train.md b/docs/tutorials/sparse/train.md
index 6f4e808..e31f046 100644
--- a/docs/tutorials/sparse/train.md
+++ b/docs/tutorials/sparse/train.md
@@ -10,7 +10,7 @@ then train a linear regression model using sparse symbols
with the Module API.
To complete this tutorial, we need:
-- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html).
+- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html).
- [Jupyter Notebook](http://jupyter.org/index.html) and [Python
Requests](http://docs.python-requests.org/en/master/) packages.
```
@@ -214,8 +214,8 @@ The function you will explore is: *y = x<sub>1</sub> +
2x<sub>2</sub> + ... 10
### Preparing the Data
-In MXNet, both
[mx.io.LibSVMIter](https://mxnet.incubator.apache.org/versions/master/api/python/io.html#mxnet.io.LibSVMIter)
-and
[mx.io.NDArrayIter](https://mxnet.incubator.apache.org/versions/master/api/python/io.html#mxnet.io.NDArrayIter)
+In MXNet, both
[mx.io.LibSVMIter](https://mxnet.incubator.apache.org/versions/master/api/python/io/io.html#mxnet.io.LibSVMIter)
+and
[mx.io.NDArrayIter](https://mxnet.incubator.apache.org/versions/master/api/python/io/io.html#mxnet.io.NDArrayIter)
support loading sparse data in CSR format. In this example, we'll use the
`NDArrayIter`.
You may see some warnings from SciPy. You don't need to worry about those for
this example.
diff --git a/docs/tutorials/vision/large_scale_classification.md
b/docs/tutorials/vision/large_scale_classification.md
index 1cf2270..17701e6 100644
--- a/docs/tutorials/vision/large_scale_classification.md
+++ b/docs/tutorials/vision/large_scale_classification.md
@@ -3,7 +3,7 @@
Training a neural network with a large number of images presents several
challenges. Even with the latest GPUs, it is not possible to train large
networks using a large number of images in a reasonable amount of time using a
single GPU. This problem can be somewhat mitigated by using multiple GPUs in a
single machine. But there is a limit to the number of GPUs that can be attached
to one machine (typically 8 or 16). This tutorial explains how to train large
networks with terabytes of dat [...]
## Prerequisites
-- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/get_started/install.html).
+- MXNet. See the instructions for your operating system in [Setup and
Installation](http://mxnet.io/install/index.html).
- [OpenCV Python library](http://opencv.org/opencv-3-2.html)
@@ -247,7 +247,7 @@ It is often straightforward to achieve a reasonable
validation accuracy, but ach
- Increase --data-nthreads (default is 4) to use more threads for data
preprocessing.
- Data preprocessing is done by opencv. If opencv is compiled from source
code, check if it is configured correctly.
- Use `--benchmark 1` to use randomly generated data rather than real data to
narrow down where the bottleneck is.
-- Check [this](http://mxnet.io/how_to/perf.html) page for more details.
+- Check [this](http://mxnet.io/faq/perf.html) page for more details.
### Memory
If the batch size is too big, it can exhaust GPU memory. If this happens,
you’ll see the error message “cudaMalloc failed: out of memory” or something
similar. There are a couple of ways to fix this:
diff --git a/example/caffe/README.md b/example/caffe/README.md
index 2a28e01..466305c 100644
--- a/example/caffe/README.md
+++ b/example/caffe/README.md
@@ -2,7 +2,7 @@
[Caffe](http://caffe.berkeleyvision.org/) has been a well-known and
widely-used deep learning framework. Now MXNet has supported calling most caffe
operators(layers) and loss functions directly in its symbolic graph! Using
one's own customized caffe layer is also effortless.
-Besides Caffe, MXNet has already embedded Torch modules and its tensor
mathematical functions.
([link](https://github.com/dmlc/mxnet/blob/master/docs/how_to/torch.md))
+Besides Caffe, MXNet has already embedded Torch modules and its tensor
mathematical functions.
([link](https://github.com/dmlc/mxnet/blob/master/docs/faq/torch.md))
This blog demonstrates two steps to use Caffe op in MXNet:
diff --git a/example/image-classification/README.md
b/example/image-classification/README.md
index 8a64b55..2967605 100644
--- a/example/image-classification/README.md
+++ b/example/image-classification/README.md
@@ -205,7 +205,7 @@ python fine-tune.py --pretrained-model
imagenet11k-resnet-152 --gpus 0,1,2,3,4,5
We obtained 87.3% top-1 validation accuracy, and the training log is available
[here](https://gist.github.com/mli/900b810258e2e0bc26fa606977a3b043#file-finetune-caltech265).
See
-the [python notebook](http://mxnet.io/how_to/finetune.html) for more
+the [python notebook](http://mxnet.io/faq/finetune.html) for more
explanations.
## Distributed Training
@@ -242,7 +242,7 @@ For more usages:
- One can use
[benchmark.py](https://github.com/dmlc/mxnet/blob/master/example/image-classification/benchmark.py)
to run distributed benchmarks (also for multiple GPUs with single machine)
-- A how-to [tutorial](http://mxnet.io/how_to/multi_devices.html) with more
+- A how-to [tutorial](http://mxnet.io/faq/multi_devices.html) with more
explanation.
- A
[blog](https://aws.amazon.com/blogs/compute/distributed-deep-learning-made-easy/)
@@ -357,7 +357,7 @@ aspects:
codes, check if it is configured correctly.
- Use `--benchmark 1` to use randomly generated data rather than real data.
-Refer to [how_to/performance](http://mxnet.io/how_to/perf.html) for more
details
+Refer to [faq/performance](http://mxnet.io/faq/perf.html) for more details
about CPU, GPU and multi-device performance.
### Memory
diff --git a/example/recommenders/crossentropy.py
b/example/recommenders/crossentropy.py
index d8577ed..ff44808 100644
--- a/example/recommenders/crossentropy.py
+++ b/example/recommenders/crossentropy.py
@@ -25,7 +25,7 @@ import time
import numpy as np
import mxnet as mx
-# ref: http://mxnet.io/how_to/new_op.html
+# ref: http://mxnet.io/faq/new_op.html
class CrossEntropyLoss(mx.operator.CustomOp):
"""An output layer that calculates gradient for cross-entropy loss
diff --git a/example/recommenders/randomproj.py
b/example/recommenders/randomproj.py
index ba080a0..83ce3a1 100644
--- a/example/recommenders/randomproj.py
+++ b/example/recommenders/randomproj.py
@@ -23,7 +23,7 @@ import numpy as np
import mxnet as mx
-# ref: http://mxnet.io/how_to/new_op.html
+# ref: http://mxnet.io/faq/new_op.html
class RandomBagOfWordsProjection(mx.operator.CustomOp):
"""Random projection layer for sparse bag-of-words (n-hot) inputs.
diff --git a/example/rnn/bucketing/README.md b/example/rnn/bucketing/README.md
index 0481609..b46642b 100644
--- a/example/rnn/bucketing/README.md
+++ b/example/rnn/bucketing/README.md
@@ -32,5 +32,5 @@ This folder contains RNN examples using high level mxnet.rnn
interface.
### Performance Note:
-More ```MXNET_GPU_WORKER_NTHREADS``` may lead to better performance. For
setting ```MXNET_GPU_WORKER_NTHREADS```, please refer to [Environment
Variables](http://mxnet.incubator.apache.org/how_to/env_var.html).
+More ```MXNET_GPU_WORKER_NTHREADS``` may lead to better performance. For
setting ```MXNET_GPU_WORKER_NTHREADS```, please refer to [Environment
Variables](http://mxnet.incubator.apache.org/faq/env_var.html).
diff --git a/example/rnn/old/README.md b/example/rnn/old/README.md
index 7540481..c03b36a 100644
--- a/example/rnn/old/README.md
+++ b/example/rnn/old/README.md
@@ -15,4 +15,4 @@ Run `get_ptb_data.sh` to download PenTreeBank data.
Performance Note:
-More ```MXNET_GPU_WORKER_NTHREADS``` may lead to better performance. For
setting ```MXNET_GPU_WORKER_NTHREADS```, please refer to [Environment
Variables](https://mxnet.readthedocs.org/en/latest/how_to/env_var.html).
+More ```MXNET_GPU_WORKER_NTHREADS``` may lead to better performance. For
setting ```MXNET_GPU_WORKER_NTHREADS```, please refer to [Environment
Variables](https://mxnet.readthedocs.org/en/latest/faq/env_var.html).
diff --git a/example/sparse/linear_classification/README.md
b/example/sparse/linear_classification/README.md
index 7e2a7ad..926d923 100644
--- a/example/sparse/linear_classification/README.md
+++ b/example/sparse/linear_classification/README.md
@@ -2,7 +2,7 @@ Linear Classification Using Sparse Matrix Multiplication
===========
This examples trains a linear model using the sparse feature in MXNet. This is
for demonstration purpose only.
-The example utilizes the sparse data loader
([mx.io.LibSVMIter](https://mxnet.incubator.apache.org/versions/master/api/python/io.html#mxnet.io.LibSVMIter)),
+The example utilizes the sparse data loader
([mx.io.LibSVMIter](https://mxnet.incubator.apache.org/versions/master/api/python/io/io.html#mxnet.io.LibSVMIter)),
the sparse dot operator and [sparse gradient
updaters](https://mxnet.incubator.apache.org/versions/master/api/python/ndarray/sparse.html#updater)
to train a linear model on the
[Avazu](https://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/binary.html#avazu)
click-through-prediction dataset.
diff --git a/example/ssd/tools/caffe_converter/README.md
b/example/ssd/tools/caffe_converter/README.md
index 5d40024..2e74fc5 100644
--- a/example/ssd/tools/caffe_converter/README.md
+++ b/example/ssd/tools/caffe_converter/README.md
@@ -10,7 +10,7 @@ python convert_caffe_modelzoo.py resnet-50
```
Please refer to
-[docs/how_to/caffe.md](../../docs/how_to/caffe.md) for more details.
+[docs/faq/caffe.md](../../docs/faq/caffe.md) for more details.
### How to use
To convert ssd caffemodels, Use: `python convert_model.py prototxt caffemodel
outputprefix`
diff --git a/perl-package/AI-MXNet/lib/AI/MXNet/Gluon/Trainer.pm
b/perl-package/AI-MXNet/lib/AI/MXNet/Gluon/Trainer.pm
index 63f521c..c2e8f31 100644
--- a/perl-package/AI-MXNet/lib/AI/MXNet/Gluon/Trainer.pm
+++ b/perl-package/AI-MXNet/lib/AI/MXNet/Gluon/Trainer.pm
@@ -39,7 +39,7 @@ use Mouse;
The set of parameters to optimize.
optimizer : str or Optimizer
The optimizer to use. See
- `help
<http://mxnet.io/api/python/optimization.html#the-mxnet-optimizer-package>`_
+ `help
<http://mxnet.io/api/python/optimization/optimization.html#the-mxnet-optimizer-package>`_
on Optimizer for a list of available optimizers.
optimizer_params : dict
Key-word arguments to be passed to optimizer constructor. For example,
diff --git a/plugin/caffe/README.md b/plugin/caffe/README.md
index 2a28e01..466305c 100644
--- a/plugin/caffe/README.md
+++ b/plugin/caffe/README.md
@@ -2,7 +2,7 @@
[Caffe](http://caffe.berkeleyvision.org/) has been a well-known and
widely-used deep learning framework. Now MXNet has supported calling most caffe
operators(layers) and loss functions directly in its symbolic graph! Using
one's own customized caffe layer is also effortless.
-Besides Caffe, MXNet has already embedded Torch modules and its tensor
mathematical functions.
([link](https://github.com/dmlc/mxnet/blob/master/docs/how_to/torch.md))
+Besides Caffe, MXNet has already embedded Torch modules and its tensor
mathematical functions.
([link](https://github.com/dmlc/mxnet/blob/master/docs/faq/torch.md))
This blog demonstrates two steps to use Caffe op in MXNet:
diff --git a/python/mxnet/context.py b/python/mxnet/context.py
index beccaeb..eb47614 100644
--- a/python/mxnet/context.py
+++ b/python/mxnet/context.py
@@ -29,7 +29,7 @@ class Context(object):
See also
----------
- `How to run MXNet on multiple CPU/GPUs
<http://mxnet.io/how_to/multi_devices.html>`
+ `How to run MXNet on multiple CPU/GPUs
<http://mxnet.io/faq/multi_devices.html>`
for more details.
Parameters
diff --git a/scala-package/README.md b/scala-package/README.md
index b2d3e9a..1494c0e 100644
--- a/scala-package/README.md
+++ b/scala-package/README.md
@@ -80,7 +80,7 @@ java -Xmx4G -cp \
```
If you've compiled with `USE_DIST_KVSTORE` enabled, the python tools in
`mxnet/tracker` can be used to launch distributed training.
-The following command runs the above example using 2 worker nodes (and 2
server nodes) in local. Refer to [Distributed
Training](http://mxnet.io/how_to/multi_devices.html) for more details.
+The following command runs the above example using 2 worker nodes (and 2
server nodes) in local. Refer to [Distributed
Training](http://mxnet.io/faq/multi_devices.html) for more details.
```bash
tracker/dmlc_local.py -n 2 -s 2 \
diff --git a/setup-utils/install-mxnet-osx-python.sh
b/setup-utils/install-mxnet-osx-python.sh
index 3cb5fcd..d0e9d5a 100755
--- a/setup-utils/install-mxnet-osx-python.sh
+++ b/setup-utils/install-mxnet-osx-python.sh
@@ -520,7 +520,7 @@ END
echo ":-)"
echo " "
echo "FYI : You can fine-tune MXNet run-time behavior using
environment variables described at:"
- echo " http://mxnet.io/how_to/env_var.html"
+ echo " http://mxnet.io/faq/env_var.html"
echo " "
echo "NEXT: Try the tutorials at: http://mxnet.io/tutorials"
echo " "
diff --git a/src/operator/custom/custom.cc b/src/operator/custom/custom.cc
index beb5f3d..164c2cc 100644
--- a/src/operator/custom/custom.cc
+++ b/src/operator/custom/custom.cc
@@ -364,7 +364,7 @@ NNVM_REGISTER_OP(Custom)
Custom operators should override required methods like `forward` and
`backward`.
The custom operator must be registered before it can be used.
-Please check the tutorial here: http://mxnet.io/how_to/new_op.html.
+Please check the tutorial here: http://mxnet.io/faq/new_op.html.
)code" ADD_FILELINE)
.set_num_inputs([](const NodeAttrs& attrs){
diff --git a/tools/caffe_converter/README.md b/tools/caffe_converter/README.md
index ac88fa1..d8ffc5c 100644
--- a/tools/caffe_converter/README.md
+++ b/tools/caffe_converter/README.md
@@ -10,4 +10,4 @@ python convert_caffe_modelzoo.py resnet-50
```
Please refer to
-[docs/how_to/caffe.md](../../docs/how_to/caffe.md) for more details.
+[docs/faq/caffe.md](../../docs/faq/caffe.md) for more details.
diff --git a/tools/caffe_translator/README.md b/tools/caffe_translator/README.md
index 1d5a77c..ad11161 100644
--- a/tools/caffe_translator/README.md
+++ b/tools/caffe_translator/README.md
@@ -27,9 +27,9 @@ Here is the list of command line parameters accepted by the
Caffe Translator:
- *solver-prototxt*: specifies the path to the solver prototxt to be
translated.
- *output-file*: specifies the file to write the translated output into.
- *params-file* (optional): specifies the .caffemodel file to initialize
parameters from.
-- *custom-data-layers* (optional): Specifies a comma-separated list of types
of the custom data layers used in the prototxt. The translator will use
[`CaffeDataIter`](https://mxnet.incubator.apache.org/how_to/caffe.html#use-io-caffedataiter)
to translate these layers to MXNet.
+- *custom-data-layers* (optional): Specifies a comma-separated list of types
of the custom data layers used in the prototxt. The translator will use
[`CaffeDataIter`](https://mxnet.incubator.apache.org/faq/caffe.html#use-io-caffedataiter)
to translate these layers to MXNet.
-**Note:** Translated code uses
[`CaffeDataIter`](https://mxnet.incubator.apache.org/how_to/caffe.html#use-io-caffedataiter)
to read from LMDB files. `CaffeDataIter` requires the number of examples in
LMDB file to be specified as a parameter. You can provide this information
before translation using a `#CaffeToMXNet` directive like shown below:
+**Note:** Translated code uses
[`CaffeDataIter`](https://mxnet.incubator.apache.org/faq/caffe.html#use-io-caffedataiter)
to read from LMDB files. `CaffeDataIter` requires the number of examples in
LMDB file to be specified as a parameter. You can provide this information
before translation using a `#CaffeToMXNet` directive like shown below:
```
data_param {
diff --git a/tools/caffe_translator/faq.md b/tools/caffe_translator/faq.md
index 81cdfb9..99d19fe 100644
--- a/tools/caffe_translator/faq.md
+++ b/tools/caffe_translator/faq.md
@@ -4,9 +4,9 @@
There is a couple of reasons why Caffe is required to run the translated code:
-1. The translator does not convert Caffe data layer to native MXNet code
because MXNet cannot read from LMDB files. Translator instead generates code
that uses
[`CaffeDataIter`](https://mxnet.incubator.apache.org/how_to/caffe.html#use-io-caffedataiter)
which can read LMDB files. `CaffeDataIter` needs Caffe to run.
+1. The translator does not convert Caffe data layer to native MXNet code
because MXNet cannot read from LMDB files. Translator instead generates code
that uses
[`CaffeDataIter`](https://mxnet.incubator.apache.org/faq/caffe.html#use-io-caffedataiter)
which can read LMDB files. `CaffeDataIter` needs Caffe to run.
-2. If the Caffe code to be translated uses custom layers, or layers that don't
have equivalent MXNet layers, the translator will generate code that will use
[CaffeOp](https://mxnet.incubator.apache.org/how_to/caffe.html#use-sym-caffeop).
CaffeOp needs Caffe to run.
+2. If the Caffe code to be translated uses custom layers, or layers that don't
have equivalent MXNet layers, the translator will generate code that will use
[CaffeOp](https://mxnet.incubator.apache.org/faq/caffe.html#use-sym-caffeop).
CaffeOp needs Caffe to run.
[**What version of Caffe prototxt can the translator
translate?**](#what_version_of_prototxt)
--
To stop receiving notification emails like this one, please contact
[email protected].