👏

On Mon, Oct 2, 2017 at 8:02 PM Seb Kiureghian <[email protected]> wrote:

> It would be awesome if MXNet were the first DL framework to support Nvidia
> Volta. What do you all think about cutting a v0.12 release once that
> integration is ready?
>
> On Wed, Sep 27, 2017 at 10:38 PM, Jun Wu <[email protected]> wrote:
>
> > I had been working on the sparse tensor project with Haibin. After it was
> > wrapped up for the first stage, I started my work on the quantization
> > project (INT-8 inference). The benefits of using quantized models for
> > inference include much higher inference throughput than FP32 model with
> > acceptable accuracy loss and compact models saved on small devices. The
> > work currently aims at quantizing ConvNets, and we will consider
> expanding
> > it to RNN networks after getting good results for images. Meanwhile, it's
> > expected to support quantization on CPU, GPU, and mobile devices.
> >
>
-- 


Dominic Divakaruni
206.475.9200 Cell

Reply via email to