Thank you for the explanation. Sorry I didn't realize the proposal is for 
inference only.

Then how do you think the amp_cast and amp_multicast in this proposal can work 
with the existing INT8 quantization workflow which I think should also be 
considered as 'mixed precision'.

-----Original Message-----
From: Anirudh Subramanian [mailto:[email protected]] 
Sent: Monday, April 29, 2019 10:25 PM
To: [email protected]
Subject: Re: Proposal for Conversion from FP32 to Mixed Precision Models

Hi Tao,

The APIs proposed: "convert_model" and "convert_block" are mainly for inference 
use cases, where customers bring a FP32 model to convert it to a mixed 
precision model to get improved performance while not losing out on the 
accuracy.
The PR: https://github.com/apache/incubator-mxnet/pull/14173 is supposed to 
handle the training use cases and this proposal doesn't cover the AMP feature 
added in the PR. I think ptrendx@ and canoerst@ are better equipped to answer 
questions 1 and 2.

> - more generally, what will be saved when users want to serialize 
> their
model to disk?

Lets say users want to save converted mixed precision model used for inference 
to disk. It will save both, the symbol with the amp_cast and amp_multicast 
operators and the params (which are casted if necessary).

Anirudh


On Mon, Apr 29, 2019 at 6:55 AM Lv, Tao A <[email protected]> wrote:

> Thank you for sharing this, Anirudh.
>
> Curious to know:
> - what will be saved in a training checkpoint or snapshot? Can it be 
> resumed on another platform which might not support the lower 
> precision the previous one used?
> - what will be saved in the final symbol.json and params file when 
> training is finished?
> - more generally, what will be saved when users want to serialize 
> their model to disk?
>
> Thank you,
> -tao
>
> -----Original Message-----
> From: Anirudh Subramanian [mailto:[email protected]]
> Sent: Monday, April 29, 2019 7:00 PM
> To: [email protected]
> Subject: Proposal for Conversion from FP32 to Mixed Precision Models
>
> Hi all,
>
> I have created a doc for conversion from FP32 to Mixed Precision Models:
>
> https://cwiki.apache.org/confluence/display/MXNET/Conversion+from+FP32
> +to+Mixed+Precision+Models
>
> I look forward to your feedback on the same.
>
> Thanks,
> Anirudh
>

Reply via email to