huanmei9 commented on issue #8717:
URL: https://github.com/apache/tvm/issues/8717#issuecomment-963892608


   Hi, @Alex-grovety . I have some questions in your reply 
   
   > Hello @Mousius, I would like to clarify something about the part of the 
problem related to casting to int16. Please look at 
`/tvm/python/tvm/relay/qnn/op/legalizations.py->_qnn_conv2d_legalize_arm_cpu(attrs,
 inputs, types)` Here we see that `use_int8_on_arm == false` for depthwise 
Conv2d operators, and is_fast_int8_on_arm() also returns false if the target 
string doesn’t contain `-mattr=+v8.2a,+dotprod`.
   > 
   > So, if we have target string like `llvm -device=arm_cpu`, we will cast to 
int16. Could you please point, which target string do you use and tell if my 
conclusion is correct or not.
   > 
   > PS: I may be misunderstanding what it means “the unusual cast operator at 
the end is making the inter-fused-operator tensors 16-bit”, if that’s true, 
could you please clarify this for me?
   > 
   > PS2: I use whole MobileNet, maybe I should cut out just one operator and 
check it?
   > 
   > Thank you!
   
   My questions are:
   
   1. What kind of arm device support ` is_fast_int8_on_arm()`? Could you give 
some examples?
   2. Why the operator need to cast to int16 if device neither 
`use_int8_on_arm` nor `is_fast_int8_on_arm`?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to