leandron commented on a change in pull request #15:
URL: https://github.com/apache/tvm-rfcs/pull/15#discussion_r684111762



##########
File path: rfcs/0013_Arm_CMSIS-NN_Integration.md
##########
@@ -0,0 +1,113 @@
+- Feature Name: [RFC] Use CMSIS-NN with TVM
+- Start Date: July 2021
+- RFC PR: https://github.com/apache/tvm-rfcs/pull/15
+- GitHub Issue: https://github.com/apache/tvm/issues/8646
+
+# Summary
+
+This RFC introduces plan of integration of CMSIS-NN library into TVM. It 
consists of efficient kernels targetted for Arm's Cortex-M architecture.
+
+Please refer to the following pages for more details on CMSIS-NN.
+https://arm-software.github.io/CMSIS_5/NN/html/index.html
+https://github.com/ARM-software/CMSIS_5/tree/develop/CMSIS/NN
+
+First PR in the series of PRs to fullfill this integration would be graph 
partitioner for softmax int8. Detailed plan can found below in this RFC.

Review comment:
       ```suggestion
   First PR in the series of PRs to fulfill this integration would be graph 
partitioner for softmax int8. Detailed plan can found below in this RFC.
   ```

##########
File path: rfcs/0013_Arm_CMSIS-NN_Integration.md
##########
@@ -0,0 +1,113 @@
+- Feature Name: [RFC] Use CMSIS-NN with TVM
+- Start Date: July 2021
+- RFC PR: https://github.com/apache/tvm-rfcs/pull/15
+- GitHub Issue: https://github.com/apache/tvm/issues/8646
+
+# Summary
+
+This RFC introduces plan of integration of CMSIS-NN library into TVM. It 
consists of efficient kernels targetted for Arm's Cortex-M architecture.

Review comment:
       ```suggestion
   This RFC introduces plan of integration of CMSIS-NN library into TVM. It 
consists of efficient kernels targeted for Arm's Cortex-M architecture.
   ```

##########
File path: rfcs/0013_Arm_CMSIS-NN_Integration.md
##########
@@ -0,0 +1,113 @@
+- Feature Name: [RFC] Use CMSIS-NN with TVM
+- Start Date: July 2021
+- RFC PR: https://github.com/apache/tvm-rfcs/pull/15
+- GitHub Issue: https://github.com/apache/tvm/issues/8646
+
+# Summary
+
+This RFC introduces plan of integration of CMSIS-NN library into TVM. It 
consists of efficient kernels targetted for Arm's Cortex-M architecture.
+
+Please refer to the following pages for more details on CMSIS-NN.
+https://arm-software.github.io/CMSIS_5/NN/html/index.html
+https://github.com/ARM-software/CMSIS_5/tree/develop/CMSIS/NN
+
+First PR in the series of PRs to fullfill this integration would be graph 
partitioner for softmax int8. Detailed plan can found below in this RFC.
+
+
+# Motivation
+
+CMSIS-NN library consists of hand-tuned kernels that are suitable for Cortex-M 
and are compliant with the quantization scheme used in Tensorflow Lite. They 
have been optimized for better performance and small memory footprint which is 
required on these embedded devices and it would make sense for TVM to reuse 
these while generating code for Cortex-M. They have been integrated with the 
TensorFlow Lite Micro project.
+
+
+# Guide-level explanation
+
+TVM's external code generation infrastructure allows for the automatic 
partitoning and code generation using the external compiler. Partitioned 
subgraphs containing operator(s) targetted for Cortex-M can then be translated 
into the CMSIS-NN C APIs which eventually become part of MLF. For this 
integration, we are heavily dependent on the TVM's infrastructure for external 
code generation.

Review comment:
       nit: introducing the acronym in the first usage.
   
   ```suggestion
   TVM's external code generation infrastructure allows for the automatic 
partitioning and code generation using the external compiler. Partitioned 
subgraphs containing operator(s) targetted for Cortex-M can then be translated 
into the CMSIS-NN C APIs which eventually become part of Model Library Format 
(MLF). For this integration, we are heavily dependent on the TVM's 
infrastructure for external code generation.
   ```

##########
File path: rfcs/0013_Arm_CMSIS-NN_Integration.md
##########
@@ -0,0 +1,113 @@
+- Feature Name: [RFC] Use CMSIS-NN with TVM
+- Start Date: July 2021
+- RFC PR: https://github.com/apache/tvm-rfcs/pull/15
+- GitHub Issue: https://github.com/apache/tvm/issues/8646
+
+# Summary
+
+This RFC introduces plan of integration of CMSIS-NN library into TVM. It 
consists of efficient kernels targetted for Arm's Cortex-M architecture.
+
+Please refer to the following pages for more details on CMSIS-NN.
+https://arm-software.github.io/CMSIS_5/NN/html/index.html
+https://github.com/ARM-software/CMSIS_5/tree/develop/CMSIS/NN
+
+First PR in the series of PRs to fullfill this integration would be graph 
partitioner for softmax int8. Detailed plan can found below in this RFC.
+
+
+# Motivation
+
+CMSIS-NN library consists of hand-tuned kernels that are suitable for Cortex-M 
and are compliant with the quantization scheme used in Tensorflow Lite. They 
have been optimized for better performance and small memory footprint which is 
required on these embedded devices and it would make sense for TVM to reuse 
these while generating code for Cortex-M. They have been integrated with the 
TensorFlow Lite Micro project.
+
+
+# Guide-level explanation
+
+TVM's external code generation infrastructure allows for the automatic 
partitoning and code generation using the external compiler. Partitioned 
subgraphs containing operator(s) targetted for Cortex-M can then be translated 
into the CMSIS-NN C APIs which eventually become part of MLF. For this 
integration, we are heavily dependent on the TVM's infrastructure for external 
code generation.
+
+If a user runs tvmc, they will get a MLF format archive which calls out to the 
CMSIS operators.
+
+```
+tvmc --target=c,cmsisnn --output-format=mlf --executor=aot

Review comment:
       Small nit on the tvmc command line, that will probably look more like 
"offload all that is possible to CMSIS-NN and the rest to CPU with the C 
backend", hence a small semantic fix:
   
   ```suggestion
   tvmc --target="cmsisnn,c" --output-format=mlf --executor=aot
   ```




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to