NicolaLancellotti commented on code in PR #10959:
URL: https://github.com/apache/tvm/pull/10959#discussion_r863945160
##########
src/tir/contrib/ethosu/passes.cc:
##########
@@ -110,6 +110,98 @@ tvm::transform::Pass HoistAllocates() {
TVM_REGISTER_GLOBAL("tir.contrib.ethos-u.HoistAllocates").set_body_typed(HoistAllocates);
+/*!
+ * \brief Reorders copy and compute nodes in such a way that independent DMA
copies,
+ * and computes happen in parallel.
+ * Copies to buffers with local scope are not reordered, indeed they copy LUT
+ * into the SHRAM which already happens in parallel with copying weights into
+ * the weights encoder.
+ */
+class CopyComputeReorderingMutator : public StmtExprMutator {
+ public:
+ CopyComputeReorderingMutator(int max_copy_movements) :
_max_copy_movements{max_copy_movements} {}
+
+ PrimFunc operator()(PrimFunc main_func) {
+ if (_max_copy_movements > 0) {
+ auto n{main_func.CopyOnWrite()};
+ n->body = this->VisitStmt(main_func->body);
+ return GetRef<PrimFunc>(n);
+ }
+ return main_func;
+ }
+
+ private:
+ Stmt VisitStmt_(const SeqStmtNode* op) override {
+ if (op->size() <= 1) {
+ return StmtExprMutator::VisitStmt_(op);
+ }
+
+ auto seq_stmt{GetRef<SeqStmt>(op)};
+ std::vector<Stmt> new_seq(seq_stmt->size());
+ std::copy(seq_stmt->seq.begin(), seq_stmt->seq.end(), new_seq.begin());
+
+ for (size_t index{}; index < new_seq.size(); ++index) {
+ for (int offset{}; offset < _max_copy_movements; ++offset) {
+ auto i{index - offset};
+ if (i > 0 && !stmt_is_copy(new_seq[i - 1]) &&
stmt_is_global_copy(new_seq[i])) {
+ std::swap(new_seq[i], new_seq[i - 1]);
+ } else {
+ break;
+ }
+ }
+ }
+
+ auto n{CopyOnWrite(op)};
+ n->seq = std::move(new_seq);
+ return Stmt{n};
+ }
+
+ tvm::runtime::Array<tvm::PrimExpr> get_stmt_args(Stmt stmt) {
Review Comment:
Done.
##########
src/tir/contrib/ethosu/passes.cc:
##########
@@ -110,6 +110,98 @@ tvm::transform::Pass HoistAllocates() {
TVM_REGISTER_GLOBAL("tir.contrib.ethos-u.HoistAllocates").set_body_typed(HoistAllocates);
+/*!
+ * \brief Reorders copy and compute nodes in such a way that independent DMA
copies,
+ * and computes happen in parallel.
+ * Copies to buffers with local scope are not reordered, indeed they copy LUT
+ * into the SHRAM which already happens in parallel with copying weights into
+ * the weights encoder.
+ */
+class CopyComputeReorderingMutator : public StmtExprMutator {
+ public:
+ CopyComputeReorderingMutator(int max_copy_movements) :
_max_copy_movements{max_copy_movements} {}
+
+ PrimFunc operator()(PrimFunc main_func) {
+ if (_max_copy_movements > 0) {
+ auto n{main_func.CopyOnWrite()};
+ n->body = this->VisitStmt(main_func->body);
+ return GetRef<PrimFunc>(n);
+ }
+ return main_func;
+ }
+
+ private:
+ Stmt VisitStmt_(const SeqStmtNode* op) override {
+ if (op->size() <= 1) {
+ return StmtExprMutator::VisitStmt_(op);
+ }
+
+ auto seq_stmt{GetRef<SeqStmt>(op)};
+ std::vector<Stmt> new_seq(seq_stmt->size());
+ std::copy(seq_stmt->seq.begin(), seq_stmt->seq.end(), new_seq.begin());
+
+ for (size_t index{}; index < new_seq.size(); ++index) {
+ for (int offset{}; offset < _max_copy_movements; ++offset) {
+ auto i{index - offset};
+ if (i > 0 && !stmt_is_copy(new_seq[i - 1]) &&
stmt_is_global_copy(new_seq[i])) {
+ std::swap(new_seq[i], new_seq[i - 1]);
+ } else {
+ break;
+ }
+ }
+ }
+
+ auto n{CopyOnWrite(op)};
+ n->seq = std::move(new_seq);
+ return Stmt{n};
+ }
+
+ tvm::runtime::Array<tvm::PrimExpr> get_stmt_args(Stmt stmt) {
+ auto eval_node{stmt.as<EvaluateNode>()};
+ ICHECK(eval_node) << "Expected statement to be an evaluate node, but was "
+ << stmt->GetTypeKey();
+ auto call_node{eval_node->value.as<CallNode>()};
+ ICHECK(call_node) << "Expected expression to be a call node, but was "
+ << eval_node->value->GetTypeKey();
+ return call_node->args;
+ }
+
+ bool stmt_is_copy(Stmt stmt) {
Review Comment:
Done.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]