were commented on issue #5741:
URL: https://github.com/apache/incubator-tvm/issues/5741#issuecomment-640097428


   #5498 Kind of related to this one.
   The current TensorCore code generation is tricky --- since we have no 
fundamental wrap reduction support in TVM, generating TensorCore code 
inevitably messes up the thread bind. No threadIdxwas bound to a reduce axis 
before, but we inevitably need a threadIdx for that reduce axis. 
   
   Is it possible to somehow represent the warp wise reduce in the schedule, 
and then the TIR analyzer and rewriter can detect this opportunity of matching 
TensorCore. 


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to