fhieber commented on pull request #20375: URL: https://github.com/apache/incubator-mxnet/pull/20375#issuecomment-868284387
> Would it be more beneficial to have the full multihead attention primitive exposed in the API instead (it could still be implemented by the interleaved matmuls)? Both Keras and pyTorch have it as a layer, and this could make the usage easier. What do you think @barry-jin? Can we have both? We'd like to upgrade Sockeye to MX2 without changing too many things at the same time. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [email protected]
