onnxruntime
8d78f96d - [CUDA] Fuse add bias and transpose into one kernel in Attention (#12670)

Commit
3 years ago
[CUDA] Fuse add bias and transpose into one kernel in Attention (#12670) * fuse add bias and transpose in attention
Author
Parents
Loading