[CUDA] Fuse add bias and transpose into one kernel in Attention #12670
fuse add bias and transpose in attention
d2680c5e
tianleiwu
marked this pull request as draft 3 years ago
tianleiwu
changed the title Fuse add bias and transpose into one kernel in Attention [CUDA] Fuse add bias and transpose into one kernel in Attention 3 years ago
format
933ee8dc
tianleiwu
marked this pull request as ready for review 3 years ago
wangyems
approved these changes
on 2022-08-22
tianleiwu
merged
8d78f96d
into main 3 years ago
tianleiwu
deleted the tlwu/bert_bias_transpose branch 3 years ago
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub