onnxruntime
a8df6c35
- Support flash attention on 2d attention mask for gpt2 left padding. (#14215)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
2 years ago
Support flash attention on 2d attention mask for gpt2 left padding. (#14215)
References
#14215 - Support flash attention on 2d attention mask for gpt2 left padding.
Author
zhanghuanrong
Parents
30b9f5dd
Loading