transformers
81233c06
- Flash-Attn: fix generation when no attention mask or no pading (#32241)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
Flash-Attn: fix generation when no attention mask or no pading (#32241) * fix * fix prev test (half of failures) * [run-slow] llama, gemma2 * [run-slow] llama, gemma2
References
#32241 - Flash-Attn: fix generation when no attention mask or no pading
Author
zucchini-nlp
Parents
27c7f971
Loading