transformers
81233c06 - Flash-Attn: fix generation when no attention mask or no pading (#32241)

Commit
1 year ago
Flash-Attn: fix generation when no attention mask or no pading (#32241) * fix * fix prev test (half of failures) * [run-slow] llama, gemma2 * [run-slow] llama, gemma2
Author
Parents
Loading