transformers
7edc9931 - don't zero out the attention_mask when using sliding window with flash attention (#31670)

Commit
1 year ago
don't zero out the attention_mask when using sliding window with flash attention (#31670) * don't zero out the attention_mask when using sliding window with flash attention * chore: lint
Author
Committer
Parents
Loading