transformers
7edc9931
- don't zero out the attention_mask when using sliding window with flash attention (#31670)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
don't zero out the attention_mask when using sliding window with flash attention (#31670) * don't zero out the attention_mask when using sliding window with flash attention * chore: lint
Author
winglian
Committer
ArthurZucker
Parents
e3cb841c
Loading