transformers
87be5595
- Fix attention mask in mamba layers (#41790)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
197 days ago
Fix attention mask in mamba layers (#41790) * not all mamba models are like LFM * compile friendly * adjust slow tests expectation * naming
References
#41790 - Fix attention mask in mamba layers
Author
zucchini-nlp
Parents
2ca506ca
Loading