transformers
87be5595 - Fix attention mask in mamba layers (#41790)

Commit
197 days ago
Fix attention mask in mamba layers (#41790) * not all mamba models are like LFM * compile friendly * adjust slow tests expectation * naming
Author
Parents
Loading