Fix attention mask in mamba layers #41790
vasqu
commented
on 2025-10-22
not all mamba models are like LFM
3462502d
zucchini-nlp
force pushed
from
efc2ee0f
to
3462502d
203 days ago
Merge branch 'main' into fix-mamba-pad-masking
a57ffa63
compile friendly
f07dbf49
vasqu
commented
on 2025-10-22
adjust slow tests expectation
4f760ec0
naming
a8b35a02
vasqu
approved these changes
on 2025-10-22
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub