transformers
d03a3ca6 - [`OPT`] Fix attention scaling (#38290)

Commit
202 days ago
[`OPT`] Fix attention scaling (#38290) * fix opt attention scaling * add comment to why we do this
Author
Parents
Loading