Add Flash Attention 2 to M2M100 model (#30256)
* Added flash attention 2.
* Fixes.
* Fix inheritance.
* Fixed init.
* Remove stuff.
* Added documentation.
* Add FA2 to M2M100 documentation.
* Add test.
* Fixed documentation.
* Update src/transformers/models/m2m_100/modeling_m2m_100.py
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
* Update docs/source/en/model_doc/nllb.md
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Fixed variable name.
---------
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>