[`FA2`] Add flash attention for `GPT-Neo` (#26486)
* added flash attention for gpt-neo
* small change
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* readme updated
* .
* changes
* removed padding_mask
* Update src/transformers/models/gpt_neo/modeling_gpt_neo.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
---------
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>