[Flash Attention 2] Add flash attention 2 for GPT-J #28295
initial implementation of flash attention for gptj
6419b041
modify flash attention and overwrite test_flash_attn_2_generate_paddi…
3a9e31f9
update flash attention support list
cb656559
remove the copy line in the `CodeGenBlock`
e47ef133
bytebarde
force pushed
to
e47ef133
2 years ago
bytebarde
changed the title [Flash Attention 2] [WIP] Add flash attention 2 for GPT-J [Flash Attention 2] Add flash attention 2 for GPT-J 2 years ago
address copy mechanism
0c31cb3f
Update src/transformers/models/gptj/modeling_gptj.py
def626ef
Merge branch 'huggingface:main' into flash_attn_gptj
6e9c7070
Add GPTJ attention classes
af0752ea
Merge branch 'huggingface:main' into flash_attn_gptj
cd73e337
add expected outputs in the gptj test
cb265c73
Ensure repo consistency with 'make fix-copies'
2b489b06
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub