transformers
[`FA-2`] Add Flash Attention to `Phi`
#27661
Merged

[`FA-2`] Add Flash Attention to `Phi` #27661

susnato
susnato
younesbelkada
susnato
younesbelkada
susnato
younesbelkada
younesbelkada approved these changes on 2023-11-23
younesbelkada younesbelkada requested a review from ArthurZucker ArthurZucker 2 years ago
ArthurZucker
ArthurZucker commented on 2023-11-23
younesbelkada
younesbelkada approved these changes on 2023-12-06
susnato
susnato
younesbelkada
younesbelkada commented on 2023-12-06
susnato add FA and modify doc file
cdcd671c
susnato test_flash_attn_2_generate_padding_right test overwritten
17185d89
susnato comment
a33fa732
susnato modify persimmon modeling file
e16090b1
susnato added speedup graph
30cac2e5
susnato more changes
9e22498e
susnato susnato force pushed to 9e22498e 2 years ago
susnato
younesbelkada
younesbelkada approved these changes on 2023-12-06
younesbelkada younesbelkada requested a review from ArthurZucker ArthurZucker 2 years ago
ArthurZucker
ArthurZucker approved these changes on 2023-12-07
ArthurZucker ArthurZucker merged f84d85ba into main 2 years ago
fxmarty

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone