transformers
Fix llama model sdpa attention forward function masking bug when output_attentions=True
#30652
Merged

Fix llama model sdpa attention forward function masking bug when output_attentions=True #30652

Aladoro
Aladoro Fix llama model forward function with attention=True, same-length enc…
514c1c32
Aladoro Fix style
6c0fa7bb
Aladoro
Aladoro propagate fix to modeling_cohere, gemma, dbrx, and olmo (which copy t…
0d91bea9
Aladoro Fix style
894c14b8
ArthurZucker
ArthurZucker commented on 2024-05-06
Aladoro ignore unnecessary sdpa mask converter when output_attentions=True
25843087
Aladoro Merge branch 'huggingface:main' into fix-llama-mask-output-attn
c7bdc95d
Aladoro
ArthurZucker
ArthurZucker
Aladoro Merge branch 'huggingface:main' into fix-llama-mask-output-attn
8d793a38
Aladoro
HuggingFaceDocBuilderDev
Aladoro
ArthurZucker
ArthurZucker commented on 2024-05-09
Aladoro Merge branch 'huggingface:main' into fix-llama-mask-output-attn
fc143acf
Aladoro add tests checking sdpa and eager outputs match when output_attention…
3e0fada2
Aladoro
ArthurZucker
ArthurZucker commented on 2024-05-15
Aladoro Split if statements in two lines
9acc1190
Aladoro Merge branch 'huggingface:main' into fix-llama-mask-output-attn
08dbd4bb
ArthurZucker
ArthurZucker approved these changes on 2024-05-15
ArthurZucker
Aladoro Fix formatting
9b79aee3
Aladoro Add fix to new jetmoe model
dd699233
Aladoro Add missing output_attentions argument to jetmoe mask creation
ad4aded6
Aladoro
ArthurZucker ArthurZucker merged 4b3eb19f into main 1 year ago
ArthurZucker
Aladoro Aladoro deleted the fix-llama-mask-output-attn branch 1 year ago
gante

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone