transformers
Fix how we compute the final non-padding token for ForSequenceClassification models
#35911
Merged

Fix how we compute the final non-padding token for ForSequenceClassification models #35911

Rocketknight1
Rocketknight1 Fix how we compute the final non-padding token for Gemma (and probabl…
f74d4dda
Rocketknight1 .size() -> .shape[]
067d99a2
Rocketknight1 Propagating changes to other models
3e381f41
Rocketknight1 Propagating changes to other models
7cc13967
Rocketknight1 Change it for all ForSequenceClassification models
d3b7c994
Rocketknight1 Fix batch dim
1c11edc4
Rocketknight1 More TF fixes
13a670ee
Rocketknight1 Copy the TF fix around as well
c671810f
Rocketknight1 Correct layer name for TFCTRL
8ccde63d
Rocketknight1 Rocketknight1 force pushed from 4620592e to 8ccde63d 1 year ago
HuggingFaceDocBuilderDev
Rocketknight1
Rocketknight1
Cyrilvallez
Cyrilvallez commented on 2025-01-29
Rocketknight1 Cleaner .to()
8c69579f
Rocketknight1 Clean up the nested if-else
172cfd76
Rocketknight1 Use argmax() instead of .max().values
26d554e3
Cyrilvallez
Cyrilvallez approved these changes on 2025-01-30
Rocketknight1 Rocketknight1 merged 694aaa7f into main 1 year ago
Rocketknight1 Rocketknight1 deleted the fix_sequence_classification_padding_side branch 1 year ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone