transformers
[pagged-attention] fix off-by-1 error in pagged attention generation
#39258
Merged

[pagged-attention] fix off-by-1 error in pagged attention generation #39258

kashif
kashif fix off-by-1 error in pagged attention generation
72f8cbd6
kashif kashif requested a review from ArthurZucker ArthurZucker 207 days ago
kashif formatting
9c6262f4
kashif use update_with_token
ec8e895e
kashif Merge branch 'main' into paged-attention-max-gen
5af10598
HuggingFaceDocBuilderDev
ArthurZucker
ArthurZucker approved these changes on 2025-07-08
kashif kashif merged db05e4ff into main 206 days ago
kashif kashif deleted the paged-attention-max-gen branch 206 days ago
kashif kashif added for patch

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone