onnxruntime
Fix GPT-2 no-past attention fusion for transformers >= 4.27
#27449
Merged

Fix GPT-2 no-past attention fusion for transformers >= 4.27 #27449

Rishi-Dave
Rishi-Dave
xadupre
xadupre commented on 2026-02-25
Rishi-Dave
xadupre
xadupre dismissed these changes on 2026-02-26
tianleiwu
azure-pipelines
Rishi-Dave Rishi-Dave dismissed their stale review via 65bee872 14 days ago
Rishi-Dave Rishi-Dave force pushed from 444e9629 to 65bee872 14 days ago
tianleiwu
Rishi-Dave Fix GPT-2 no-past attention fusion for transformers >= 4.27
2030a97f
Rishi-Dave Use verify_fusion in no-past attention test per review feedback
4b881b5a
Rishi-Dave Fix line length lint in no-past attention test
604aa892
Rishi-Dave Rishi-Dave force pushed from 65bee872 to 604aa892 11 days ago
Rishi-Dave
tianleiwu
azure-pipelines
tianleiwu
tianleiwu approved these changes on 2026-03-05
tianleiwu tianleiwu enabled auto-merge (squash) 10 days ago
tianleiwu tianleiwu merged 01a56ce6 into main 10 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone