[webgpu] Optimize AttentionPrepare #26850
[webgpu] Optimize attentionPrepare
e15bd722
change the splitQKV to BNSH format
8d38f51b
fix the bugs
dfb98e3f
remove debugging codes
cbb8ef22
SplitPackedQKV with BSD format
0067a2d9
make SplitPackedQKV work on GQA
efdc783d
Add component support to SplitPackedQKV
ceea4d88
qjia7
marked this pull request as ready for review 105 days ago
address comments from Copilot
419e2630
guschmue
approved these changes
on 2026-01-05
qjia7
merged
5bc10a39
into main 91 days ago
qjia7
deleted the attention_prepare branch 91 days ago
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub