xla
[Kernel] support kv cache quantization in ragged attention kernel
#9249
Merged

[Kernel] support kv cache quantization in ragged attention kernel #9249

qihqi merged 6 commits into master from chengji/improve-attn
yaochengji
yaochengji [Kernel] support kv cache quantization in ragged attention kernel
b57e4578
yaochengji yaochengji requested a review from vanbasten23 vanbasten23 318 days ago
yaochengji modify python op
d3655f89
yaochengji fix test
52a00099
yaochengji yaochengji force pushed from 057c40b0 to 52a00099 318 days ago
yaochengji fix test
5749e4c6
yaochengji fix test
070138a4
yaochengji yaochengji force pushed from cfee66b5 to 070138a4 318 days ago
vanbasten23
vanbasten23 commented on 2025-05-28
vanbasten23
vanbasten23 commented on 2025-05-28
vanbasten23
vanbasten23 commented on 2025-05-28
vanbasten23
vanbasten23 commented on 2025-05-28
vanbasten23
vanbasten23 commented on 2025-05-28
yaochengji fix comments
80412afe
vanbasten23
vanbasten23 approved these changes on 2025-05-29
qihqi qihqi merged 2edcd2e3 into master 316 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone