vllm
[ROCm] Support MLA with nhead<16 and FP8 KV cache for TP=8 (Kimi K2.5/Linear)
#35850
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
2
Changes
View On
GitHub
[ROCm] Support MLA with nhead<16 and FP8 KV cache for TP=8 (Kimi K2.5/Linear)
#35850
tjtanaa
merged 2 commits into
vllm-project:main
from
ChuanLi1101:fix/mla-nhead-fp8-kv-upstream
ChuanLi1101
requested a review
from
tjtanaa
9 days ago
mergify
added
rocm
mergify
added
v1
gemini-code-assist
commented on 2026-03-03
zejunchen-zejun
approved these changes on 2026-03-03
tjtanaa
approved these changes on 2026-03-05
tjtanaa
added
ready
tjtanaa
enabled auto-merge (squash)
7 days ago
mergify
added
needs-rebase
[ROCm] Support MLA with nhead<16 and FP8 KV cache for TP=8 (Kimi K2.5…
71cb317a
disabled auto-merge
6 days ago
Head branch was pushed to by a user without write access
ChuanLi1101
force pushed
from
db465b38
to
71cb317a
6 days ago
mergify
removed
needs-rebase
Merge branch 'main' into fix/mla-nhead-fp8-kv-upstream
04adb9e7
tjtanaa
enabled auto-merge (squash)
6 days ago
tjtanaa
merged
c188749b
into main
6 days ago
Login to write a write a comment.
Login via GitHub
Reviewers
tjtanaa
zejunchen-zejun
gemini-code-assist
Assignees
No one assigned
Labels
rocm
ready
v1
Milestone
No milestone
Login to write a write a comment.
Login via GitHub