vllm
[ROCm] Enable chunked prefill/paged attention in MLA on ROCm
#14316
Merged

[ROCm] Enable chunked prefill/paged attention in MLA on ROCm #14316

SageMoore
SageMoore init
ae056e14
github-actions
vincent-4
vincent-4 requested changes on 2025-03-05
hongxiayang hongxiayang added rocm
SageMoore cleanup boolean logic
f1dbffb0
shajrawi
shajrawi approved these changes on 2025-03-06
shajrawi
houseroad
houseroad approved these changes on 2025-03-06
LucasWilkinson
LucasWilkinson requested changes on 2025-03-07
SageMoore comments
8f9664db
hongxiayang
hongxiayang approved these changes on 2025-03-10
LucasWilkinson
LucasWilkinson approved these changes on 2025-03-10
LucasWilkinson LucasWilkinson enabled auto-merge (squash) 279 days ago
github-actions github-actions added ready
vincent-4
vincent-4 approved these changes on 2025-03-11
SageMoore Merge branch 'main' of https://github.com/neuralmagic/vllm into sage/…
e9673140
LucasWilkinson LucasWilkinson merged d9f83d62 into main 277 days ago
SageMoore SageMoore deleted the sage/amd-deepseek branch 277 days ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone