vllm
a1257fd1
- [Kernel] Add FP8 KV cache support to Triton MLA decode attention (#34597)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
52 days ago
[Kernel] Add FP8 KV cache support to Triton MLA decode attention (#34597) Signed-off-by: grimulkan <grimulkan@gmail.com>
References
#34597 - [Kernel] Add FP8 KV cache support to Triton MLA decode attention
Author
grimulkan
Parents
abcffbba
Loading