vllm
32aa74c0 - [ROCm][FP8][Kernel] FP8 quantization fused into Custom Paged Attention (#17139)

Commit
349 days ago
[ROCm][FP8][Kernel] FP8 quantization fused into Custom Paged Attention (#17139) Signed-off-by: Gregory Shtrasberg <Gregory.Shtrasberg@amd.com>
Author
Parents
Loading