vllm
[XPU][CT] support per-channel quantization in xpu fp8 linear method
#38316
Merged

[XPU][CT] support per-channel quantization in xpu fp8 linear method #38316

jikunshang merged 1 commit into vllm-project:main from yma11:per-channel
yma11
yma11 yma11 requested a review from mgoin mgoin 43 days ago
yma11 yma11 requested a review from robertgshaw2-redhat robertgshaw2-redhat 43 days ago
yma11 yma11 requested a review from tlrmchlsmth tlrmchlsmth 43 days ago
yma11 yma11 requested a review from yewentao256 yewentao256 43 days ago
yma11 yma11 requested a review from pavanimajety pavanimajety 43 days ago
claude
claude commented on 2026-03-27
gemini-code-assist
gemini-code-assist commented on 2026-03-27
yma11
mergify mergify added intel-gpu
mergify
mergify mergify added needs-rebase
yma11 Add per-channel quantized model in compressed-tensors
ed9a974d
yma11 yma11 force pushed to ed9a974d 28 days ago
mergify mergify removed needs-rebase
jikunshang
jikunshang approved these changes on 2026-04-12
jikunshang jikunshang enabled auto-merge (squash) 27 days ago
github-actions github-actions added ready
disabled auto-merge 27 days ago
Manually disabled by user
jikunshang jikunshang changed the title [XPU] Add per-channel quantized model in compressed-tensors [XPU] support per-channel quantization in fp8 linear method 27 days ago
jikunshang jikunshang changed the title [XPU] support per-channel quantization in fp8 linear method [XPU] support per-channel quantization in xpu fp8 linear method 27 days ago
jikunshang jikunshang changed the title [XPU] support per-channel quantization in xpu fp8 linear method [XPU][CT] support per-channel quantization in xpu fp8 linear method 27 days ago
jikunshang jikunshang enabled auto-merge (squash) 27 days ago
jikunshang jikunshang merged 394ff869 into main 27 days ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone