vllm
gptq marlin quantization support for fused moe with lora
#30254
Open

gptq marlin quantization support for fused moe with lora #30254

Bhanu068 wants to merge 1 commit into vllm-project:main from Bhanu068:gptq_moe_lora_feat
Bhanu068
Bhanu068 gptq marlin quantization support for fused moe with lora
6a6d0809
Bhanu068 Bhanu068 requested a review from mgoin mgoin 18 hours ago
Bhanu068 Bhanu068 requested a review from robertgshaw2-redhat robertgshaw2-redhat 18 hours ago
Bhanu068 Bhanu068 requested a review from tlrmchlsmth tlrmchlsmth 18 hours ago
Bhanu068 Bhanu068 requested a review from yewentao256 yewentao256 18 hours ago
Bhanu068 Bhanu068 requested a review from pavanimajety pavanimajety 18 hours ago
gemini-code-assist
gemini-code-assist commented on 2025-12-08
chatgpt-codex-connector
chatgpt-codex-connector commented on 2025-12-08
jeejeelee
jeejeelee commented on 2025-12-09

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone