vllm
Revert "[Llama4,Quantization] Simplify and generalize logic for Q/K permutations in quantized self-attn layers "
#34997
Merged

Revert "[Llama4,Quantization] Simplify and generalize logic for Q/K permutations in quantized self-attn layers " #34997

LucasWilkinson
LucasWilkinson Revert "[Llama4,Quantization] Simplify and generalize logic for Q/K p…
365da4f8
dosubot
mergify mergify added llama
gemini-code-assist
gemini-code-assist commented on 2026-02-20
LucasWilkinson LucasWilkinson added ready
LucasWilkinson LucasWilkinson requested a review from DarkLight1337 DarkLight1337 44 days ago
LucasWilkinson LucasWilkinson requested a review from DarkLight1337 DarkLight1337 44 days ago
LucasWilkinson LucasWilkinson requested a review from robertgshaw2-redhat robertgshaw2-redhat 44 days ago
LucasWilkinson LucasWilkinson requested a review from robertgshaw2-redhat robertgshaw2-redhat 44 days ago
mgoin
mgoin approved these changes on 2026-02-20
mgoin mgoin added ci-failure
mgoin mgoin enabled auto-merge (squash) 44 days ago
vllm-bot vllm-bot merged 0e22cd61 into main 44 days ago
vllm-bot vllm-bot deleted the revert-34471-fix-llama4-quant-loading branch 44 days ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone