vllm
[MISC] Fix Tensor Parallelism for Quantized Mamba Models with n_groups=1
#33257
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
3
Changes
View On
GitHub
[MISC] Fix Tensor Parallelism for Quantized Mamba Models with n_groups=1
#33257
tlrmchlsmth
merged 3 commits into
vllm-project:main
from
CentML:vadim/fix-falcon-fp8-tp
fix tp>1 for quntized mamba models
88743733
vadiklyutiy
requested a review
from
tdoublep
37 days ago
vadiklyutiy
requested a review
from
tlrmchlsmth
37 days ago
vadiklyutiy
assigned
vadiklyutiy
37 days ago
gemini-code-assist
commented on 2026-01-28
cursor
commented on 2026-01-28
fix
b3878efe
tomeras91
commented on 2026-01-28
vadiklyutiy
force pushed
to
b3878efe
32 days ago
vadiklyutiy
force pushed
to
b3878efe
32 days ago
Unify MambaMixer2 TP sharding to use custom weight loader
d5d6d0b8
vadiklyutiy
added
ready
mgoin
added
bug
tlrmchlsmth
commented on 2026-02-03
tlrmchlsmth
approved these changes on 2026-02-03
tlrmchlsmth
merged
a372f3f4
into main
30 days ago
Login to write a write a comment.
Login via GitHub
Reviewers
tlrmchlsmth
cursor
tomeras91
gemini-code-assist
tdoublep
Assignees
vadiklyutiy
Labels
bug
ready
Milestone
No milestone
Login to write a write a comment.
Login via GitHub