vllm
[MISC] Fix Tensor Parallelism for Quantized Mamba Models with n_groups=1
#33257
Merged

[MISC] Fix Tensor Parallelism for Quantized Mamba Models with n_groups=1 #33257

vadiklyutiy
vadiklyutiy fix tp>1 for quntized mamba models
88743733
vadiklyutiy vadiklyutiy requested a review from tdoublep tdoublep 37 days ago
vadiklyutiy vadiklyutiy requested a review from tlrmchlsmth tlrmchlsmth 37 days ago
vadiklyutiy vadiklyutiy assigned vadiklyutiy vadiklyutiy 37 days ago
gemini-code-assist
gemini-code-assist commented on 2026-01-28
vadiklyutiy
cursor
cursor commented on 2026-01-28
vadiklyutiy fix
b3878efe
tomeras91
tomeras91 commented on 2026-01-28
vadiklyutiy vadiklyutiy force pushed to b3878efe 32 days ago
vadiklyutiy vadiklyutiy force pushed to b3878efe 32 days ago
vadiklyutiy Unify MambaMixer2 TP sharding to use custom weight loader
d5d6d0b8
vadiklyutiy
vadiklyutiy vadiklyutiy added ready
vadiklyutiy
tomeras91
vadiklyutiy
mgoin mgoin added bug
tlrmchlsmth
tlrmchlsmth commented on 2026-02-03
tlrmchlsmth
tlrmchlsmth approved these changes on 2026-02-03
tlrmchlsmth tlrmchlsmth merged a372f3f4 into main 30 days ago

Login to write a write a comment.

Login via GitHub

Assignees
Labels
Milestone