vllm
[Bugfix] Fix mismatch between global and local attention heads in tensor-parallel mode for param2moe model
#39707
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
11
Changes
View On
GitHub
[Bugfix] Fix mismatch between global and local attention heads in tensor-parallel mode for param2moe model
#39707
DarkLight1337
merged 11 commits into
vllm-project:main
from
Bharatgen-Tech:main
added files to support BharatGen's param2moe architecture
64d317d3
Update vllm/model_executor/models/param2moe.py
48bf0991
Updated Param2MoEForCausalLM class to maintain alphabetical order in …
a014075a
refactored weight loading into Param2MoEModel and adopted AutoWeights…
ebf14b16
Add Param2MoEForCausalLM to registry (tests/models/registry.py)
6c61d2c5
Fix formatting (ruff)
3c46652c
Merge branch 'vllm-project:main' into main
1acafe61
Merge branch 'vllm-project:main' into main
d2d41897
Refactor attention heads and normalization logic, patch for TP loading
05d55913
mergify
added
bug
gemini-code-assist
commented on 2026-04-13
Merge branch 'vllm-project:main' into main
d94bc190
DarkLight1337
added
ready
DarkLight1337
approved these changes on 2026-04-14
DarkLight1337
enabled auto-merge (squash)
25 days ago
Fix formatting (ruff)
a55d5322
disabled auto-merge
25 days ago
Head branch was pushed to by a user without write access
DarkLight1337
merged
f7e62e3d
into main
25 days ago
Login to write a write a comment.
Login via GitHub
Reviewers
DarkLight1337
gemini-code-assist
Assignees
No one assigned
Labels
bug
ready
Milestone
No milestone
Login to write a write a comment.
Login via GitHub