vllm
[Bugfix] Fix mismatch between global and local attention heads in tensor-parallel mode for param2moe model
#39707
Merged

[Bugfix] Fix mismatch between global and local attention heads in tensor-parallel mode for param2moe model #39707

bhargav-patel-29
bhargav-patel-29 added files to support BharatGen's param2moe architecture
64d317d3
bhargav-patel-29 Update vllm/model_executor/models/param2moe.py
48bf0991
bhargav-patel-29 Updated Param2MoEForCausalLM class to maintain alphabetical order in …
a014075a
bhargav-patel-29 refactored weight loading into Param2MoEModel and adopted AutoWeights…
ebf14b16
bhargav-patel-29 Add Param2MoEForCausalLM to registry (tests/models/registry.py)
6c61d2c5
bhargav-patel-29 Fix formatting (ruff)
3c46652c
bhargav-patel-29 Merge branch 'vllm-project:main' into main
1acafe61
bhargav-patel-29 Merge branch 'vllm-project:main' into main
d2d41897
bhargav-patel-29 Refactor attention heads and normalization logic, patch for TP loading
05d55913
mergify mergify added bug
gemini-code-assist
gemini-code-assist commented on 2026-04-13
bhargav-patel-29 Merge branch 'vllm-project:main' into main
d94bc190
bhargav-patel-29
DarkLight1337 DarkLight1337 added ready
DarkLight1337
DarkLight1337 approved these changes on 2026-04-14
DarkLight1337 DarkLight1337 enabled auto-merge (squash) 25 days ago
mergify
bhargav-patel-29 Fix formatting (ruff)
a55d5322
disabled auto-merge 25 days ago
Head branch was pushed to by a user without write access
bhargav-patel-29
DarkLight1337 DarkLight1337 merged f7e62e3d into main 25 days ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone