vllm
[Benchmark] Add support for multiple batch size benchmark through CLI in `benchmark_moe.py` + Add Triton Fused MoE kernel config for FP8 E=16 on B200
#20516
Merged

[Benchmark] Add support for multiple batch size benchmark through CLI in `benchmark_moe.py` + Add Triton Fused MoE kernel config for FP8 E=16 on B200 #20516

DarkLight1337 merged 2 commits into vllm-project:main from fix/benchmark
b8zhong
github-actions
b8zhong fix
b9345566
b8zhong b8zhong force pushed to b9345566 213 days ago
mergify mergify added performance
gemini-code-assist
gemini-code-assist commented on 2025-07-05
gemini-code-assist
gemini-code-assist commented on 2025-07-05
b8zhong add config file
3a5a0847
DarkLight1337
DarkLight1337 approved these changes on 2025-07-06
DarkLight1337 DarkLight1337 enabled auto-merge (squash) 213 days ago
github-actions github-actions added ready
DarkLight1337 DarkLight1337 merged cede942b into main 213 days ago
b8zhong b8zhong changed the title [Benchmark] Add support for multiple batch size benchmark through CLI in `benchmark_moe.py` [Benchmark] Add support for multiple batch size benchmark through CLI in `benchmark_moe.py` + Add Triton Fused MoE kernel config for FP8 E=16 on B200 213 days ago
b8zhong b8zhong deleted the fix/benchmark branch 213 days ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone