Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
vllm-project/vllm
Pull Requests
Commits
seemethere/cuda_arm64
7snzwi-codex/change-default-logging-behavior
acc-rate
amd_dev
amd_mori
amd-ci
andy-neuma-testing
apply-refactor-to-ct
batched_triton_fallback
bench-latency
benchmark_serving_test
bind_kv_caches
build-flashinfer-aot-wheel
codex/add-auto-max-model-length-setting
codex/add-pandas-and-datasets-to-requirements
codex/change-default-logging-behavior
codex/remove-raydistributedexecutor-from-v0-engine
codex/remove-virtual-engine-from-codebase
codex/remove-vllm-v0-engine-references-from-docs
codex/update-arch-overview-md-with-vllm-v1-details
copilot/fix-31e676e9-a4af-4ed2-b74d-19d27f0a57b2
copilot/fix-584be906-f283-4e17-8776-c14111357ee7
copilot/fix-56244f30-e76a-41ed-beaf-3bc9de22a2c9
copilot/fix-870996da-9146-438e-9a52-cdc6c1743086
copilot/fix-c6914add-1b66-46d0-9948-c2e7b6f2259f
copilot/fix-cudagraph-flag-combination
correct-docs-cuda-version
dbo-cudagraph-size-cherry
deep_full_cudagraph_fix
deepep_tweaks
deepseek_optimizations_alex_rob
dependabot/github_actions/actions/checkout-5.0.0
disable-sd
dockerfile-nvcc-compress
eplb_policy_log_fix
fix_ds_eagle
fix_use_ep
fix-aiter-mixtral
fix-doc-build
fix-flashinfer-experts-quant-config-hack
fix-hashing-partial-blocks
fix-precommit
fp8_ep_dp
full_cudagraph
gemma3n-mm
ghsa-mcmc-2m55-j8jj
gpu_ids2
gpu-ids
il_tool
jax-tpu
kevin_h100
khluu/clean_apt
khluu/nccl
khluu/sync_ci_1230
khluu/test_fixed_premerge
khluu/test_latest_feat
khluu/test_pull_through_cache
khluu/test_rebase
khluu/test_us_east_1
khluu/test
khluu/try_moc
khluu/use_ccache_premerge
khluu/0.11.1
khluu/8gpu_h200
khluu-patch-1
low_latency_opt
lwilkinson/cg-support
lwilkinson/dbo-full-cudagraphs
lwilkinson/eagle-piecewise
lwilkinson/potential-cutlass-mla-fix
lwilkinson/refactor-cmake
main
mamba_tests
marlin_gptoss_swiglu
maybe_fix_hang_2
mergify/houseroad/config-update
minus_x
mk-init-refactor-poc
mla_cuda_graphs
mla_decode_any_head
mla-support-awq-marlin
moondream2
optimize-prefix-caching-scheduling
pd_scheduling
pil_image
qwen25vl
rebased_fi_moe
reduce_scatter_comm
refactor-modelopt-fp8-modular-kernel
releases/v0.9.0
releases/v0.9.1
releases/v0.9.2
releases/v0.10.0
releases/v0.10.1
releases/v0.10.2
releases/v0.11.0
releases/v0.11.1
releases/v0.11.2
releases/v0.12.0
releases/v0.13.0
remove_mamba_ssm
revert-21550-chengji/fix-ci
revert-22299-main
revert-26740-wentao-optimize-startup-log-2
revert-27532-lwilkinson/upconvert-all-2
revert-27600-torch-utils-import
revert-29385-eplb_nightly_ci
running-deque
seemethere/cuda_arm64
simon-mo-patch-1
skip-lmfe-tests
split_kv_cache_init
support_global_dp_logging
test-debug-lb
test-docker-cache
tms/distributed_timeout
topk_id_hack
torch_dynamo
tpu_v1_optimized
tpu_v1
update_from_kv_xfer_finished_race_fix
use-uv-python-for-docker
v0.8.0
v0.8.1
v0.8.2
v0.8.3
v0.8.4
v0.8.5
v1-sched-interface-2
v1_fix_profiler
verbose-prime-rl-ci
wentao-fix-python-install-ci-error
wentao-fix-qwen3vl-launch-bug
wentao-fix-torch-compile-issue
wentao-revert-torch-warning
wentao-update-torch-to-2.9.1
whisper-translate
wide_ep_working_branch
wide_ep_working_branch_2
woosuk/fa3-swa-cudagraph
woosuk/flashinfer-swa
woosuk/remove-req-idx-mapping
woosuk/rm-add-init-env
woosuk/router-nixl
woosuk/sampled-token-ids
woosuk/test-router
woosuk/v2-logit-bias
woosuk/v2-penalties
woosuk-jf
wye-refactor-w8a8-quant
zhuohan/moe-kernel-experiment
zhuohan/remove-redundant-argument
zhuohan/remove-virtual-engine
zhuohan/revert-26709
Merge branch 'main' into seemethere/cuda_arm64
mgoin
committed
143 days ago
5667ed87
[ROCm][Bugfix] Fix compilation error in topk softmax fused kernel (#22819)
kliuae
committed
143 days ago
Verified
c6cd5ca3
[CI/Build] Skip gpt_big model test because of broken HF model (#22848)
Isotr0py
committed
143 days ago
Verified
df0e0f02
[CI/Build] Fix param mismatch in `test_eagle_correctness` (#22847)
DarkLight1337
committed
143 days ago
Verified
b4b78d63
[CI] Fix `tests/v1/e2e/test_kv_sharing_fast_prefill.py` import on test (#22815)
NickLucche
committed
143 days ago
Verified
12817a8a
Update torch_cuda_arch_list
seemethere
committed
143 days ago
846aa6dc
ci: Add CUDA + arm64 relase builds
seemethere
committed
143 days ago
f8b2e006
[CI/Build] Update VLM common tests (#22841)
DarkLight1337
committed
143 days ago
Verified
c9232d41
[Bugfix] Fix MiniCPMV Image input inference failed (#22813)
jio-H
committed
143 days ago
Verified
9bd9294f
[Misc] clear and separate error messages for input too long and input + max-tokens too long (#22803)
Roger Wang
committed
143 days ago
Verified
da270519
[Core] Use individual MM items in P0/P1 cache and model runner (#22570)
DarkLight1337
committed
143 days ago
Verified
19b927e5
[Frontend] Multithreaded async multimodal load_bytes (#22710)
milesial
committed
143 days ago
Verified
20d65aa7
Fix GGUF loader for Qwen3 MoE. (#22785)
Gh0u1L5
committed
143 days ago
Verified
b159c0a6
Remove unnecessary CUDA sync of qwen image and video preprocess (#22792)
cyyever
committed
143 days ago
Verified
6772bb0f
[Bugfix][mamba] Fix type annotation of Mamba2Metadata (#22787)
heheda12345
committed
143 days ago
Verified
fceafaf5
[Nixl][CI] Fix tests (#22806)
NickLucche
committed
143 days ago
Verified
6b794c75
[FEATURE] support custom vllm tuned config path for fused moe triton kernels (#22791)
vermouth1992
committed
143 days ago
Verified
98deac38
[Frontend] Add chunked processing to handle long inputs in embedding models (#22280)
x22x22
committed
143 days ago
Verified
653124bd
[Platform] Custom ops support for FusedMoe (#22509)
wangxiyuan
committed
143 days ago
Verified
0b1bdac6
[V1] Add tree drafting tests for eagle spec decoding (#22705)
TheEpicDolphin
committed
143 days ago
Verified
d94e3026
[Doc] Add max_lora_rank configuration guide (#22782)
chi2liu
committed
143 days ago
Verified
3f52738d
[Bugfix] Fix Nemotron VL image processing (#22739)
ducviet00
committed
143 days ago
Verified
a01e0018
[Model] Add missing prefix to glm4_1v (#22716)
zRzRzRzRzRzRzR
committed
143 days ago
Verified
9e7e5baa
[Model] Add option to run Step3VisionEncoder in DP (#22697)
zzh142857
committed
143 days ago
Verified
d16aa3da
[gpt-oss] upgrade gpt-oss to v0.0.3 and add version check (#22768)
heheda12345
committed
144 days ago
Verified
6807af8f
[Perf] Support topk softmax fused kernel for broader num_experts (#22211)
shixianc
committed
144 days ago
Verified
4c558cf6
[Bug] Fix Unexpected Keyword Argument 'w1_bias' (#22757)
yewentao256
committed
144 days ago
Verified
77a6bf07
Remove unneeded ROCm platform import when using CUDA (#22765)
mgoin
committed
144 days ago
Verified
4082338a
Force TRTLLM attention for gpt-oss on SM100 (#22678)
mgoin
committed
144 days ago
Verified
c6b92879
[Bugfix] Fix default enable for CUTLASS MLA on SM100 (#22738)
mgoin
committed
144 days ago
Verified
b1361c72
Older