Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
vllm-project/vllm
Pull Requests
Commits
tms/distributed_timeout
7snzwi-codex/change-default-logging-behavior
acc-rate
aiter-fp8-mk
amd_dev
amd_mori
amd-ci
andy-neuma-testing
batched_triton_fallback
bench-latency
benchmark_serving_test
bind_kv_caches
build-flashinfer-aot-wheel
codex/add-auto-max-model-length-setting
codex/add-pandas-and-datasets-to-requirements
codex/change-default-logging-behavior
codex/remove-raydistributedexecutor-from-v0-engine
codex/remove-virtual-engine-from-codebase
codex/remove-vllm-v0-engine-references-from-docs
codex/update-arch-overview-md-with-vllm-v1-details
copilot/fix-31e676e9-a4af-4ed2-b74d-19d27f0a57b2
copilot/fix-584be906-f283-4e17-8776-c14111357ee7
copilot/fix-56244f30-e76a-41ed-beaf-3bc9de22a2c9
copilot/fix-870996da-9146-438e-9a52-cdc6c1743086
copilot/fix-c6914add-1b66-46d0-9948-c2e7b6f2259f
copilot/fix-cudagraph-flag-combination
correct-docs-cuda-version
dbo-cudagraph-size-cherry
deep_full_cudagraph_fix
deepep_tweaks
deepseek_optimizations_alex_rob
dependabot/github_actions/actions/checkout-5.0.0
disable-sd
dockerfile-nvcc-compress
fix_ds_eagle
fix_use_ep
fix-doc-build
fix-hashing-partial-blocks
fix-precommit
fp8_ep_dp
full_cudagraph
gemma3n-mm
ghsa-mcmc-2m55-j8jj
gpu_ids2
gpu-ids
il_tool
jax-tpu
kevin_h100
khluu/clean_apt
khluu/nccl
khluu/test_fixed_premerge
khluu/test_latest_feat
khluu/test_pull_through_cache
khluu/test_rebase
khluu/test_us_east_1
khluu/test
khluu/try_moc
khluu/use_ccache_premerge
khluu/0.11.1
khluu/8gpu_h200
khluu-patch-1
low_latency_opt
lwilkinson/cg-support
lwilkinson/dbo-full-cudagraphs
lwilkinson/eagle-piecewise
lwilkinson/potential-cutlass-mla-fix
lwilkinson/refactor-cmake
main
mamba_tests
marlin_gptoss_swiglu
maybe_fix_hang_2
mergify/houseroad/config-update
minus_x
mla_cuda_graphs
mla_decode_any_head
mla-support-awq-marlin
moe-refactor-modelopt-fp8
moondream2
optimize-prefix-caching-scheduling
pd_scheduling
pil_image
qwen25vl
rebased_fi_moe
reduce_scatter_comm
refactor-modelopt-fp8-modular-kernel
releases/v0.9.0
releases/v0.9.1
releases/v0.9.2
releases/v0.10.0
releases/v0.10.1
releases/v0.10.2
releases/v0.11.0
releases/v0.11.1
releases/v0.11.2
releases/v0.12.0
releases/v0.13.0
remove_mamba_ssm
revert-21550-chengji/fix-ci
revert-22299-main
revert-26740-wentao-optimize-startup-log-2
revert-27532-lwilkinson/upconvert-all-2
revert-27600-torch-utils-import
revert-29385-eplb_nightly_ci
running-deque
seemethere/cuda_arm64
simon-mo-patch-1
skip-lmfe-tests
split_kv_cache_init
support_global_dp_logging
test-debug-lb
test-docker-cache
tms/distributed_timeout
topk_id_hack
torch_dynamo
tpu_v1_optimized
tpu_v1
update_from_kv_xfer_finished_race_fix
use-uv-python-for-docker
v0.8.0
v0.8.1
v0.8.2
v0.8.3
v0.8.4
v0.8.5
v1-sched-interface-2
v1_fix_profiler
verbose-prime-rl-ci
wentao-fix-python-install-ci-error
wentao-fix-qwen3vl-launch-bug
wentao-fix-torch-compile-issue
wentao-update-torch-to-2.9.1
whisper-translate
wide_ep_working_branch
wide_ep_working_branch_2
woosuk/fa3-swa-cudagraph
woosuk/flashinfer-swa
woosuk/remove-req-idx-mapping
woosuk/rm-add-init-env
woosuk/router-nixl
woosuk/sampled-token-ids
woosuk/test-router
woosuk/v2-logit-bias
woosuk/v2-penalties
woosuk-jf
wye-refactor-w8a8-quant
zhuohan/moe-kernel-experiment
zhuohan/remove-redundant-argument
zhuohan/remove-virtual-engine
zhuohan/revert-26709
Fix precommit
tlrmchlsmth
committed
156 days ago
2f86f710
Add VLLM_DISTRIBUTED_INIT_TIMEOUT_SECONDS
tlrmchlsmth
committed
156 days ago
feeb1730
[Kernel] Apply torch.Tag.needs_fixed_stride_order only for torch==2.6.0 (#19346)
zou3519
committed
156 days ago
Verified
b2eb2b5a
[CI] Update CODEOWNERS for vllm/compilation (#21185)
zou3519
committed
156 days ago
Verified
21274ab4
Let GraniteMoeAttention use YaRN (#21174)
tdoublep
committed
156 days ago
Verified
ed8cbfed
[Core] Set pooling params based on task and model (#21128)
DarkLight1337
committed
156 days ago
Verified
45badd05
[Bugfix] Allocate less memory in non-batched CUTLASS MoE (#21121)
ElizaWszola
committed
156 days ago
Verified
4adc66f6
[Doc] Fix typo in model name (#21178)
DarkLight1337
committed
156 days ago
Verified
55ad6487
[Bugfix] The special_tokens in tokenizer should also be controlled by do_lower_case in encoder_config. (#20750)
noooop
committed
156 days ago
Verified
5895afd7
[Model] Re-add the implicit conversion feature for as_seq_cls_model (#21103)
noooop
committed
156 days ago
Verified
ca4eb82b
[Misc] Make MM embedding merge interface explicit in model runner (#21147)
Roger Wang
committed
156 days ago
Verified
ba2dfbb0
[benchmark] Sending request strictly follows the random intervals (#21108)
Jialin
committed
156 days ago
Verified
1bf65138
[Misc] Do not print async output warning for v1 (#21151)
WoosukKwon
committed
156 days ago
Verified
54cf1cae
[Perf] Add swap_ab to SM90 FP8 non-block CUTLASS moe grouped gemm (#20911)
shixianc
committed
156 days ago
Verified
5780121c
[Core] FlashInfer CUTLASS fused MoE backend (NVFP4) (#20037)
wenscarl
committed
156 days ago
Verified
c7d8724e
[Doc] Add inplace weights loading example (#19640)
22quinn
committed
156 days ago
Verified
b38baabc
[Attention] Make local attention backend agnostic (#21093)
LucasWilkinson
committed
156 days ago
Verified
89cab4d0
[Docs] Update supported models documentation with missing models (#20844)
luccafong
committed
156 days ago
Verified
b9a21e91
[Docs] Add minimal demo of Ray Data API usage (#21080)
crypdick
committed
156 days ago
Verified
c4e3b125
[Bugfix] Fix the tensor non-contiguous issue for Flashinfer TRT-LLM backend attention kernel (#21133)
elvischenv
committed
156 days ago
Verified
8dfb45ca
[Log] Debugging Log with more Information (#20770)
yewentao256
committed
156 days ago
Verified
8a8fc946
[V0 deprecation] Remove V0 HPU backend (#21131)
WoosukKwon
committed
156 days ago
Verified
4de71463
On environments where numa cannot be detected we get 0 (#21115)
ericcurtin
committed
157 days ago
Verified
ac9fb732
[Misc] Qwen MoE model supports LoRA (#20932)
jeejeelee
committed
157 days ago
Verified
a3a6c695
[Model] Update pooling model interface (#21058)
DarkLight1337
committed
157 days ago
Verified
90bd2ab6
[Performance] Performance improvements in non-blockwise fp8 CUTLASS MoE (#20762)
ElizaWszola
committed
157 days ago
Verified
9fb2d220
[Docs] Move code block out of admonition now that it's short (#21118)
hmellor
committed
157 days ago
Verified
2d6a3820
[Misc] Avoid unnecessary import (#21106)
wangxiyuan
committed
157 days ago
Verified
89e3c4e9
[Docs] Improve docstring formatting for `FusedMoEParallelConfig.make` (#21117)
hmellor
committed
157 days ago
Verified
fe8a2c54
[VLM] Add Nemotron-Nano-VL-8B-V1 support (#20349)
kylehh
committed
157 days ago
Verified
4ef00b5c
Older