Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
huggingface/text-generation-inference
Pull Requests
Commits
feat/better_tokens
20250708-ci-fixes
add_L4
add_api_key
add_batch_dimension
add_chunked_atn
add_chunked_attn
add_deepseekv3
add_gptq_docs
add_integration_test
add_readme_dashboard
add_tunable_prefill
add_vlm_chunking
add-chat-response-format
add-google-cloud-provider
add-quickstart-script
add-rotary-embed-tests
add-small-ttft-script
add-test-for-warmup-and-kvcache
adding_docs
adjust-mllama-test-output
adjust-where-request-max-tokens-is-defaulted
aiter_kernels
amd-ci-fx
auto_length
automodel-supports-flash-paged-attention
avoid-cuda-graph-during-warmup-if-oom
avoid-zero-seed
backends/trtllm
backends/trtllm-executor
baichuan2-13b
bnb4
bugfix/add_tools_prompt
bugfix/moe-kernels-imports
bugfix/phi-exl2
bump-client-0.6.2
bump-kernel-versions
bump-poetry-and-requirements
chunked_attn_l4
ci_amd
ci_amd2
ci_amd3
ci_amd4
ci2
ci-amihalik-update-chat-completion-messages
ci-new-cluster
ci-patch
ci-run-openai-function-calling-compatible-support
ci-update_xpu_image
ci-xpu
ci-xpu2
close_dl_thread
compat_logger
compile-grammar-in-router
cuda_ipc_allreduce
debug/gemma2
debug-gpt2
debug-request-id
debug-torch-23
debugging-timeouts
deploy/aml
dev
development-guide
dummy
enable_non_divisible_embeddings
enable-non-grammar-constrained-tools
enable-qwen2vl-video
enable-transformers-vlm
exl2
experiment/moe
explore-static-triton-kernels
explore-t4-gemma-issues
feat/add-load-test
feat/attention_sinks
feat/backend_abstraction
feat/backend_feature
feat/better_tokens
feat/cuda_12
feat/flash_decoding
feat/improve_max_tokens
feat/max_queue_size
feat/page_re_alloc
feat/parse_logs
feat/support_deepspeed
feat-backend-llamacpp
feature/machete
feature/moe-kernels
feature/no_repeat_ngram_size_ci
feature/no_repeat_ngram_size
feature/phi-3-small
feature/prefix
feature/radix-prefix-cache
feature/radix-prefix-cache-bench
feature/vlm-prefix-caching
fix/allow-top-p-0
fix/avoid_record_streams
fix_default_arg
fix_exl2
fix_fp8_llama3.2
fix_leak
fix_mistral2
fix_neox_rotary_emb
fix/op-trace-id
fix/parse-mamba-config
fix_phi3
fix-cudagraph-bug
fix-gemma-tokenization
fix-grammar-cleanup-bug
fix-grammar-fsm-batching
fix-mixtral-adapter-loading
fix-release-tests
fix-repack-for-marlin
fix-tool-call-def
fix-tp
fix-version-install
flashinfer
flashinfer-0.2.5
fp8_kvcache
fp8_rocm
gaudi_llama4_tmp
git_v2.1.0
git_v2.1.1
git_v2.2.0
git_v2.3.0
git_v2.3.1
git_v2.4.0
git_v2.4.1
git_v3.0.0
git_v3.0.1
git_v3.0.2
git_v3.1.0
git_v3.2.2
git_v3.2.3
git_v3.3.3
git_v3.3.4
git_v3.3.5
git_2.0.4
git_3.1.1
git_3.2.0
git_3.2.1
git_3.3.0
git_3.3.1
git_3.3.2
improve_defaults
improve_launcher_defaults
improve-docs
improve-dynamic-message-content
improve-json-schema-field
improve-tool-call-and-response-ids
inlcude-latest-release-on-commit-builds-tags
ipex-moe
kvrouter
kvrouter-endpoints
llama-fused-compiled-mlp
main
maintenance/docker-network
maintenance/merge-vlm-input-prep
mamba2
martinigoyanes-fix-frequency-penalty
medusa
megatron
message-more-info
mi300-temp
mllama
model_compat_log
more_logs
multi-lora
new_minor_version
nix/cargo-clippy
nix/docker2
nix_integration_tests
nix/pytorch-2.5.1
nix_test2
no_root_user
no_root_user2
op-compilation-benchmarking
origin/slind_window_fix
osanseviero-patch-1
pip-installable
pr-1869-ci-run
pr-2076-ci-run
pr-2290-ci-runner
pr-2366-ci-branch
pr-2444-ci-branch
pr-2517-ci-branch
pr-2711-ci-branch
pr-2784-ci-branch
pr-2840-ci-branch
pr-2954-ci-branch
pr-3002-ci-branch
pr-3004-ci-branch
pr-3018-ci-branch
precompile-kernels-workflow
prefix_chunk
prefix_default
proxy_sse_engine_state
quantization
quantization-0.1
refactor-lora-linear
release-3.2.4
remove_post_load_weights
response-header-metrics
revert
rocm_6.2_fixes
rocm-ci-build
router-grammar-compile
s3-cache
self-generating-docs
set-num-blocks
simpler_exllama
skip-mistral-test
speculative
streaming_conceptual
support-granite-vision
support-logit-bias-in-chat
support-phi3-small
support-phi-model
support-pre-compile-kernels
temp_work
test_docs
test_rocm
test-batch-speedup-amount
tmp_invariants
tmp_medusa
tmp_torch_compile
transformers-ci
triton_fix
trtllm/executor_stats
trtllm-stop-words
tuna
update_docs2
update_internal_version
update_peft
update_readme
update-flake-deps-and-logit-processor
update-jsonschema
upgrade_mlp_speculator
upgrade-outlines
use_g6
use_updated_kernels
vllm/setup
zstd
fmt
OlivierDehaene
committed
2 years ago
a4fd6905
feat(server): use encoding to get prefill tokens
OlivierDehaene
committed
2 years ago
83e442ca
fix(server): fix warpers on CPU (#472)
OlivierDehaene
committed
2 years ago
Verified
53aa9194
feat(server): improve flash attention import errors (#465)
OlivierDehaene
committed
2 years ago
Verified
ece7ffa4
feat(router): add ngrok integration (#453)
OlivierDehaene
committed
2 years ago
Verified
f59fb8b6
feat(server): pre-allocate past key values for flash causal LM (#412)
OlivierDehaene
committed
2 years ago
Verified
5ce89059
fix(makefile): Fix typo and use POSIX comparison in the makefile (#443)
piratos
committed
2 years ago
Verified
ca650e5b
docs(launcher): fix CUDA_VISIBLE_DEVICES helper comment (#441)
antferdom
committed
2 years ago
Verified
d4eb60f4
feat(server): optimize dist ops (#434)
OlivierDehaene
committed
2 years ago
Verified
e496c9ba
feat(server): Rework model loading (#344)
Narsil
committed
2 years ago
Verified
abd58ff8
chore: update openapi schema
OlivierDehaene
committed
2 years ago
19c41824
feat(server): batch tokenization for flash causal lm (#411)
OlivierDehaene
committed
2 years ago
Verified
6abec14a
feat(server): only compute prefill logprobs when asked (#406)
OlivierDehaene
committed
2 years ago
Verified
895c5f15
feat(launcher): parse oom signal (#404)
OlivierDehaene
committed
2 years ago
Verified
83b84486
feat(sagemaker): add trust remote code to entrypoint (#394)
OlivierDehaene
committed
2 years ago
Verified
62fc4010
v0.8.2
OlivierDehaene
committed
2 years ago
e7248fe9
feat(server): load santacoder/starcoder models with safetensors (#393)
OlivierDehaene
committed
2 years ago
Verified
95d35469
feat(server): remove trust_remote_code requirement for falcon models (#396)
OlivierDehaene
committed
2 years ago
Verified
c0928e6f
fix(server): fix has_position_ids (#395)
OlivierDehaene
committed
2 years ago
Verified
d69a0633
v0.8.1
OlivierDehaene
committed
2 years ago
db2ebe39
fix(server): fix bnb quantization for CausalLM models (#385)
OlivierDehaene
committed
2 years ago
Verified
337afb28
feat(server): add retry on download (#384)
OlivierDehaene
committed
2 years ago
Verified
87dc034b
increase health checks
OlivierDehaene
committed
2 years ago
444400b4
v0.8.0
OlivierDehaene
committed
2 years ago
081b9265
feat(server): support RefinedWeb models (#379)
OlivierDehaene
committed
2 years ago
Verified
b8b950b3
fix(server): fix quantization
OlivierDehaene
committed
2 years ago
bf7f1d54
fix(launcher): parse num cuda devices from CUDA_VISIBLE_DEVICES and NVIDIA_VISIBLE_DEVICES
OlivierDehaene
committed
2 years ago
49a6c8c1
fix(launcher): parse num cuda devices from CUDA_VISIBLE_DEVICES and NVIDIA_VISIBLE_DEVICES
OlivierDehaene
committed
2 years ago
146e72c3
Fix issue when load AutoModelForSeq2SeqLM model (#370)
CL-Shang
committed
2 years ago
Verified
5fde8d99
feat(server): support vectorized warpers in flash causal lm (#317)
OlivierDehaene
committed
2 years ago
Verified
62f91f78
Older