Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
huggingface/text-generation-inference
Pull Requests
Commits
feat-backend-llamacpp
20250708-ci-fixes
add_L4
add_api_key
add_batch_dimension
add_chunked_atn
add_chunked_attn
add_deepseekv3
add_gptq_docs
add_integration_test
add_readme_dashboard
add_tunable_prefill
add_vlm_chunking
add-chat-response-format
add-google-cloud-provider
add-quickstart-script
add-rotary-embed-tests
add-small-ttft-script
add-test-for-warmup-and-kvcache
adding_docs
adjust-mllama-test-output
adjust-where-request-max-tokens-is-defaulted
aiter_kernels
amd-ci-fx
auto_length
automodel-supports-flash-paged-attention
avoid-cuda-graph-during-warmup-if-oom
avoid-zero-seed
backends/trtllm
backends/trtllm-executor
baichuan2-13b
bnb4
bugfix/add_tools_prompt
bugfix/moe-kernels-imports
bugfix/phi-exl2
bump-client-0.6.2
bump-kernel-versions
bump-poetry-and-requirements
chunked_attn_l4
ci_amd
ci_amd2
ci_amd3
ci_amd4
ci2
ci-amihalik-update-chat-completion-messages
ci-new-cluster
ci-patch
ci-run-openai-function-calling-compatible-support
ci-update_xpu_image
ci-xpu
ci-xpu2
close_dl_thread
compat_logger
compile-grammar-in-router
cuda_ipc_allreduce
debug/gemma2
debug-gpt2
debug-request-id
debug-torch-23
debugging-timeouts
deploy/aml
dev
development-guide
dummy
enable_non_divisible_embeddings
enable-non-grammar-constrained-tools
enable-qwen2vl-video
enable-transformers-vlm
exl2
experiment/moe
explore-static-triton-kernels
explore-t4-gemma-issues
feat/add-load-test
feat/attention_sinks
feat/backend_abstraction
feat/backend_feature
feat/better_tokens
feat/cuda_12
feat/flash_decoding
feat/improve_max_tokens
feat/max_queue_size
feat/page_re_alloc
feat/parse_logs
feat/support_deepspeed
feat-backend-llamacpp
feature/machete
feature/moe-kernels
feature/no_repeat_ngram_size_ci
feature/no_repeat_ngram_size
feature/phi-3-small
feature/prefix
feature/radix-prefix-cache
feature/radix-prefix-cache-bench
feature/vlm-prefix-caching
fix/allow-top-p-0
fix/avoid_record_streams
fix_default_arg
fix_exl2
fix_fp8_llama3.2
fix_leak
fix_mistral2
fix_neox_rotary_emb
fix/op-trace-id
fix/parse-mamba-config
fix_phi3
fix-cudagraph-bug
fix-gemma-tokenization
fix-grammar-cleanup-bug
fix-grammar-fsm-batching
fix-mixtral-adapter-loading
fix-release-tests
fix-repack-for-marlin
fix-tool-call-def
fix-tp
fix-version-install
flashinfer
flashinfer-0.2.5
fp8_kvcache
fp8_rocm
gaudi_llama4_tmp
git_v2.1.0
git_v2.1.1
git_v2.2.0
git_v2.3.0
git_v2.3.1
git_v2.4.0
git_v2.4.1
git_v3.0.0
git_v3.0.1
git_v3.0.2
git_v3.1.0
git_v3.2.2
git_v3.2.3
git_v3.3.3
git_v3.3.4
git_v3.3.5
git_2.0.4
git_3.1.1
git_3.2.0
git_3.2.1
git_3.3.0
git_3.3.1
git_3.3.2
improve_defaults
improve_launcher_defaults
improve-docs
improve-dynamic-message-content
improve-json-schema-field
improve-tool-call-and-response-ids
inlcude-latest-release-on-commit-builds-tags
ipex-moe
kvrouter
kvrouter-endpoints
llama-fused-compiled-mlp
main
maintenance/docker-network
maintenance/merge-vlm-input-prep
mamba2
martinigoyanes-fix-frequency-penalty
medusa
megatron
message-more-info
mi300-temp
mllama
model_compat_log
more_logs
multi-lora
new_minor_version
nix/cargo-clippy
nix/docker2
nix_integration_tests
nix/pytorch-2.5.1
nix_test2
no_root_user
no_root_user2
op-compilation-benchmarking
origin/slind_window_fix
osanseviero-patch-1
pip-installable
pr-1869-ci-run
pr-2076-ci-run
pr-2290-ci-runner
pr-2366-ci-branch
pr-2444-ci-branch
pr-2517-ci-branch
pr-2711-ci-branch
pr-2784-ci-branch
pr-2840-ci-branch
pr-2954-ci-branch
pr-3002-ci-branch
pr-3004-ci-branch
pr-3018-ci-branch
precompile-kernels-workflow
prefix_chunk
prefix_default
proxy_sse_engine_state
quantization
quantization-0.1
refactor-lora-linear
release-3.2.4
remove_post_load_weights
response-header-metrics
revert
rocm_6.2_fixes
rocm-ci-build
router-grammar-compile
s3-cache
self-generating-docs
set-num-blocks
simpler_exllama
skip-mistral-test
speculative
streaming_conceptual
support-granite-vision
support-logit-bias-in-chat
support-phi3-small
support-phi-model
support-pre-compile-kernels
temp_work
test_docs
test_rocm
test-batch-speedup-amount
tmp_invariants
tmp_medusa
tmp_torch_compile
transformers-ci
triton_fix
trtllm/executor_stats
trtllm-stop-words
tuna
update_docs2
update_internal_version
update_peft
update_readme
update-flake-deps-and-logit-processor
update-jsonschema
upgrade_mlp_speculator
upgrade-outlines
use_g6
use_updated_kernels
vllm/setup
zstd
misc: use return Ok(())
mfuntowicz
committed
1 year ago
Verified
182ffaf0
feat(backend): use c++ defined types for llama.cpp
mfuntowicz
committed
1 year ago
e0dda9b6
feat(backend): better map exception throw on C++ side
mfuntowicz
committed
1 year ago
c9f6c3a8
feat(backend): add mimalloc memory allocator to the container
mfuntowicz
committed
1 year ago
db41776a
feat(backend): correctly link to all libraries
mfuntowicz
committed
1 year ago
f5c4cee3
feat: Fix Cmakelist to allow building on Darwin platform (#2785)
Hugoch
committed
1 year ago
Verified
59b0ef30
feat(backend): use new batch API to generate tokens
mfuntowicz
committed
1 year ago
b10eaab9
feat(backend): create llama_context_params with default factory
mfuntowicz
committed
1 year ago
dc6435e3
feat(backend): update llama.cpp to 4215
mfuntowicz
committed
1 year ago
b1ebc8f7
misc(offline): update model creation as std::shared_ptr
mfuntowicz
committed
1 year ago
6c5a75b5
feat(backend): add missing temperature parameter
mfuntowicz
committed
1 year ago
9d659f1e
feat(backend): add guard in case top_k = 0
mfuntowicz
committed
1 year ago
df72c56b
feat(backend): add some test to the backend for core allocation
mfuntowicz
committed
1 year ago
929a2fc7
feat(backend): fix when num_cores_per_instance is equals to zero with the size of the generated core allocation
mfuntowicz
committed
1 year ago
298367cd
feat(backend): use the new batch api from llama
mfuntowicz
committed
1 year ago
8e897935
feat(backend): remove core overriding in the Rust backend
mfuntowicz
committed
1 year ago
274cfce4
Update Dockerfile.llamacpp as per review
mfuntowicz
committed
1 year ago
Verified
d918e6a1
Update Dockerfile.llamacpp as per review
mfuntowicz
committed
1 year ago
Verified
bbe95ca9
chore: remove unrelated change to trtllm
mfuntowicz
committed
1 year ago
9025a26c
misc(doc): rust documentation
mfuntowicz
committed
1 year ago
862a519f
misc(doc): c++ documentation
mfuntowicz
committed
1 year ago
b9c04b9c
misc(license): update LICENSE
mfuntowicz
committed
1 year ago
4ee2ee58
misc(backend): allow rebinding numa core affinity
mfuntowicz
committed
1 year ago
2d9465d1
misc(docker): add numa lib as dependency
mfuntowicz
committed
1 year ago
30ae9963
feat(backend): rely on multi consumer queue to scheduler workers
mfuntowicz
committed
1 year ago
5a856616
feat(backend): correctly setup llama_context providing n_threads and n_ubatch
mfuntowicz
committed
1 year ago
84eead21
feat(backend): bind thread and memory affinity for thread
mfuntowicz
committed
1 year ago
50c37661
feat(backend): multistream inference on CPU
mfuntowicz
committed
1 year ago
5335bf97
misc(build): improve build process
mfuntowicz
committed
1 year ago
23d2bcf2
feat(backend): update llamacpp to 4077
mfuntowicz
committed
1 year ago
70c90ad9
Older