vllm
0b53bec6
- [DOC]: Add warning about max_num_batched_tokens and max_model_len when chunked prefill is disabled (#33109)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
4 days ago
[DOC]: Add warning about max_num_batched_tokens and max_model_len when chunked prefill is disabled (#33109) Signed-off-by: Vincent Gimenes <147169146+VincentG1234@users.noreply.github.com>
References
#33109 - [DOC]: Add warning about max_num_batched_tokens and max_model_len when chunked prefill is disabled
Author
VincentG1234
Parents
c568581f
Loading