vllm
0b53bec6 - [DOC]: Add warning about max_num_batched_tokens and max_model_len when chunked prefill is disabled (#33109)

Commit
4 days ago
[DOC]: Add warning about max_num_batched_tokens and max_model_len when chunked prefill is disabled (#33109) Signed-off-by: Vincent Gimenes <147169146+VincentG1234@users.noreply.github.com>
Author
Parents
Loading