llama.cpp
c830f99c
- server : support max_completion_tokens request property (#19831)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
6 days ago
server : support max_completion_tokens request property (#19831) "max_tokens" is deprectated in favor of "max_completion_tokens" which sets the upper bound for reasoning+output token. Closes: #13700
References
#19831 - server : support max_completion_tokens request property
Author
rgerganov
Parents
aa6f918c
Loading