llama.cpp
2b6b55a5 - server : include usage statistics only when user request them (#16052)

Commit
4 days ago
server : include usage statistics only when user request them (#16052) * server : include usage statistics only when user request them When serving the OpenAI compatible API, we should check if {"stream_options": {"include_usage": true} is set in the request when deciding whether we should send usage statistics closes: #16048 * add unit test
Author
Parents
Loading