llama.cpp
bcf75461 - server : add arg for disabling prompt caching (#18776)

Commit
3 days ago
server : add arg for disabling prompt caching (#18776) * server : add arg for disabling prompt caching Disabling prompt caching is useful for clients who are restricted to sending only OpenAI-compat requests and want deterministic responses. * address review comments * address review comments
Author
Parents
Loading