llama.cpp
0a019ed8
- server: add --models-memory-max parameter to allow dynamically unloading models when they exceed a memory size threshold
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
15 days ago
server: add --models-memory-max parameter to allow dynamically unloading models when they exceed a memory size threshold
Author
0cc4m
Committer
0cc4m
Parents
05e141a6
Loading