llama.cpp
8185710a
- CUDA: use only 1 thread if fully offloaded (#2915)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
CUDA: use only 1 thread if fully offloaded (#2915)
References
#2915 - CUDA: use 1 thread if model is fully offloaded
Author
JohannesGaessler
Parents
7eb41179
Loading