llama.cpp
e5ca3937
- llama : do not cap thread count when MoE on CPU (#5419)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
llama : do not cap thread count when MoE on CPU (#5419) * Not capping thread count when MoE inference is running on CPU * Whitespace
References
#5419 - Not capping thread count when MoE inference is running on CPU
Author
ptsochantaris
Parents
e4124c24
Loading