llama.cpp
970b5ab7
- ggml-cuda : add TQ2_0 support
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
ggml-cuda : add TQ2_0 support
References
#11183 - ggml-cuda : add TQ2_0 kernels, for ternary inference on GPU
Author
compilade
Parents
5cd85b5e
Loading