llama.cpp
3fe81781
- CUDA: faster q8_0 -> f16 dequantization (#4895)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
CUDA: faster q8_0 -> f16 dequantization (#4895)
References
#4895 - CUDA: faster q8_0 -> f16 dequantization
Author
JohannesGaessler
Parents
e7e4df03
Loading