llama.cpp
81f28f25
- Remove call to ggml_cuda_mul_mat_get_wsize
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
2 years ago
Remove call to ggml_cuda_mul_mat_get_wsize
References
#1237 - Generalize `quantize_fns` for simpler FP16 handling
Author
sw
Parents
f9c585f0
Loading