llama.cpp
17c97fb0
- CUDA: mul_mat_vec_q max. batch size 8 -> 4 (#5370)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
2 years ago
CUDA: mul_mat_vec_q max. batch size 8 -> 4 (#5370)
References
#5370 - CUDA: mul_mat_vec_q max. batch size 8 -> 4
Author
JohannesGaessler
Parents
b08f22c8
Loading