llama.cpp
f4003cfb
- fix nwarps > batch size
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
fix nwarps > batch size
References
#7527 - CUDA: quantized KV support for FA vec
Author
JohannesGaessler
Committer
JohannesGaessler
Parents
f0877604
Loading