llama.cpp
5ca49cbe
- ggml: implement quantized KV cache for FA (#7372)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Hide Minimap (CTRL+M)
Commit
1 year ago
ggml: implement quantized KV cache for FA (#7372)
References
#7372 - ggml: implement quantized KV cache for FA
Author
JohannesGaessler
Parents
1b01f06d
Files
1
ggml.c
Loading