llama.cpp
ggml: implement quantized KV cache for FA
#7372
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
1
Changes
View On
GitHub
ggml: implement quantized KV cache for FA
#7372
JohannesGaessler
merged 1 commit into
ggml-org:master
from
JohannesGaessler:fa-quantize-3
JohannesGaessler
force pushed
1 year ago
ggerganov
approved these changes on 2024-05-19
ggml: implement quantized KV cache for FA
b7da2e86
JohannesGaessler
force pushed
to
b7da2e86
1 year ago
JohannesGaessler
merged
5ca49cbe
into master
1 year ago
Login to write a write a comment.
Login via GitHub
Reviewers
ggerganov
Assignees
No one assigned
Labels
None yet
Milestone
No milestone
Login to write a write a comment.
Login via GitHub