transformers
Enable Quantize KV Cache for Mistral Model
#35042
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
1
Changes
View On
GitHub
Enable Quantize KV Cache for Mistral Model
#35042
zucchini-nlp
merged 1 commit into
huggingface:main
from
Bojun-Feng:enh/enable_quant_kv_mistral
fix #35041
89eb601d
zucchini-nlp
approved these changes on 2024-12-03
zucchini-nlp
requested a review
from
Rocketknight1
1 year ago
zucchini-nlp
merged
96618960
into main
288 days ago
Login to write a write a comment.
Login via GitHub
Reviewers
zucchini-nlp
Rocketknight1
Assignees
No one assigned
Labels
None yet
Milestone
No milestone
Login to write a write a comment.
Login via GitHub