llama.cpp
server : avoid breaking KV cache when prompt >= n_ctx (#6855)
#8359
Closed
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
1
Changes
View On
GitHub
server : avoid breaking KV cache when prompt >= n_ctx (#6855)
#8359
prfd
wants to merge 1 commit into
ggml-org:master
from server
github-actions
added
examples
github-actions
added
python
github-actions
added
server
ggerganov
commented on 2024-07-09
HanClinto
commented on 2024-07-10
server : avoid breaking KV cache when prompt >= n_ctx (#6855)
e6a5a6c6
mofosyne
added
Review Complexity : Medium
prfd
closed this
171 days ago
Login to write a write a comment.
Login via GitHub
Reviewers
ggerganov
HanClinto
Assignees
No one assigned
Labels
examples
python
Review Complexity : Medium
server
Milestone
No milestone
Login to write a write a comment.
Login via GitHub