llama.cpp
server : avoid breaking KV cache when prompt >= n_ctx (#6855)
#8359
Closed

server : avoid breaking KV cache when prompt >= n_ctx (#6855) #8359

prfd wants to merge 1 commit into ggml-org:master from server
prfd
github-actions github-actions added examples
github-actions github-actions added python
github-actions github-actions added server
ggerganov
ggerganov commented on 2024-07-09
HanClinto
HanClinto commented on 2024-07-10
HanClinto
server : avoid breaking KV cache when prompt >= n_ctx (#6855)
e6a5a6c6
mofosyne mofosyne added Review Complexity : Medium
prfd prfd closed this 171 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone