llama.cpp
llama: (proposal) propagating the results of `graph_compute` to the user interface
#9525
Merged

llama: (proposal) propagating the results of `graph_compute` to the user interface #9525

Xarbirus
Xarbirus
slaren
Xarbirus
Xarbirus Xarbirus force pushed 1 year ago
vignesh1507
vignesh1507 approved these changes on 2024-10-06
Xarbirus Xarbirus force pushed 1 year ago
Xarbirus Xarbirus force pushed 1 year ago
Xarbirus
slaren
ggerganov
slaren
ggerganov
Xarbirus
Xarbirus
slaren
Xarbirus
slaren
ggerganov
compilade
Xarbirus llama: propagating the results of `graph_compute` to the user interface
5e354e3c
Xarbirus llama: reverting kv_cache in case of failed compute
47018932
Xarbirus llama: `llama_kv_cache_state` was removed, only the result of `llama_…
acb95283
Xarbirus llama: restore a kv_cache in case of failed computation
0026c810
Xarbirus llama: correct reverting of the entire batch.
ee599f90
Xarbirus
Xarbirus Xarbirus force pushed to ee599f90 1 year ago
ggerganov
Xarbirus llama: updated comments
0638c448
Xarbirus
ggerganov llama : add comments about KV cache state after error
9ef5d089
ggerganov ggerganov merged fb4a0ec0 into master 1 year ago
Xarbirus Xarbirus deleted the ggml_status_to_user branch 1 year ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone