llama: (proposal) propagating the results of `graph_compute` to the user interface #9525
llama: propagating the results of `graph_compute` to the user interface
5e354e3c
llama: reverting kv_cache in case of failed compute
47018932
llama: `llama_kv_cache_state` was removed, only the result of `llama_…
acb95283
llama: restore a kv_cache in case of failed computation
0026c810
llama: correct reverting of the entire batch.
ee599f90
Xarbirus
force pushed
to
ee599f90
1 year ago
llama: updated comments
0638c448
llama : add comments about KV cache state after error
9ef5d089
ggerganov
merged
fb4a0ec0
into master 1 year ago
Xarbirus
deleted the ggml_status_to_user branch 1 year ago
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub