llama.cpp
llama: (proposal) propagating the results of `graph_compute` to the user interface
#9525
Merged

llama: (proposal) propagating the results of `graph_compute` to the user interface #9525

Xarbirus
Xarbirus
slaren
Xarbirus
Xarbirus Xarbirus force pushed from 95ce058c to 20510ea0 349 days ago
vignesh1507
vignesh1507 approved these changes on 2024-10-06
Xarbirus Xarbirus force pushed from 20510ea0 to 5535683e 329 days ago
Xarbirus Xarbirus force pushed from 5535683e to 059e78cf 329 days ago
Xarbirus
slaren
ggerganov
slaren
ggerganov
Xarbirus
Xarbirus
slaren
Xarbirus
slaren
ggerganov
compilade
Xarbirus llama: propagating the results of `graph_compute` to the user interface
5e354e3c
Xarbirus llama: reverting kv_cache in case of failed compute
47018932
Xarbirus llama: `llama_kv_cache_state` was removed, only the result of `llama_…
acb95283
Xarbirus llama: restore a kv_cache in case of failed computation
0026c810
Xarbirus llama: correct reverting of the entire batch.
ee599f90
Xarbirus
Xarbirus Xarbirus force pushed from bbf27cc2 to ee599f90 303 days ago
ggerganov
Xarbirus llama: updated comments
0638c448
Xarbirus
ggerganov llama : add comments about KV cache state after error
9ef5d089
ggerganov ggerganov merged fb4a0ec0 into master 299 days ago
Xarbirus Xarbirus deleted the ggml_status_to_user branch 289 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone