llama.cpp
server : implement prompt processing progress report in stream mode
#15827
Merged

server : implement prompt processing progress report in stream mode #15827

ngxson merged 7 commits into master from xsn/server_progress_api
ngxson
ngxson server : implement `return_progress`
29f1d50e
github-actions github-actions added examples
github-actions github-actions added server
ngxson add timings.cache_n
4404ad86
ngxson add progress.time_ms
e166a550
ngxson add test
f4213ccd
github-actions github-actions added python
ngxson fix test for chat/completions
ebcef910
ngxson ngxson marked this pull request as ready for review 103 days ago
ngxson readme: add docs on timings
b6ac24c6
ngxson ngxson requested a review from ggerganov ggerganov 103 days ago
ngxson ngxson requested a review from allozaur allozaur 103 days ago
ggerganov
ggerganov approved these changes on 2025-09-06
ggerganov
ggerganov commented on 2025-09-06
ngxson use ggml_time_us
053dc6b3
ngxson ngxson merged 61bdfd52 into master 103 days ago
ExtReMLapin
BradHutchings
BradHutchings
narendrachaudhary51
ngxson
ngxson ngxson deleted the xsn/server_progress_api branch 74 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone