llama.cpp
Prompt lookup decoding
#4484
Merged

Prompt lookup decoding #4484

LeonEricsson
LeonEricsson initial commit, going through initializations
cae8f50b
LeonEricsson main loop finished, starting to debug
0ec5fdb5
LeonEricsson BUG: generates gibberish/repeating tokens after a while
1665ad8b
LeonEricsson Merge branch 'ggerganov:master' into prompt-lookup
34048416
LeonEricsson
ggerganov
LeonEricsson
LeonEricsson kv_cache management
21431197
LeonEricsson Merge branch 'prompt-lookup' of github.com:LeonEricsson/llama.cpp int…
45b8032b
stduhpf
LeonEricsson
LeonEricsson Added colors to distinguish drafted tokens (--color). Updated README
1b26d715
ggerganov lookup : fix token positions in the draft batch
5b279754
ggerganov
ggerganov
ggerganov approved these changes on 2023-12-17
stduhpf
ggerganov lookup : use n_draft from CLI params
d8ed670c
ggerganov
LeonEricsson
stduhpf
ggerganov
stduhpf
LeonEricsson
apoorvumang
ggerganov
ggml-org ggml-org deleted a comment from ghost on 2023-12-20
apoorvumang
apoorvumang
apoorvumang
LeonEricsson
apoorvumang
kalomaze
kalomaze
stduhpf
kalomaze
ggerganov
stduhpf
LeonEricsson
sorasoras
ggerganov lookup : final touches
50ea1ef7
ggerganov ggerganov merged 7082d24c into master 1 year ago
LeonEricsson
x4080
stduhpf
apoorvumang
x4080
x4080

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone