llama.cpp
server: Ensure batches are either all embed or all completion (#8076)
#8420
Merged

server: Ensure batches are either all embed or all completion (#8076) #8420

ggerganov merged 2 commits into ggml-org:master from iamlemec:server-embed
iamlemec
iamlemec make sure batches are all embed or all non-embed
40c99abb
github-actions github-actions added examples
github-actions github-actions added server
iamlemec iamlemec changed the title server: Ensure batches are either all embed or all completion server: Ensure batches are either all embed or all completion (#8076) 1 year ago
compilade
compilade commented on 2024-07-10
iamlemec non-embedding batch for sampled tokens; fix unused params warning
371cb8df
ggerganov
iamlemec
ggerganov
ggerganov approved these changes on 2024-07-12
ggerganov ggerganov merged c3ebcfa1 into master 1 year ago
okigan
iamlemec
okigan
iamlemec
okigan

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone