llama.cpp
c3ebcfa1 - server : ensure batches are either all embed or all completion (#8420)

Commit
1 year ago
server : ensure batches are either all embed or all completion (#8420) * make sure batches are all embed or all non-embed * non-embedding batch for sampled tokens; fix unused params warning
Author
Parents
Loading