llama.cpp
c3ebcfa1
- server : ensure batches are either all embed or all completion (#8420)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Hide Minimap (CTRL+M)
Commit
345 days ago
server : ensure batches are either all embed or all completion (#8420) * make sure batches are all embed or all non-embed * non-embedding batch for sampled tokens; fix unused params warning
References
#8420 - server: Ensure batches are either all embed or all completion (#8076)
Author
iamlemec
Parents
8a4441ea
Files
1
examples/server
server.cpp
Loading