llama.cpp
c3ebcfa1 - server : ensure batches are either all embed or all completion (#8420)

Commit
345 days ago
server : ensure batches are either all embed or all completion (#8420) * make sure batches are all embed or all non-embed * non-embedding batch for sampled tokens; fix unused params warning
Author
Parents
  • examples/server
    • File
      server.cpp