llama.cpp
e9b7a5cb - llama : use n_threads_batch only when n_tokens >= 32

Commit
1 year ago
llama : use n_threads_batch only when n_tokens >= 32 ggml-ci
Author
Committer
Parents
  • File
    llama.cpp