llama.cpp
d894f352 - perplexity : support using multiple sequences to allow larger batch sizes (#5946)

Commit
1 year ago
perplexity : support using multiple sequences to allow larger batch sizes (#5946) * perplexity : support using multiple sequences to allow larger batch sizes ggml-ci * set cparams.n_parallel to the number of sequences * print tested n_ctx, add assert
Author
Parents
Loading