llama.cpp
7f5ff558 - server: stop generation at `n_ctx_train` if `n_predict` is not set (#6638)

Commit
1 year ago
server: stop generation at `n_ctx_train` if `n_predict` is not set (#6638) * server: cap n_predict if not set to n_ctx_train * server: fix infinite loop * server: infinite loop, move in process_token server: infinite loop: set stop limit to true * minor: spaces * minor: spaces * server: include prompt tokens in the EOS limit
Author
Parents
Loading