llama.cpp
c9f670a1 - Implement non-greedy tokenizer that tries to maximize token lengths (#242)

Commit
2 years ago
Implement non-greedy tokenizer that tries to maximize token lengths (#242) * Implement non-greedy tokenizer that tries to maximize token lengths * Insert single space in front of the prompt - this is to match original llama tokenizer behavior --------- Co-authored-by: Jakub Horak <jakub.horak@ibawizard.net>
Author
Parents
Loading