llama.cpp
eb425962
- llama : do not use KV cache for non-causal models
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
llama : do not use KV cache for non-causal models ggml-ci
References
#5796 - llama : fix embeddings
Author
ggerganov
Committer
ggerganov
Parents
d0347840
Loading