llama.cpp
3600cc28
- llama : use n_swa + n_ubatch cells for SWA cache (#13833)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
98 days ago
llama : use n_swa + n_ubatch cells for SWA cache (#13833) * llama : use n_swa + n_ubatch cells for SWA cache ggml-ci * llama : add warning about multi-sqeuence SWA contexts
References
#13833 - llama : use n_swa + n_ubatch cells for SWA cache
Author
ggerganov
Parents
c7e0a205
Loading