llama.cpp
1fb033fd
- ggml : ggml_rope now takes a vector with positions instead of n_past
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
ggml : ggml_rope now takes a vector with positions instead of n_past
References
#3234 - llama : store non-RoPEd K cache
#3228 - llama : custom attention mask + parallel decoding + no context swaps
Author
ggerganov
Committer
ggerganov
Parents
3b4bab6a
Loading