llama.cpp
a9b5fe98
- fix: Fix logic for initializing inputs and attn layers for hybrid caches
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
197 days ago
fix: Fix logic for initializing inputs and attn layers for hybrid caches Branch: GraniteFour Signed-off-by: Gabe Goodhart <ghart@us.ibm.com>
References
#13979 - Hybrid recurrent cache
Author
gabe-l-hart
Committer
gabe-l-hart
Parents
e3c16315
Loading