llama.cpp
d8c929ff - feat: Allow custom layer filters for hybrid recurrent

Commit
322 days ago
feat: Allow custom layer filters for hybrid recurrent This should help support architectures like Falcon H1 where there is overlap between layers that need attention and recurrent caches. https://github.com/ggml-org/llama.cpp/pull/13979#discussion_r2140748922 Branch: HybridRecurrentCache Signed-off-by: Gabe Goodhart <ghart@us.ibm.com>
Author
Committer
Parents
Loading