llama.cpp
bb4f7a9e
- memory : fix broken batch splits for recurrent cache (#14575)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
71 days ago
memory : fix broken batch splits for recurrent cache (#14575) Splits producing more than one ubatch per batch for recurrent models were broken with #14512. This fixes it by moving the completeness check after the ubatch split loop.
References
#14575 - memory : fix broken batch splits for recurrent cache
Author
compilade
Parents
b8eeb874
Loading