llama.cpp
3ba12fed
- kv-cache : extend cache quantization checks (#21586)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
29 days ago
kv-cache : extend cache quantization checks (#21586) to also check for enabled flash attention, instead of just auto.
References
#21586 - kv-cache : extend cache quantization checks
Author
Green-Sky
Parents
54739490
Loading