llama.cpp
9ca2e677
- server : add speculative decoding support (#10455)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
286 days ago
server : add speculative decoding support (#10455) * server : add speculative decoding support ggml-ci * server : add helper function slot.can_speculate() ggml-ci
References
#10455 - server : add speculative decoding support
Author
ggerganov
Parents
5931c1f2
Loading