llama.cpp
daf4c6d3
- llama : fix worst case graph build
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Hide Minimap (CTRL+M)
Commit
1 year ago
llama : fix worst case graph build
References
#3228 - llama : custom attention mask + parallel decoding + no context swaps
Author
ggerganov
Parents
fa0e6778
Files
3
common
common.cpp
examples/llama-bench
llama-bench.cpp
llama.cpp
Loading