llama.cpp
e536426d
- llamafile : disable sgemm for batch-size 1 (#9330)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
llamafile : disable sgemm for batch-size 1 (#9330)
References
#9330 - Only enable sgemm for prompt processing, not for inference
Author
netrunnereve
Parents
1b9ae518
Loading