llama.cpp
88540445
- Clarify default MMQ for CUDA and LLAMA_CUDA_FORCE_MMQ flag (#8115)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
Clarify default MMQ for CUDA and LLAMA_CUDA_FORCE_MMQ flag (#8115) * Add message about int8 support * Add suggestions from review Co-authored-by: Johannes Gäßler <johannesg@5d6.de> --------- Co-authored-by: Johannes Gäßler <johannesg@5d6.de>
References
#8115 - Clarify default MMQ for CUDA and LLAMA_CUDA_FORCE_MMQ flag
Author
isaac-mcfadyen
Parents
c8771ab5
Loading