llama.cpp
88540445 - Clarify default MMQ for CUDA and LLAMA_CUDA_FORCE_MMQ flag (#8115)

Commit
1 year ago
Clarify default MMQ for CUDA and LLAMA_CUDA_FORCE_MMQ flag (#8115) * Add message about int8 support * Add suggestions from review Co-authored-by: Johannes Gäßler <johannesg@5d6.de> --------- Co-authored-by: Johannes Gäßler <johannesg@5d6.de>
Parents
Loading