llama.cpp
Clarify default MMQ for CUDA and LLAMA_CUDA_FORCE_MMQ flag
#8115
Merged

Clarify default MMQ for CUDA and LLAMA_CUDA_FORCE_MMQ flag #8115

JohannesGaessler merged 2 commits into ggml-org:master from mmq-readme-update
isaac-mcfadyen
isaac-mcfadyen Add message about int8 support
b6cd699d
mofosyne mofosyne added Review Complexity : Low
slaren
slaren approved these changes on 2024-06-25
slaren slaren requested a review from JohannesGaessler JohannesGaessler 1 year ago
JohannesGaessler
JohannesGaessler requested changes on 2024-06-25
isaac-mcfadyen Add suggestions from review
37ff7098
JohannesGaessler
JohannesGaessler approved these changes on 2024-06-26
JohannesGaessler JohannesGaessler merged 88540445 into master 1 year ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone