llama.cpp
CUDA: use MMQ instead of cuBLAS by default
#8075
Merged

CUDA: use MMQ instead of cuBLAS by default #8075

JohannesGaessler
github-actions github-actions added build
github-actions github-actions added Nvidia GPU
github-actions github-actions added ggml
Dampfinchen
JohannesGaessler
Nexesenex
slaren
JohannesGaessler
Nexesenex
slaren
slaren
slaren commented on 2024-06-24
JohannesGaessler JohannesGaessler force pushed from cc029763 to 5479853c 1 year ago
mofosyne mofosyne added Review Complexity : Medium
slaren
slaren approved these changes on 2024-06-24
JohannesGaessler CUDA: use MMQ instead of cuBLAS by default
61f3cb6e
JohannesGaessler JohannesGaessler force pushed from 5479853c to 61f3cb6e 1 year ago
JohannesGaessler JohannesGaessler merged a818f302 into master 1 year ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone