llama.cpp
CUDA: fix MMQ nwarps for AMD with warp_size==32
#15014
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
1
Changes
View On
GitHub
CUDA: fix MMQ nwarps for AMD with warp_size==32
#15014
IMbackK
merged 1 commit into
ggml-org:master
from
JohannesGaessler:cuda-mmq-fix-nwarps
CUDA: fix MMQ nwarps for AMD with warp_size==32
50e1ca96
JohannesGaessler
requested a review
from
IMbackK
37 days ago
github-actions
added
Nvidia GPU
github-actions
added
ggml
IMbackK
approved these changes on 2025-08-01
IMbackK
merged
9c35706b
into master
37 days ago
Login to write a write a comment.
Login via GitHub
Reviewers
IMbackK
Assignees
No one assigned
Labels
Nvidia GPU
ggml
Milestone
No milestone
Login to write a write a comment.
Login via GitHub