llama.cpp
CUDA: fix MMQ nwarps for AMD with warp_size==32
#15014
Merged

CUDA: fix MMQ nwarps for AMD with warp_size==32 #15014

JohannesGaessler
JohannesGaessler CUDA: fix MMQ nwarps for AMD with warp_size==32
50e1ca96
JohannesGaessler JohannesGaessler requested a review from IMbackK IMbackK 37 days ago
github-actions github-actions added Nvidia GPU
github-actions github-actions added ggml
IMbackK
IMbackK approved these changes on 2025-08-01
IMbackK IMbackK merged 9c35706b into master 37 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone