llama.cpp
CUDA: Fix loop unrolling for BW in mul_mat_q_stream_k_fixup
#19053
Merged

CUDA: Fix loop unrolling for BW in mul_mat_q_stream_k_fixup #19053

ORippler
ORippler CUDA: Fix loop unrolling for BW in mul_mat_q_stream_k_fixup
390146ec
ORippler ORippler requested a review from JohannesGaessler JohannesGaessler 14 days ago
github-actions github-actions added Nvidia GPU
github-actions github-actions added ggml
JohannesGaessler
ORippler
JohannesGaessler
ORippler
ORippler
JohannesGaessler
JohannesGaessler
JohannesGaessler approved these changes on 2026-02-03
JohannesGaessler JohannesGaessler merged 1f1e57f2 into master 4 days ago
ORippler ORippler deleted the osimons/fix_bw_mmq_fixup_kernel branch 4 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone