llama.cpp
CUDA: GEMM for FP32/FP16/BF16 and ne11 <= 16
#15131
Merged

CUDA: GEMM for FP32/FP16/BF16 and ne11 <= 16 #15131

JohannesGaessler
JohannesGaessler CUDA: GEMM for FP32/FP16/BF16 and ne11 <= 16
e70fa55f
github-actions github-actions added Nvidia GPU
github-actions github-actions added ggml
slaren
slaren
slaren approved these changes on 2025-08-06
JohannesGaessler JohannesGaessler force pushed from 140e30d8 to 94d7361b 64 days ago
JohannesGaessler JohannesGaessler force pushed from 94d7361b to cd393a6c 64 days ago
JohannesGaessler JohannesGaessler force pushed from cd393a6c to 285f342d 64 days ago
JohannesGaessler JohannesGaessler force pushed from 285f342d to 6eeaf230 64 days ago
JohannesGaessler JohannesGaessler force pushed from 6eeaf230 to 781cbe04 64 days ago
JohannesGaessler try CI fix
52d9ccce
JohannesGaessler JohannesGaessler force pushed from 781cbe04 to 52d9ccce 64 days ago
IMbackK
JohannesGaessler
IMbackK
IMbackK
IMbackK
JohannesGaessler
IMbackK
slaren
JohannesGaessler JohannesGaessler force pushed from 66eab7a2 to 52d9ccce 64 days ago
JohannesGaessler JohannesGaessler merged 1d72c841 into master 64 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone