llama.cpp
CUDA: GEMM for FP32/FP16/BF16 and ne11 <= 16
#15131
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
2
Changes
View On
GitHub
CUDA: GEMM for FP32/FP16/BF16 and ne11 <= 16
#15131
JohannesGaessler
merged 2 commits into
ggml-org:master
from
JohannesGaessler:cuda-mmf-3
CUDA: GEMM for FP32/FP16/BF16 and ne11 <= 16
e70fa55f
github-actions
added
Nvidia GPU
github-actions
added
ggml
slaren
approved these changes on 2025-08-06
JohannesGaessler
force pushed
from
140e30d8
to
94d7361b
64 days ago
JohannesGaessler
force pushed
from
94d7361b
to
cd393a6c
64 days ago
JohannesGaessler
force pushed
from
cd393a6c
to
285f342d
64 days ago
JohannesGaessler
force pushed
from
285f342d
to
6eeaf230
64 days ago
JohannesGaessler
force pushed
from
6eeaf230
to
781cbe04
64 days ago
try CI fix
52d9ccce
JohannesGaessler
force pushed
from
781cbe04
to
52d9ccce
64 days ago
JohannesGaessler
force pushed
from
66eab7a2
to
52d9ccce
64 days ago
JohannesGaessler
merged
1d72c841
into master
64 days ago
Login to write a write a comment.
Login via GitHub
Reviewers
slaren
Assignees
No one assigned
Labels
Nvidia GPU
ggml
Milestone
No milestone
Login to write a write a comment.
Login via GitHub