llama.cpp
Avoid fp32->fp16->fp32 conversion on cdna in ggml_cuda_op_mul_mat_cublas
#11356
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
1
Changes
View On
GitHub
Avoid fp32->fp16->fp32 conversion on cdna in ggml_cuda_op_mul_mat_cublas
#11356
JohannesGaessler
merged 1 commit into
ggml-org:master
from
IMbackK:cdna_opt
IMbackK
requested a review
from
JohannesGaessler
354 days ago
github-actions
added
Nvidia GPU
github-actions
added
ggml
JohannesGaessler
approved these changes on 2025-01-24
IMbackK
force pushed
352 days ago
Avoid fp32->fp16->fp32 conversion on cdna in ggml_cuda_op_mul_mat_cublas
9d9ac6aa
IMbackK
force pushed
to
9d9ac6aa
352 days ago
JohannesGaessler
merged
9fbadaef
into master
352 days ago
Login to write a write a comment.
Login via GitHub
Reviewers
JohannesGaessler
Assignees
No one assigned
Labels
Nvidia GPU
ggml
Milestone
No milestone
Login to write a write a comment.
Login via GitHub