llama.cpp
Force FP32 compute in GLM4 FFN Down
#13101
Merged

Force FP32 compute in GLM4 FFN Down #13101

city96
city96 Force FP32 compute in cuBLAS GEMM
6efd8727
github-actions github-actions added Nvidia GPU
github-actions github-actions added ggml
Beinsezii
JohannesGaessler
JohannesGaessler requested changes on 2025-04-25
city96
JohannesGaessler
city96 Revert "Force FP32 compute in cuBLAS GEMM"
db52579a
city96 Force F32 compute in GLM4 ffn down
70975676
city96
JohannesGaessler
JohannesGaessler approved these changes on 2025-04-25
city96 city96 changed the title CUDA: Force FP32 compute in cuBLAS GEMM Force FP32 compute in GLM4 FFN Down 232 days ago
city96 Edit comment to clarify issue
06113f00
city96
JohannesGaessler JohannesGaessler merged 558a7647 into master 232 days ago
JohannesGaessler
Mushoz
city96
city96 city96 deleted the glm4_cublas_fix branch 232 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone