llama.cpp
Vulkan: Support fp32 accumulator in quantized matmul to fix GLM4-32B incoherence
#13607
Merged

Vulkan: Support fp32 accumulator in quantized matmul to fix GLM4-32B incoherence #13607

0cc4m merged 1 commit into master from 0cc4m/vulkan-mmq-fp32-acc-glm4
0cc4m
0cc4m Vulkan: Add f32 accumulator support to quantized mul mat to fix GLM4 …
35d675d8
0cc4m 0cc4m requested a review from jeffbolznv jeffbolznv 214 days ago
github-actions github-actions added Vulkan
github-actions github-actions added ggml
jeffbolznv
0cc4m
JohannesGaessler
jeffbolznv
0cc4m
jeffbolznv
JohannesGaessler
0cc4m
jeffbolznv
0cc4m
0cc4m
jeffbolznv
0cc4m
jeffbolznv
jeffbolznv approved these changes on 2025-05-18
ggerganov
0cc4m
0cc4m 0cc4m marked this pull request as ready for review 212 days ago
ggerganov
LostRuins
jeffbolznv
0cc4m
0cc4m 0cc4m merged 8960efd0 into master 212 days ago
0cc4m 0cc4m deleted the 0cc4m/vulkan-mmq-fp32-acc-glm4 branch 212 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone