llama.cpp
8960efd0
- Vulkan: Add f32 accumulator support to quantized mul mat to fix GLM4 32B incoherence (#13607)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
212 days ago
Vulkan: Add f32 accumulator support to quantized mul mat to fix GLM4 32B incoherence (#13607)
References
#13607 - Vulkan: Support fp32 accumulator in quantized matmul to fix GLM4-32B incoherence
Author
0cc4m
Parents
725f23f1
Loading