llama.cpp
c9c64dee
- Set GLM4 blk.*.attn_output.weight, kqv_out-* matmul to GGML_PREC_F32 to fix infinity values in output (#13639)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
212 days ago
Set GLM4 blk.*.attn_output.weight, kqv_out-* matmul to GGML_PREC_F32 to fix infinity values in output (#13639)
References
#13639 - Fix GLM4 incoherence with fp16 accumulators
Author
0cc4m
Parents
c00a2634
Loading