llama.cpp
llama:use F32 precision in GLM4 attention and no FA
#9130
Merged

llama:use F32 precision in GLM4 attention and no FA #9130

piDack
piDack fix glm GGG err
8a6ba03c
piDack piDack changed the title fix glm GGG err Fix glm4 GGG err 1 year ago
piDack piDack changed the title Fix glm4 GGG err llama:use F32 precision in GLM4 attention and no FA 1 year ago
ggerganov
ggerganov approved these changes on 2024-08-23
ggerganov ggerganov merged a07c32ea into master 1 year ago
piDack
ThiloteE

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone