llama:use F32 precision in GLM4 attention and no FA #9130
fix glm GGG err
8a6ba03c
piDack
changed the title fix glm GGG err Fix glm4 GGG err 1 year ago
piDack
changed the title Fix glm4 GGG err llama:use F32 precision in GLM4 attention and no FA 1 year ago
ggerganov
approved these changes
on 2024-08-23
ggerganov
merged
a07c32ea
into master 1 year ago
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub