llama.cpp
metal : use F32 attention accumulators in FA kernels
#13975
Merged

metal : use F32 attention accumulators in FA kernels #13975

ggerganov merged 1 commit into master from gg/metal-fa-acc-f32
ggerganov
ggerganov metal : use F32 accumulators in FA kernels
21be70ec
github-actions github-actions added ggml
github-actions github-actions added Apple Metal
ggerganov ggerganov merged ea394d7a into master 107 days ago
ggerganov ggerganov deleted the gg/metal-fa-acc-f32 branch 107 days ago
Animaxx

Login to write a write a comment.

Login via GitHub

Reviewers
No reviews
Assignees
No one assigned
Labels
Milestone