llama.cpp
Fix FlashAttention debug test, FP32 assert
#7684
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
1
Changes
View On
GitHub
Fix FlashAttention debug test, FP32 assert
#7684
JohannesGaessler
merged 1 commit into
ggml-org:master
from
JohannesGaessler:cuda-fa-quant-fixup
Fix FlashAttention debug test, FP32 assert
45102363
github-actions
added
testing
github-actions
added
Nvidia GPU
github-actions
added
ggml
slaren
approved these changes on 2024-06-01
JohannesGaessler
merged
e141ce62
into master
1 year ago
Login to write a write a comment.
Login via GitHub
Reviewers
slaren
Assignees
No one assigned
Labels
testing
Nvidia GPU
ggml
Milestone
No milestone
Login to write a write a comment.
Login via GitHub