llama.cpp
d8919424
- CUDA: fix FlashAttention on Turing (#13415)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
220 days ago
CUDA: fix FlashAttention on Turing (#13415)
References
#13415 - CUDA: fix FlashAttention on Turing
Author
JohannesGaessler
Parents
7fef1176
Loading