llama.cpp
21c84b5d
- CUDA: fix Volta FlashAttention logic (#11615)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Hide Minimap (CTRL+M)
Commit
145 days ago
CUDA: fix Volta FlashAttention logic (#11615)
References
#11615 - CUDA: fix Volta FlashAttention logic
Author
JohannesGaessler
Parents
d92cb67e
Files
2
ggml/src/ggml-cuda
fattn-wmma-f16.cu
fattn.cu
Loading