llama.cpp
133d99c5
- CUDA: deduplicate FlashAttention code (#7352)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
CUDA: deduplicate FlashAttention code (#7352)
References
#7352 - CUDA: deduplicate FlashAttention code
Author
JohannesGaessler
Parents
cb42c294
Loading