llama.cpp
71b69aa7
- cuda : fix flash_attn kernel to produce same results as CPU
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
cuda : fix flash_attn kernel to produce same results as CPU
Author
ggerganov
Committer
ggerganov
Parents
fd878f71
Loading