llama.cpp
7a221b67
- llama : use F32 precision in Qwen2 attention and no FA (#8412)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
llama : use F32 precision in Qwen2 attention and no FA (#8412)
References
#8412 - llama : use F32 precision in Qwen2 attention and no FA
Author
ggerganov
Parents
278d0e18
Loading