llama.cpp
7a221b67 - llama : use F32 precision in Qwen2 attention and no FA (#8412)

Commit
1 year ago
llama : use F32 precision in Qwen2 attention and no FA (#8412)
Author
Parents
Loading