llama.cpp
865af990 - ggml : ggml_flash_attn_ext() support ALiBi (CUDA)

Commit
1 year ago
ggml : ggml_flash_attn_ext() support ALiBi (CUDA) ggml-ci
Author
Parents
Loading