llama.cpp
ec68e84c
- ggml : support bcast ggml_soft_max_ext, ggml_flash_attn_ext (#14435)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
126 days ago
ggml : support bcast ggml_soft_max_ext, ggml_flash_attn_ext (#14435) ggml-ci
References
#14435 - ggml : support broadcast for ggml_soft_max_ext and ggml_flash_attn_ext
Author
ggerganov
Committer
ggerganov
Parents
307e79d3
Loading