llama.cpp
8875523e
- vulkan: support softmax/FA batch and broadcast (#14449)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
166 days ago
vulkan: support softmax/FA batch and broadcast (#14449)
References
#14435 - ggml : support broadcast for ggml_soft_max_ext and ggml_flash_attn_ext
Author
jeffbolznv
Committer
ggerganov
Parents
ec68e84c
Loading