llama.cpp
33f890e5 - vulkan: support flash attention GQA/split_k with small batches (#18938)

Commit
49 days ago
vulkan: support flash attention GQA/split_k with small batches (#18938)
Author
Parents
Loading