llama.cpp
9067487c - ggml : fix FA mask dim 2 and 3 (#14505)

Commit
164 days ago
ggml : fix FA mask dim 2 and 3 (#14505) * ggml : fix FA mask dim 2 and 3 ggml-ci * backends : unsupport batched FA in CUDA and Vulkan ggml-ci * vulkan : disable FA for mask->ne[2] != 1
Author
Parents
Loading