llama.cpp
44e18ef9 - vulkan: fix coopmat2 flash attention for non-contiguous inputs (#11281)

Commit
266 days ago
vulkan: fix coopmat2 flash attention for non-contiguous inputs (#11281) Add code similar to mul_mm_cm2 to force alignment of strides, to avoid a performance regression. Add noncontiguous FA tests in test-backend-ops. Fixes #11268.
Author
Parents
Loading