llama.cpp
44e18ef9
- vulkan: fix coopmat2 flash attention for non-contiguous inputs (#11281)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
266 days ago
vulkan: fix coopmat2 flash attention for non-contiguous inputs (#11281) Add code similar to mul_mm_cm2 to force alignment of strides, to avoid a performance regression. Add noncontiguous FA tests in test-backend-ops. Fixes #11268.
References
#11281 - vulkan: fix coopmat2 flash attention for non-contiguous inputs
Author
jeffbolznv
Parents
3edfa7d3
Loading