llama.cpp
vulkan: support arbitrary KV dimension in flash attention
#16160
Merged

vulkan: support arbitrary KV dimension in flash attention #16160

jeffbolznv
jeffbolznv jeffbolznv requested a review from 0cc4m 0cc4m 18 days ago
github-actions github-actions added Vulkan
github-actions github-actions added ggml
jeffbolznv
0cc4m
0cc4m
0cc4m approved these changes on 2025-09-27
jeffbolznv vulkan: support arbitrary KV dimension in flash attention
88fea950
jeffbolznv jeffbolznv force pushed from 48cbf213 to 88fea950 12 days ago
0cc4m 0cc4m merged e6d65fb0 into master 12 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone