llama.cpp
vulkan: support arbitrary KV dimension in flash attention
#16160
Merged

vulkan: support arbitrary KV dimension in flash attention #16160

jeffbolznv
jeffbolznv jeffbolznv requested a review from 0cc4m 0cc4m 187 days ago
github-actions github-actions added Vulkan
github-actions github-actions added ggml
jeffbolznv
0cc4m
0cc4m
0cc4m approved these changes on 2025-09-27
jeffbolznv vulkan: support arbitrary KV dimension in flash attention
88fea950
jeffbolznv jeffbolznv force pushed to 88fea950 181 days ago
0cc4m 0cc4m merged e6d65fb0 into master 181 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone