llama.cpp
vulkan: support arbitrary KV dimension in flash attention
#16160
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
1
Changes
View On
GitHub
vulkan: support arbitrary KV dimension in flash attention
#16160
0cc4m
merged 1 commit into
ggml-org:master
from
jeffbolznv:KV_bounds_check
jeffbolznv
requested a review
from
0cc4m
187 days ago
github-actions
added
Vulkan
github-actions
added
ggml
0cc4m
approved these changes on 2025-09-27
vulkan: support arbitrary KV dimension in flash attention
88fea950
jeffbolznv
force pushed
to
88fea950
181 days ago
0cc4m
merged
e6d65fb0
into master
181 days ago
Login to write a write a comment.
Login via GitHub
Reviewers
0cc4m
Assignees
No one assigned
Labels
Vulkan
ggml
Milestone
No milestone
Login to write a write a comment.
Login via GitHub