llama.cpp
d6a50940 - ggml-webgpu: Fix bug in FlashAttention support check (#22492)

Commit
11 days ago
ggml-webgpu: Fix bug in FlashAttention support check (#22492) * Fix flashattention support check for devices that don't support subgroups * set path to none if kv_tile doesn't fit
Author
Parents
Loading