llama.cpp
CUDA: use mma FA kernel for gqa > 4 on RTX 4000
#15035
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
1
Changes
View On
GitHub
Commits
CUDA: use mma FA kernel for gqa > 4 on RTX 4000
JohannesGaessler
committed
219 days ago
Loading