llama.cpp
1425f587 - CUDA: attention sinks for mma FlashAttention (#15157)

Commit
30 days ago
CUDA: attention sinks for mma FlashAttention (#15157)
Parents
Loading