llama.cpp
21c84b5d - CUDA: fix Volta FlashAttention logic (#11615)

Commit
145 days ago
CUDA: fix Volta FlashAttention logic (#11615)
Parents
  • ggml/src/ggml-cuda
    • File
      fattn-wmma-f16.cu
    • File
      fattn.cu