llama.cpp
CUDA: fix Volta FlashAttention logic
#11615
Merged

CUDA: fix Volta FlashAttention logic #11615

JohannesGaessler
github-actions github-actions added Nvidia GPU
github-actions github-actions added ggml
JohannesGaessler JohannesGaessler requested a review from ggerganov ggerganov 276 days ago
ggerganov
JohannesGaessler CUDA: fix Volta FlashAttention logic
5ee63ee4
JohannesGaessler JohannesGaessler force pushed from ff0d3f67 to 5ee63ee4 276 days ago
JohannesGaessler
ggerganov
ggerganov approved these changes on 2025-02-03
ggerganov ggerganov merged 21c84b5d into master 276 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone