llama.cpp
CUDA: add dynamic shared mem to softmax, refactor general usage
#14497
Merged

CUDA: add dynamic shared mem to softmax, refactor general usage #14497

am17an
am17an am17an requested a review from JohannesGaessler JohannesGaessler 128 days ago
JohannesGaessler
JohannesGaessler commented on 2025-07-02
github-actions github-actions added testing
github-actions github-actions added Nvidia GPU
github-actions github-actions added ggml
am17an am17an force pushed from 6429086b to 4c7bcaab 128 days ago
am17an am17an force pushed from 4c7bcaab to a67ef5c0 128 days ago
am17an am17an requested a review from JohannesGaessler JohannesGaessler 128 days ago
am17an CUDA: add dynamic shared mem to softmax, refactor general usage
b9bcb7d7
am17an Review: refactor switch statement, change cross_entropy to use full size
34e5142d
am17an rebase
7b162818
am17an am17an force pushed from a67ef5c0 to 7b162818 127 days ago
JohannesGaessler
JohannesGaessler approved these changes on 2025-07-02
JohannesGaessler
CISC
CISC
am17an am17an merged 55c2646b into master 127 days ago
am17an am17an deleted the cuda_increase_shared_mem_limits branch 127 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone