llama.cpp
CUDA: use shared mem for ssm_conv
#20128
Merged

CUDA: use shared mem for ssm_conv #20128

am17an merged 5 commits into ggml-org:master from am17an:cuda_ssm_conv
am17an
am17an CUDA: use shared mem for ssm_conv
0141e9c0
am17an am17an requested a review from ggerganov ggerganov 24 days ago
am17an am17an requested a review from JohannesGaessler JohannesGaessler 24 days ago
github-actions github-actions added testing
github-actions github-actions added Nvidia GPU
github-actions github-actions added ggml
am17an fuse silu + ssm_conv
7ba1b0ae
am17an fuse unary + mul
de3856d4
am17an
CISC
CISC commented on 2026-03-05
JohannesGaessler
am17an
am17an enable for fp16
d440e643
JohannesGaessler
JohannesGaessler approved these changes on 2026-03-06
am17an formatting
43f3f54f
am17an am17an merged 1e38a7a6 into master 23 days ago
am17an am17an deleted the cuda_ssm_conv branch 23 days ago
IMbackK
IMbackK commented on 2026-03-06

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone