llama.cpp
c54bba86
- ggml : optimize cuda cumsum fallback kernel (#18343)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
16 days ago
ggml : optimize cuda cumsum fallback kernel (#18343)
References
#18343 - ggml : optimize cuda cumsum fallback (~2.5x speedup vs CUB)
Author
Aadeshveer
Parents
f5acfb2f
Loading