llama.cpp
a15ef8f8
- CUDA: fix partial offloading for ne0 % 256 != 0 (#8572)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Hide Minimap (CTRL+M)
Commit
328 days ago
CUDA: fix partial offloading for ne0 % 256 != 0 (#8572)
References
#8572 - CUDA: fix partial offloading for ne0 % 256 != 0
Author
JohannesGaessler
Parents
705b7ecf
Files
4
ggml
include
ggml-backend.h
src
ggml-alloc.c
ggml-backend.c
ggml-cuda.cu
Loading