llama.cpp
Support broadcast add & mul on CUDA (fixed)
#2192
Merged

Loading