llama.cpp
ggml-cuda : perform cublas fp16 matrix multiplication as fp16
#3370
Merged

Loading