llama.cpp
da040034 - ggml-cuda : perform cublas fp16 matrix multiplication as fp16 (#3370)

Commit
1 year ago
ggml-cuda : perform cublas fp16 matrix multiplication as fp16 (#3370) * ggml-cuda : perform cublas fp16 matrix multiplication as fp16 * try to fix rocm build * restrict fp16 mat mul to volta and up
Author
Parents
Loading