llama.cpp
da040034
- ggml-cuda : perform cublas fp16 matrix multiplication as fp16 (#3370)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
ggml-cuda : perform cublas fp16 matrix multiplication as fp16 (#3370) * ggml-cuda : perform cublas fp16 matrix multiplication as fp16 * try to fix rocm build * restrict fp16 mat mul to volta and up
References
#3370 - ggml-cuda : perform cublas fp16 matrix multiplication as fp16
Author
slaren
Parents
e5196210
Loading