llama.cpp
ggml-cuda : perform cublas fp16 matrix multiplication as fp16
#3370
Merged

ggml-cuda : perform cublas fp16 matrix multiplication as fp16 #3370

ggerganov merged 3 commits into master from cublas-f16
slaren
slaren ggml-cuda : perform cublas fp16 matrix multiplication as fp16
79fe5a1f
slaren try to fix rocm build
32ada53c
cebtenzzre
cebtenzzre commented on 2023-09-27
slaren restrict fp16 mat mul to volta and up
7d5674dd
slaren
bobqianic
ggerganov
ggerganov approved these changes on 2023-09-28
ggerganov ggerganov merged da040034 into master 1 year ago
ggerganov
ggerganov
slaren slaren deleted the cublas-f16 branch 1 year ago
whoreson
whoreson
cebtenzzre
ByerRA

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone