llama.cpp
cuda: Add Q5_1, Q5_0, Q4_1 and Q4_0 to F32 conversion support. (#10976)
#12000
Merged

cuda: Add Q5_1, Q5_0, Q4_1 and Q4_0 to F32 conversion support. (#10976) #12000

gcp
github-actions github-actions added Nvidia GPU
github-actions github-actions added ggml
JohannesGaessler
JohannesGaessler commented on 2025-02-21
gcp gcp force pushed 303 days ago
gcp cuda: Add Q5_1, Q5_0, Q4_1 and Q4_0 to F32 conversion support. (#10976)
295573fc
gcp gcp force pushed to 295573fc 303 days ago
gcp
JohannesGaessler
JohannesGaessler approved these changes on 2025-02-22
JohannesGaessler JohannesGaessler merged d7090842 into master 302 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone