llama.cpp
4d196981 - convert : force patch_embd weights to F16 or F32 to avoid broken GGUFs (#15367)

Commit
137 days ago
convert : force patch_embd weights to F16 or F32 to avoid broken GGUFs (#15367) * force patch_embd weights to f32 * use MmprojModel base tensor_force_quant instead
Author
Parents
Loading