llama.cpp
e80f06d2 - llama : correction of the attn.v.weight quantization for IQ3_XS (#6209)

Commit
1 year ago
llama : correction of the attn.v.weight quantization for IQ3_XS (#6209) IQ3_XS was not mentioned, IQ3_S and IQ3_M were present twice. That PR corrects this in the manner which was probably intended initially.
Author
Parents
Loading