llama.cpp
e80f06d2
- llama : correction of the attn.v.weight quantization for IQ3_XS (#6209)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
1 year ago
llama : correction of the attn.v.weight quantization for IQ3_XS (#6209) IQ3_XS was not mentioned, IQ3_S and IQ3_M were present twice. That PR corrects this in the manner which was probably intended initially.
References
#6209 - correction of the attn.v.weight quantization for IQ3_XS
Author
Nexesenex
Parents
f77a8ffd
Loading