llama.cpp
75923754 - Better perplexity for 2- and 3-bit quantization for LLaMA-v2-70B (#2807)

Commit
2 years ago
Better perplexity for 2- and 3-bit quantization for LLaMA-v2-70B (#2807) * Better perplexity for 2- and 3-bit quantization for the 70B model * PR comment --------- Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com>
Author
Parents
Loading