llama.cpp
75923754
- Better perplexity for 2- and 3-bit quantization for LLaMA-v2-70B (#2807)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
2 years ago
Better perplexity for 2- and 3-bit quantization for LLaMA-v2-70B (#2807) * Better perplexity for 2- and 3-bit quantization for the 70B model * PR comment --------- Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com>
References
#2807 - Better perplexity for 2- and 3-bit quantization for LLaMA-v2-70B
Author
ikawrakow
Parents
771551a7
Loading