auto-round
a52c81ef
- Reduce RAM usage of quantizing VLM models and fix issues of quantizing gemma4
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
2 days ago
Reduce RAM usage of quantizing VLM models and fix issues of quantizing gemma4 Signed-off-by: lvliang-intel <liang1.lv@intel.com>
References
#1791 - Reduce RAM usage of quantizing VLM models and fix some issues of quantizing gemma4
Author
lvliang-intel
Committer
lvliang-intel
Parents
5bbe39ba
Loading