llama.cpp
gguf-py : add Numpy MXFP4 de/quantization support
#15111
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
2
Changes
View On
GitHub
Commits
gguf-py : add MXFP4 de/quantization support
compilade
committed
36 days ago
ggml-quants : handle zero amax for MXFP4
compilade
committed
36 days ago
Loading