llama.cpp
e54d41be
- gguf-py : add Numpy MXFP4 de/quantization support (#15111)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
29 days ago
gguf-py : add Numpy MXFP4 de/quantization support (#15111) * gguf-py : add MXFP4 de/quantization support * ggml-quants : handle zero amax for MXFP4
References
#15111 - gguf-py : add Numpy MXFP4 de/quantization support
Author
compilade
Parents
4850b52a
Loading