llama.cpp
gguf-py : add Numpy MXFP4 de/quantization support
#15111
Merged

Commits
  • gguf-py : add MXFP4 de/quantization support
    compilade committed 36 days ago
  • ggml-quants : handle zero amax for MXFP4
    compilade committed 36 days ago
Loading