llama.cpp
58b515ca - convert-hf : add --outtype auto-f16

Comment changes are shownComment changes are hidden
Commit
1 year ago
convert-hf : add --outtype auto-f16 A reason for this to exist is for model quantizers who want an initial GGUF with the most fidelity to the original model while still using a 16-bit float type instead of 32-bit floats.
Author
Committer
Parents
  • File
    convert-hf-to-gguf.py
  • gguf-py/gguf
    • File
      constants.py
Loading