transformers
2230d149 - fix get_keys_to_not_convert() to return correct modules for full precision inference (#25105)

Commit
2 years ago
fix get_keys_to_not_convert() to return correct modules for full precision inference (#25105) * add test for `get_keys_to_not_convert` * add minimum patch to keep mpt lm_head from 8bit quantization * add reivsion to
Author
Parents
Loading