transformers
75e39856 - Fix Break change of AWQ FusedModules due to Attention Refactor (#41909)

Commit
158 days ago
Fix Break change of AWQ FusedModules due to Attention Refactor (#41909) * fix awq bc due to attention refactor * feat: support more rope_types for awq fusion * feat: add test for llama3 * fix ruff format * propagate changes in modeling_llama
Author
Parents
Loading