DeepSpeed
8d98e171 - Enable mixtral 8x7b autotp (#5257)

Comment changes are shownComment changes are hidden
Commit
1 year ago
Enable mixtral 8x7b autotp (#5257) This PR aims to enable mixtral 8x7b (MoE model) autotp. Co-authored-by: Logan Adams <114770087+loadams@users.noreply.github.com>
Author
Parents
  • deepspeed/module_inject
    • File
      auto_tp.py