DeepSpeed
tolerating missing optimizer states for MoE [2nd attempt]
#4120
Merged

tolerating missing optimizer states for MoE [2nd attempt] #4120

clumsy
clumsy clumsy requested a review from jeffra jeffra 2 years ago
clumsy clumsy requested a review from tjruwase tjruwase 2 years ago
clumsy clumsy requested a review from mrwyattii mrwyattii 2 years ago
clumsy clumsy force pushed from 343f7ef2 to 6f1be41c 2 years ago
clumsy clumsy force pushed from 6f1be41c to f19a973f 2 years ago
clumsy
clumsy
tjruwase
tjruwase
tjruwase commented on 2023-08-17
clumsy
clumsy
clumsy clumsy force pushed from f19a973f to 6561f327 2 years ago
clumsy
clumsy commented on 2023-08-18
tjruwase
tjruwase commented on 2023-08-21
clumsy clumsy force pushed from 830d495a to 39aaad86 2 years ago
clumsy
azzhipa skipping redundant MoE optimizer state loading
b8927b8e
clumsy clumsy force pushed from 39aaad86 to b8927b8e 2 years ago
Merge branch 'master' into fix/missing_moe_optim_states
275bce6a
tjruwase
tjruwase approved these changes on 2023-08-28
clumsy
Merge branch 'master' into fix/missing_moe_optim_states
a95b97a4
clumsy
tjruwase tjruwase merged e801e6d7 into master 2 years ago
clumsy clumsy deleted the fix/missing_moe_optim_states branch 2 years ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone