diffusers
b9d52fca
- [train_lcm_distill_lora_sdxl.py] Fix the LR schedulers when num_train_epochs is passed in a distributed training env (#8446)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Hide Minimap (CTRL+M)
Commit
1 year ago
[train_lcm_distill_lora_sdxl.py] Fix the LR schedulers when num_train_epochs is passed in a distributed training env (#8446) fix num_train_epochs Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
References
#8446 - [train_lcm_distill_lora_sdxl.py] Fix the LR schedulers when num_train_epochs is passed in a distributed training env
Author
rootonchair
Parents
2ada094b
Files
1
examples/consistency_distillation
train_lcm_distill_lora_sdxl.py
Loading