diffusers
a11b0f83 - Fix: training resume from fp16 for SDXL Consistency Distillation (#6840)

Commit
1 year ago
Fix: training resume from fp16 for SDXL Consistency Distillation (#6840) * Fix: training resume from fp16 for lcm distill lora sdxl * Fix coding quality - run linter * Fix 1 - shift mixed precision cast before optimizer * Fix 2 - State dict errors by removing load_lora_into_unet * Update train_lcm_distill_lora_sdxl.py - Revert default cache dir to None --------- Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
Author
Parents
Loading