Fix: training resume from fp16 for SDXL Consistency Distillation (#6840)
* Fix: training resume from fp16 for lcm distill lora sdxl
* Fix coding quality - run linter
* Fix 1 - shift mixed precision cast before optimizer
* Fix 2 - State dict errors by removing load_lora_into_unet
* Update train_lcm_distill_lora_sdxl.py - Revert default cache dir to None
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>