Describe the bug
|
transformer.to(accelerator.device, dtype=weight_dtype) |
the transformer before accelerator.prepare should not convert to fp16 in mix precision;
it will broken the grad precision and result in worse training result
Reproduction
see the "/examples/research_projects/sd3_lora_colab/train_dreambooth_lora_sd3_miniature.py"
Logs
System Info
diffusers 0.31.0
Who can help?
No response
Describe the bug
diffusers/examples/research_projects/sd3_lora_colab/train_dreambooth_lora_sd3_miniature.py
Line 717 in e8aacda
the transformer before accelerator.prepare should not convert to fp16 in mix precision;
it will broken the grad precision and result in worse training result
Reproduction
see the "/examples/research_projects/sd3_lora_colab/train_dreambooth_lora_sd3_miniature.py"
Logs
System Info
diffusers 0.31.0
Who can help?
No response