-
Notifications
You must be signed in to change notification settings - Fork 140
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question for Finetuning #36
Comments
Hi. Without training settings as well as the figures for comparison, it is hard to tell the problem. Maybe the test data distribution is very different from the validation one. |
Thanks for your very quick response. I got similar results as this issue #26. The following zip file contains the training config. I just follow your suggestion to reset the model path. And here is the results (right: result; left: input). The bottom one is a sample from validation set and it is also has been used for training. The above one did has some difference with the input, but the result is very far from the model you provided. Thanks! |
Hi. |
Thanks! By the way, how long did you spend for the fine-tuning stage, maybe just around 24 hours? |
For several days. The longer the better. One week should be enough. |
So the provided model (stable_000117.ckpt) is achieved by fine-tuned for around a week? I thought 117 epoch model do not need to spend for so much time to get. |
The speed of A100 is more than 2x than V100, and I do not remember the exact training time of the 512 model. |
Thank you so much. I will keep running the experiments and let you know if there is any update. |
Dear author, |
Dear author, |
Thanks a lot! |
I currently only conducted the fine-tuning experiments and haven't trained the CFW since I found that my fine-tuning result is not good enough to train the CFW. How about your fine-tuning results? For training the CFW, I saw there is a issue #28 that might can help you. |
I think they are making sense but yea different than the author's results hmm, maybe can try fix the seeds for inference to check if the finetuning is successful? https://huggingface.co/docs/diffusers/using-diffusers/reproducibility |
I'm wondering if it is possible to share one example that you generated using only the fine-tuned model? Thanks a lot! |
Dear author,
I tried to reproduce your work and I'm currently want to validate the result generated by the fine-tuned sd model. I used only one A100 gpu card for finetuning (currently at around 130 epoch) and test with the script "sr_val_ddpm_text_T_vqganfin_old.py" by reset the ckpt path and change the dec_w to 0.0. The rest result shows almost no difference with the input. But when I saw the training log, the validation results look pretty good. Do you have any idea for this issue?
Thanks a lot!
The text was updated successfully, but these errors were encountered: