-
Notifications
You must be signed in to change notification settings - Fork 35
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training sf:1 (debluring) #22
Comments
You only need to rewrite the class of dataset. Please refer to this bicubic dataset. |
I have changed the Dataset implemetation. However I am getting the following issue in Unet mismatch when I run it.
I am trying test it on input sizes GT= 64x64 and LQ= 64x64 Here is my implementation of the Dataset:
And here is the test config I am running it with:
|
Hi @zsyOAOA would you be able to provide any guidance on where to look in the code to enable training of same size inputs and outputs? |
@cyprian For the same size inputs and outputs, try to set the diffusion.params.sf = 1 and model.params.in_channels=51 (51=3+48) in the config file. Note that the vqgan downsamples the input 4 times, we thus unshuffles the input to the dimension of H/4 x W/4 x 48. |
@zsyOAOA the model.params.in_channels=51 still gave me tensor size mismatch on the vqgan downsampling. I went with a different approach of encoding both LQ and GT via VQGAN and that seams to work, but I am loosing some of the fidelity with this 4x downsampling. Is there a VGGAN model that does 2x downsampling? I could not find where you got the autoencoder_vq_f4.pth from. |
I haven't a VQGAN model with 2x downsampling. For the 4x model, please find it in this link. |
@zsyOAOA I would like to train my own VQGAN model to create better image representation for my image class. Can you tell me how you trained your VQGAN? |
Thank you for a quick reply. In the repo I don't see the weights for the VQGAN f4, that you are using in your code. I am trying to find config for that training so that I could just use my own Dataset. Did you download your weights from that repo? @zsyOAOA |
The checkpoint I used is extracted from the latent diffusion model for image super-resolution. |
Ok. So if I understand correctly you extracted just extracted the first_stage_model from that super resolution checkpoint? @zsyOAOA (BTW, I really appreciate your help) |
Yes. @cyprian |
Thank you for providing your code. I already tested the super resolution, and it great. Is it possible to adopt the config to do debluring on lq:256 x gt:256 so without any super resolution? What would I have to change.
The text was updated successfully, but these errors were encountered: