Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Suggestions requested for training a LAMA model on higher resolution data #287

Open
rachit221195 opened this issue Jan 5, 2024 · 1 comment

Comments

@rachit221195
Copy link

Currently all LAMA models have been trained on 256x256 crops of 512x512 images.
I would like to understand what changes should be made to train a LAMA model on a bigger image resolution - maybe 512x512 crops from 1024x1024 images.

I want suggestions which can guide me on what changes in network architecture (number of upsampling, downsampling steps, number of resnet modules) can be experimented with. Apart from network architecture, are there any other changes that might be worth experimenting with.

@Abbsalehi
Copy link

Abbsalehi commented Mar 13, 2024

There are two datasets with the below sizes mentioned in the Readme file.

Places dataset: 512 by 512 images
CelebA dataset: 256 by 256 images

Once you generate masks, you can crop your 1024x1024 images to 512 or cropping using a Python code by yourself. Please let me know if you have any questions.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants