Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

the input_size and out_size of big-lama #83

Open
ray0809 opened this issue Jan 17, 2022 · 4 comments
Open

the input_size and out_size of big-lama #83

ray0809 opened this issue Jan 17, 2022 · 4 comments

Comments

@ray0809
Copy link

ray0809 commented Jan 17, 2022

Hi,
Firstly, Thank you for making such a great project open source.
I found the out_size in released big-lama config.yaml is 256, was the big-lama model trained with images' size 256?

@windj007
Copy link
Collaborator

windj007 commented Jan 19, 2022

Hi! All LaMa models in the paper were trained using 256x256 crops from Places. Original resolution of images in Places is approximately 512

@windj007
Copy link
Collaborator

Feel free to reopen the issue if you have further questions

@hjq133
Copy link

hjq133 commented Mar 11, 2022

Hi, @windj007. When using Places for training, why doesn't lama scale the image to 256 before cropping, is it more meaningful than directly making 256x256 crops ?

@windj007
Copy link
Collaborator

Due to the nature of convolutions, the networks adapt to the scale of objects/textures - and perform best on exactly that scale. Inpainting in 256 is not very interesting in practice - so why optimize methods for such low resolution? Original resolution of images in Places is 512 - so we decided to keep that scale (i.e. average size of objects in pixels), but training directly in 512 is very expensive - so we used crops.

@windj007 windj007 reopened this Mar 11, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants