Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Influence: Amount of training data and data augmentation via Detectron2 #71

Closed
Alexander-Tack opened this issue Dec 22, 2021 · 3 comments

Comments

@Alexander-Tack
Copy link

Hi,

Thank you for sharing LaMa! The inpainting quality is really impressive!

I was wondering:

  1. Comparing "LaMa-Fourier" with "Big LaMa-Fourier": How much did the larger training data (4.5M images from the Places-Challenge dataset) contribute to the improved quality of Big LaMa-Fourier? Do you think that similar results could have also been achieved for Big LaMa-Fourier with less data?

  2. You have proposed a sophisticated approach for data augmentation. How much did the training and the inference benefit from data augmentation using segmentation masks from Detectron2?

Best wishes,
Alex

@windj007
Copy link
Collaborator

Hi! Sorry for the late reply!

How much did the larger training data (4.5M images from the Places-Challenge dataset) contribute to the improved quality of Big LaMa-Fourier?

Larger data helps, but significantly less than larger model size and other training tricks (SegmPL, large masks)

Do you think that similar results could have also been achieved for Big LaMa-Fourier with less data?

Less data = less quality, but not dramatically - reducing model size or removing SegmPL or using smaller training masks would hurt more

How much did the training and the inference benefit from data augmentation using segmentation masks from Detectron2?

We do not use segmentation masks from Detectron for training. We tried it in the very beginning of the project, but faced technical issues (slow, gpu memory consumption, cuda-reinitialization limitation) - so we could not use segmentation-based mask generation effectively during training. It is there just because we forgot to remove it when preparing a public code release. Please note that segm_proba: 0 in all the data configs.

@windj007
Copy link
Collaborator

In our experience, mask widths matter much more than mask shapes

@Alexander-Tack
Copy link
Author

Hi Roman,
awesome!! Thank you for sharing your experience! This really helps a lot and confirms my preliminary findings :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants