Skip to content

Latest commit

 

History

History
34 lines (19 loc) · 1.67 KB

dataset.md

File metadata and controls

34 lines (19 loc) · 1.67 KB

Dataset

The dataset used for this model was taken from here. The dataset contains 400,000 images of backgrounds, background-foregrounds and their corresponding masks and depth maps each. For more info on the dataset, please go to this link.

Preview

The dataset contains four types of images

Background

bg

Background-Foreground

bg_fg

Background-Foreground Mask

bg_fg_mask

Background-Foreground Depth Map

bg_fg_depth_map

Preprocessing

  • The input images (background and background-foreground) were normalized according to the values given on the dataset page.
  • No preprocessing was done on the output images except converting them into torch.Tensor type and keeping their values within the range [0, 1].
  • There was no point in applying any physical data augmentation techniques on the images as it would distort them from their corresponding labels which were not augmented.
  • So, the only option left was to use photometric augmentations. I tried HueSaturationValue and RandomContrast augmentations from the albumentations package. The code for augmentation can be seen here.

Data Loading

  • The dataset is huge, so it is not possible to load the entire dataset into memory at once. So only the images names are indexed and they are fetched on need basis. The code for data loading can be found here