Generating Large Scale Image Datasets from 3D CAD Models
Switch branches/tags
Nothing to show
Clone or download
Latest commit 7febc94 Jun 28, 2017
Permalink
Failed to load latest commit information.
LICENSE add readme and license Jun 3, 2015
README.md Update README.md Jun 28, 2017

README.md

Generating Large Scale Image Datasets from 3D CAD Models

Published at CVPR'15 Workshop on The Future of Datasets in Vision

[Paper] (http://vision.cs.uml.edu/pubs/cvpr2015_workshop_virtual_dataset.pdf)

Citation

@inproceedings{baochen15fdv,
    Author = {Baochen Sun and Xingchao Peng and Kate Saenko},
    Title = {Generating Large Scale Image Datasets from 3D CAD Models},
    Booktitle = {CVPR Workshop},
    Year = {2015}
}

Dataset One

Dataset

Dataset for "From Virtual to Reality: Fast Adaptation of Virtual Object Detectors to Real Domains" published in British Machine Vision Conference(BMVC) 2014.

Paper

Extended Abstract

Poster

Citation

@inproceedings{baochen14BMVC,
    Author = {Baochen Sun and Kate Saenko},
    Title = {From Virtual to Reality: Fast Adaptation of Virtual Object Detectors to Real Domains},
    Booktitle = {British Machine Vision Conference},
    Year = {2014}
}

Dataset Two

Dataset

Dataset for "What Do Deep CNNs Learn About Object?" published in ICLR'15 workshop.

Workshop Paper

Arxiv Full Paper

Poster: TBD

Citation

@inproceedings{peng2015learning,
  title={Learning deep object detectors from 3d models},
  author={Peng, Xingchao and Sun, Baochen and Ali, Karim and Saenko, Kate},
  booktitle={Proceedings of the IEEE International Conference on Computer Vision},
  pages={1278--1286},
  year={2015}
}

FAQ

Please go to the GitHub repo (https://github.com/UMassLowell-Vision-Group/From-Virtual-to-Reality) for source code, 3d Models, datasets of Dataset One. The source code of Dataset Two is similar to Dataset One and more explainations are in the following. The following explanations are based on Dataset One.

The source code is available in 'code' folder. The render.ms is the file for the rendering part. All the rendered images (with annotation) are in these two folders: 'virtual' and 'virtual_gray'. The 3D models is in the '3d_models' folder. Basically, here are the brief steps:

  1. Download the 3D models: https://drive.google.com/file/d/0B9ca-zqjEJYxZE5hVTk3UU9hcG8/view?usp=sharing  (In case that a password is needed to unzip the file, input 'iccv2015')

  2. Run render.ms in 3ds Max (the software we used to generate the dataset)

The are also limited comments in render.ms which might help you understand the code.

To generate more realistic images (as in the 'What Do Deep CNNs Learn About Object' paper), you may need to specify different background and texture for different category/3d model. Then you need to change the 'images_bg' and 'images_texture' in the render.ms file to point to different background and textures for different category/3d model.