Skip to content

[ECCV 2022] Neural-Sim: Learning to Generate Training Data with NeRF

License

MIT, MIT licenses found

Licenses found

MIT
LICENSE
MIT
LICENSE.md
Notifications You must be signed in to change notification settings

gyhandy/Neural-Sim-NeRF

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Neural-Sim: Learning to Generate Training Data with NeRF

[ECCV 2022] Neural-Sim: Learning to Generate Training Data with NeRF

Code are actively updating, thanks!

Overview

The code is for On-demand synthetic data generation: Given a target task and a test dataset, our approach “Neural-sim” generates data on-demand using a fully differentiable synthetic data generation pipeline which maximises accuracy for the target task.

Editor

Neural-Sim pipeline: Our pipeline finds the optimal parameters for generating views from a trained neural renderer (NeRF) to use as training data for object detection. The objective is to find the optimal NeRF rendering parameters ψ that can generate synthetic training data Dtrain, such that the model (RetinaNet, in our experiments) trained on Dtrain, maximizes accuracy on a downstream task represented by the validation set Dval

Editor

1 Installation

Start by cloning the repo:

git clone https://github.com/gyhandy/Neural-Sim-NeRF.git

1 install the requirement of nerf-pytorch

pip install -r requirements.txt

2 install detectorn2

2 NeRF models and dataset

Quick start

For quick start, you could download our pretrained NeRF models and created sample dataset with BlenderProc here. Then unzip it and place in .logs. (Note: if not download automatically, please right click, copy the link and open in a new tab.)

Train your own NeRF model with BlenderProc

(1) Generate Bop format images with BlenderProc

  • Follow the Installation instruction of BlenderProc

  • Download BOP dataset object toolkit used in BlenderProc. For instance, to download the YCB-V dataset toolkit, please download the "Base archive" and "Object models", two zip files. Then unzip ycbv_base.zip get the ycbv folder, unzip ycbv_models.zip get the models folder, move the models folder into ycbv folder. The path may look like this:

-BOP
--bop_toolkit
--ycbv_models
--ycbv
---models

Note: It would be better to create a new virtual environment for Blenderproc synthesis.

example command

python run.py examples/camera_sampling/config.yaml /PATH/OF/BOP/ ycbv /PATH/OF/BOP/bop_toolkit/ OUTPUT/PATH

(2) Process synthesized images to be admitted by nerf (OPENCV --> OPENGL)

if use BlenderProc synthesized image, please use

python data_generation-Blender.py

if use LatentFusion read BOP format data, please use

python data_generation-LINEMOD.py

(3) You could train NeRF with instructions NeRF-pytorch

3 Neural_Sim Bilelve optimization pipeline

cd ./optimization

Please use the neural-sim_main.py to run the end-to-end pipeline. E.g.,

python neural_sim_main.py --config ../configs/nerf_param_ycbv_general.txt --object_id 2 --expname exp_ycb_synthetic --psi_pose_cats_mode 5 --test_distribution 'one_1'

'--config' indicates the NeRF parameter

'--object_id' indicates the optimized ycbv object id, here is cheese box

'--expname' indicates the name of experiment

'--psi_pose_cats_mode' indicates the bin number of starting pose distribution during training

'--test_distribution' indicates the bin number of test pose distribution

Contact / Cite

If you use (part of) our code or find our work helpful, please consider citing

@article{ge2022neural,
  title={Neural-Sim: Learning to Generate Training Data with NeRF},
  author={Ge, Yunhao and Behl, Harkirat and Xu, Jiashu and Gunasekar, Suriya and Joshi, Neel and Song, Yale and Wang, Xin and Itti, Laurent and Vineet, Vibhav},
  journal={arXiv preprint arXiv:2207.11368},
  year={2022}
}

About

[ECCV 2022] Neural-Sim: Learning to Generate Training Data with NeRF

Resources

License

MIT, MIT licenses found

Licenses found

MIT
LICENSE
MIT
LICENSE.md

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages