The source codes are based on matconvnet.
Create a new folder 'data' and then create 3 subfolders 'deploynet', 'modelnet40' and 'models' in it.
Download the pretrained net from
Save the downloaded net file 'net-deployed.mat' in \data\deploynet\
For a demo, simply implement the following lines in MATLAB command window
setup; % If the dependency libraries are not compiled on your ocmputer, please compile them first through the setup command (see setup.m for details);
sa1 = cnn_output_fast('human.off');
You can also input a scene:
sa2 = cnn_output_fast('skateboard.off');
Note that the program accepts OFF or OBJ files for the input 3D models.
If you want to train it using the modelnet40 dataset, please follow the steps below:
- Download the baseline VGG19 net from the official website of matconvnet or
Save it in \data\models\
Download the ModelNet40 dataset and render all of the models using render_views_of_all_meshes_in_a_folder.m and create_data.m in the utils folder. Then save the rendered images (including 40 folders) in \data\modelnet40.
Run run_train.m where the get_imdb function only need to be implemented at the first time of training.
Please cite our paper if you use the codes:
Ran Song, Yonghuai Liu, Paul L. Rosin. Mesh Saliency via Weakly Supervised Classification-for-Saliency CNN. IEEE Transactions on Visualization and Computer Graphics, 13 pages, 2019.