Skip to content
Unstructured multi-view depth estimation using mask-based multiplane representation
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Type Name Latest commit message Commit time
Failed to load latest commit information.


Yuxin Hou · Arno Solin · Juho Kannala

Codes for the paper:

  • Yuxin Hou, Arno Solin, and Juho Kannala (2019). Unstructured multi-view depth estimation using mask-based multiplane representation. Scandinavian Conference on Image Analysis (SCIA). [preprint on arXiv]


MaskMVS is a method for depth estimation for unstructured multi-view image-pose pairs. In the plane-sweep procedure, the depth planes are sampled by histogram matching that ensures covering the depth range of interest. Unlike other plane-sweep methods, we do not rely on a cost metric to explicitly build the cost volume, but instead infer a multiplane mask representation which regularizes the learning. Compared to many previous approaches, we show that our method is lightweight and generalizes well without requiring excessive training. See the paper for further details.


Tested with:

  • Python3
  • Numpy
  • Pytorch 0.3.0
  • CUDA 9 (You can also run without CUDA, but then you need to remove all .cuda() in codes)
  • opencv
  • imageio (with freeimage plugin)

To install imageio, run conda install -c conda-forge imageio or pip install imageio. To install the freeimage plugin, run the following Python script once:

import imageio

Download pretrained models

We provide our pretrained models of our MaskNet and DispNet to run the example code. Please download the models via the link

Run the example

  1. Put both the model masknet_model_best.pth.tar and the model dispnet_model_best.pth.tar under the project folder.
  2. Then just run the jupyter notebook file example.ipynb


This software is distributed under the GNU General Public License (version 3 or later); please refer to the file LICENSE, included with the software, for details.

You can’t perform that action at this time.