DeMoN: Depth and Motion Network for Learning Monocular Stereo
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017
Abstract: In this paper we formulate structure from motion as a learning problem. We train a convolutional network end-to-end to compute depth and camera motion from successive, unconstrained image pairs. The architecture is composed of multiple stacked encoder-decoder networks, the core part being an iterative network that is able to improve its own predictions. The network estimates not only depth and motion, but additionally surface normals, optical flow between the images and confidence of the matching. A crucial component of the approach is a training loss based on spatial relative differences. Compared to traditional two-frame structure from motion methods, results are more accurate and more robust. In contrast to the popular depth-from-single-image networks, DeMoN learns the concept of matching and, thus, better generalizes to structures not seen during training.
Publisher's link
Project pageSee also
BibTex reference
@InProceedings{UZUMIDB17, author = "B. Ummenhofer and H. Zhou and J. Uhrig and N. Mayer and E. Ilg and A. Dosovitskiy and T. Brox", title = "DeMoN: Depth and Motion Network for Learning Monocular Stereo", booktitle = "IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", month = " ", year = "2017", url = "http://lmbweb.informatik.uni-freiburg.de/Publications/2017/UZUMIDB17" }