Depth estimation from a single image represents a fascinating, yet challenging problem with countless applications. Recent works proved that this task could be learned without direct supervision from ground truth labels leveraging image synthesis on sequences or stereo pairs. Focusing on this second case, in this paper we leverage stereo matching in order to improve monocular depth estimation. To this aim we propose monoResMatch, a novel deep architecture designed to infer depth from a single input image by synthesizing features from a different point of view, horizontally aligned with the input image, performing stereo matching between the two cues. In contrast to previous works sharing this rationale, our network is the first trained end-to-end from scratch. Moreover, we show how obtaining proxy ground truth annotation through traditional stereo algorithms, such as Semi-Global Matching, enables more accurate monocular depth estimation still countering the need for expensive depth labels by keeping a self-supervised approach. Exhaustive experimental results prove how the synergy between i) the proposed monoResMatch architecture and ii) proxy-supervision attains state-of-the-art for self-supervised monocular depth estimation. The code is publicly available at https://github.com/fabiotosi92/monoResMatch-Tensorflow.

Learning monocular depth estimation infusing traditional stereo knowledge / F. Tosi, F. Aleotti, M. Poggi, S. Mattoccia. - ELETTRONICO. - (2019), pp. 9799-9809. (Intervento presentato al convegno IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2019) tenutosi a Long Beach, CA, USA nel June 16-20 2019) [10.1109/CVPR.2019.01003].

Learning monocular depth estimation infusing traditional stereo knowledge

F. Tosi;F. Aleotti;M. Poggi;S. Mattoccia
2019

Abstract

Depth estimation from a single image represents a fascinating, yet challenging problem with countless applications. Recent works proved that this task could be learned without direct supervision from ground truth labels leveraging image synthesis on sequences or stereo pairs. Focusing on this second case, in this paper we leverage stereo matching in order to improve monocular depth estimation. To this aim we propose monoResMatch, a novel deep architecture designed to infer depth from a single input image by synthesizing features from a different point of view, horizontally aligned with the input image, performing stereo matching between the two cues. In contrast to previous works sharing this rationale, our network is the first trained end-to-end from scratch. Moreover, we show how obtaining proxy ground truth annotation through traditional stereo algorithms, such as Semi-Global Matching, enables more accurate monocular depth estimation still countering the need for expensive depth labels by keeping a self-supervised approach. Exhaustive experimental results prove how the synergy between i) the proposed monoResMatch architecture and ii) proxy-supervision attains state-of-the-art for self-supervised monocular depth estimation. The code is publicly available at https://github.com/fabiotosi92/monoResMatch-Tensorflow.
2019
2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
9799
9809
Learning monocular depth estimation infusing traditional stereo knowledge / F. Tosi, F. Aleotti, M. Poggi, S. Mattoccia. - ELETTRONICO. - (2019), pp. 9799-9809. (Intervento presentato al convegno IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2019) tenutosi a Long Beach, CA, USA nel June 16-20 2019) [10.1109/CVPR.2019.01003].
F. Tosi, F. Aleotti, M. Poggi, S. Mattoccia
File in questo prodotto:
File Dimensione Formato  
Tosi_Learning_Monocular_Depth_Estimation_Infusing_Traditional_Stereo_Knowledge_CVPR_2019_paper.pdf

accesso aperto

Tipo: Postprint
Licenza: Licenza per accesso libero gratuito
Dimensione 2.16 MB
Formato Adobe PDF
2.16 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/710383
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 166
  • ???jsp.display-item.citation.isi??? 114
social impact