Estimating depth from a single image is a very challenging and exciting topic in computer vision with implications in several application domains. Recently proposed deep learning approaches achieve outstanding results by tackling it as an image reconstruction task and exploiting geometry constraints (e.g., epipolar geometry) to obtain supervisory signals for training. Inspired by these works and compelling results achieved by Generative Adversarial Network (GAN) on image reconstruction and generation tasks, in this paper we propose to cast unsupervised monocular depth estimation within a GAN paradigm. The generator network learns to infer depth from the reference image to generate a warped target image. At training time, the discriminator network learns to distinguish between fake images generated by the generator and target frames acquired with a stereo rig. To the best of our knowledge, our proposal is the first successful attempt to tackle monocular depth estimation with a GAN paradigm and the extensive evaluation on CityScapes and KITTI datasets confirm that it enables to improve state-of-the-art. Additionally, we highlight a major issue with data deployed by a standard evaluation protocol widely used in this field and fix this problem using a more reliable dataset recently made available by the KITTI evaluation benchmark.

Generative Adversarial Networks for unsupervised monocular depth prediction / F. Aleotti, F. Tosi, M. Poggi, S. Mattoccia. - ELETTRONICO. - (2019), pp. 337-354. (Intervento presentato al convegno 3D Reconstruction in the Wild 2018 (3DRW2018), ECCV 2018 Workshop tenutosi a Munich, Germany nel September 14, 2018) [10.1007/978-3-030-11009-3_20].

Generative Adversarial Networks for unsupervised monocular depth prediction

F. Aleotti;F. Tosi;M. Poggi;S. Mattoccia
2019

Abstract

Estimating depth from a single image is a very challenging and exciting topic in computer vision with implications in several application domains. Recently proposed deep learning approaches achieve outstanding results by tackling it as an image reconstruction task and exploiting geometry constraints (e.g., epipolar geometry) to obtain supervisory signals for training. Inspired by these works and compelling results achieved by Generative Adversarial Network (GAN) on image reconstruction and generation tasks, in this paper we propose to cast unsupervised monocular depth estimation within a GAN paradigm. The generator network learns to infer depth from the reference image to generate a warped target image. At training time, the discriminator network learns to distinguish between fake images generated by the generator and target frames acquired with a stereo rig. To the best of our knowledge, our proposal is the first successful attempt to tackle monocular depth estimation with a GAN paradigm and the extensive evaluation on CityScapes and KITTI datasets confirm that it enables to improve state-of-the-art. Additionally, we highlight a major issue with data deployed by a standard evaluation protocol widely used in this field and fix this problem using a more reliable dataset recently made available by the KITTI evaluation benchmark.
2019
Proceedings of 3D Reconstruction in the Wild 2018 (3DRW2018), in conjunction with (ECCV 2018)
337
354
Generative Adversarial Networks for unsupervised monocular depth prediction / F. Aleotti, F. Tosi, M. Poggi, S. Mattoccia. - ELETTRONICO. - (2019), pp. 337-354. (Intervento presentato al convegno 3D Reconstruction in the Wild 2018 (3DRW2018), ECCV 2018 Workshop tenutosi a Munich, Germany nel September 14, 2018) [10.1007/978-3-030-11009-3_20].
F. Aleotti, F. Tosi, M. Poggi, S. Mattoccia
File in questo prodotto:
Eventuali allegati, non sono esposti

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/653878
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 27
  • ???jsp.display-item.citation.isi??? 51
social impact