Guided depth super-resolution aims at using a low-resolution depth map and an associated high-resolution RGB image to recover a high-resolution depth map. However, restoring precise and sharp edges near depth discontinuities and fine structures is still challenging for state-of-the-art methods. To alleviate this issue, we propose a novel multi-stage depth super-resolution network, which progressively reconstructs HR depth maps from explicit and implicit high-frequency information. We introduce an efficient transformer to obtain explicit high-frequency information. The shape bias and global context of the transformer allow our model to focus on high-frequency details between objects, i.e., depth discontinuities, rather than texture within objects. Furthermore, we project the input color images into the frequency domain for additional implicit high-frequency cues extraction. Finally, to incorporate the structural details, we develop a fusion strategy that combines depth features and high-frequency information in the multi-stage-scale framework. Exhaustive experiments on the main benchmarks show that our approach establishes a new state-of-the-art. Code will be publicly available at https://github.com/wudiqx106/DSR-EI.

Qiao, X., Ge, C., Zhang, Y., Zhou, Y., Tosi, F., Poggi, M., et al. (2023). Depth super-resolution from explicit and implicit high-frequency features. COMPUTER VISION AND IMAGE UNDERSTANDING, 237, 1-12 [10.1016/j.cviu.2023.103841].

Depth super-resolution from explicit and implicit high-frequency features

Zhang, Youmin;Tosi, Fabio;Poggi, Matteo;Mattoccia, Stefano
2023

Abstract

Guided depth super-resolution aims at using a low-resolution depth map and an associated high-resolution RGB image to recover a high-resolution depth map. However, restoring precise and sharp edges near depth discontinuities and fine structures is still challenging for state-of-the-art methods. To alleviate this issue, we propose a novel multi-stage depth super-resolution network, which progressively reconstructs HR depth maps from explicit and implicit high-frequency information. We introduce an efficient transformer to obtain explicit high-frequency information. The shape bias and global context of the transformer allow our model to focus on high-frequency details between objects, i.e., depth discontinuities, rather than texture within objects. Furthermore, we project the input color images into the frequency domain for additional implicit high-frequency cues extraction. Finally, to incorporate the structural details, we develop a fusion strategy that combines depth features and high-frequency information in the multi-stage-scale framework. Exhaustive experiments on the main benchmarks show that our approach establishes a new state-of-the-art. Code will be publicly available at https://github.com/wudiqx106/DSR-EI.
2023
Qiao, X., Ge, C., Zhang, Y., Zhou, Y., Tosi, F., Poggi, M., et al. (2023). Depth super-resolution from explicit and implicit high-frequency features. COMPUTER VISION AND IMAGE UNDERSTANDING, 237, 1-12 [10.1016/j.cviu.2023.103841].
Qiao, Xin; Ge, Chenyang; Zhang, Youmin; Zhou, Yanhui; Tosi, Fabio; Poggi, Matteo; Mattoccia, Stefano
File in questo prodotto:
File Dimensione Formato  
2303.09307.pdf

Open Access dal 24/09/2024

Tipo: Postprint
Licenza: Licenza per Accesso Aperto. Creative Commons Attribuzione - Non commerciale - Non opere derivate (CCBYNCND)
Dimensione 7.92 MB
Formato Adobe PDF
7.92 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/957752
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? 2
social impact