Image retrieval usually faces scale-variance issues as the amount of image data is rapidly increasing, which calls for more accurate retrieval technology. Besides, existing methods usually treat pair-image similarity as a binary value which indicates whether two images share either at least one common label or none of shared labels. However, such similarity definition cannot truly describe the similarity ranking for different numbers of common labels when handling the multi-label image retrieval problem. In this paper, a Feature Disentangling and Reciprocal Learning (FDRL) method is introduced with label-guided similarity to solve the above multi-label image retrieval problem. Multi-scale features are first extracted by BNInception network and then disentangled to the corresponding high- and low-correlation features under the guidance of estimated global correlations. After that, the disentangled features are combined through a reciprocal learning approach to enhance the feature representation ability. Final hash codes are learned based on the global features derived from BNInception network and the combined features generated by reciprocal learning. The whole network is optimized by the proposed label-guided similarity loss function which aims to simultaneously preserve absolute similarity for hard image pairs and relative similarity for soft image pairs. Experimental results on three public benchmark datasets demonstrate that the proposed method outperforms current state-of-the-art techniques. The code is online here: ‘https://github.com/Yong-DAI/FDRL’.

Feature disentangling and reciprocal learning with label-guided similarity for multi-label image retrieval

Luigi Di Stefano
2022

Abstract

Image retrieval usually faces scale-variance issues as the amount of image data is rapidly increasing, which calls for more accurate retrieval technology. Besides, existing methods usually treat pair-image similarity as a binary value which indicates whether two images share either at least one common label or none of shared labels. However, such similarity definition cannot truly describe the similarity ranking for different numbers of common labels when handling the multi-label image retrieval problem. In this paper, a Feature Disentangling and Reciprocal Learning (FDRL) method is introduced with label-guided similarity to solve the above multi-label image retrieval problem. Multi-scale features are first extracted by BNInception network and then disentangled to the corresponding high- and low-correlation features under the guidance of estimated global correlations. After that, the disentangled features are combined through a reciprocal learning approach to enhance the feature representation ability. Final hash codes are learned based on the global features derived from BNInception network and the combined features generated by reciprocal learning. The whole network is optimized by the proposed label-guided similarity loss function which aims to simultaneously preserve absolute similarity for hard image pairs and relative similarity for soft image pairs. Experimental results on three public benchmark datasets demonstrate that the proposed method outperforms current state-of-the-art techniques. The code is online here: ‘https://github.com/Yong-DAI/FDRL’.
2022
Dai Y.; Song W.; Li Y.; Luigi Di Stefano
File in questo prodotto:
Eventuali allegati, non sono esposti

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/903380
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? ND
social impact