Bias or spurious correlations in image backgrounds can impact neural networks, causing shortcut learning (Clever Hans Effect) and hampering generalization to real-world data. ISNet, a recently introduced architecture, proposed the optimization of Layer-Wise Relevance Propagation (LRP, an explanation technique) heatmaps, to mitigate the influence of backgrounds on deep classifiers. However, ISNet's training time scales linearly with the number of classes in an application. Here, we propose reformulated architectures, dubbed Faster ISNets, whose training time becomes independent from this number. Additionally, we introduce a concise and model-agnostic LRP implementation, LRP-Flex, which can readily explain arbitrary DNN architectures, or convert them into Faster ISNets. We challenge the proposed architectures using synthetic background bias, and COVID-19 detection in chest X-rays, an application that commonly presents background bias. The networks hindered background attention and shortcut learning, surpassing multiple state-of-the-art models on out-of-distribution test datasets. Representing a potentially massive training speed improvement over ISNet, the proposed architectures introduce LRP optimization into a gamut of applications that the original ISNet model cannot feasibly handle. Code for the Faster ISNet and LRP-Flex is available at https://github.com/PedroRASB/FasterISNet.

Bassi, P.R.A.S., Decherchi, S., Cavalli, A. (2024). Faster ISNet for Background Bias Mitigation on Deep Neural Networks. IEEE ACCESS, 12, 155151-155167 [10.1109/access.2024.3461773].

Faster ISNet for Background Bias Mitigation on Deep Neural Networks

Cavalli, Andrea
2024

Abstract

Bias or spurious correlations in image backgrounds can impact neural networks, causing shortcut learning (Clever Hans Effect) and hampering generalization to real-world data. ISNet, a recently introduced architecture, proposed the optimization of Layer-Wise Relevance Propagation (LRP, an explanation technique) heatmaps, to mitigate the influence of backgrounds on deep classifiers. However, ISNet's training time scales linearly with the number of classes in an application. Here, we propose reformulated architectures, dubbed Faster ISNets, whose training time becomes independent from this number. Additionally, we introduce a concise and model-agnostic LRP implementation, LRP-Flex, which can readily explain arbitrary DNN architectures, or convert them into Faster ISNets. We challenge the proposed architectures using synthetic background bias, and COVID-19 detection in chest X-rays, an application that commonly presents background bias. The networks hindered background attention and shortcut learning, surpassing multiple state-of-the-art models on out-of-distribution test datasets. Representing a potentially massive training speed improvement over ISNet, the proposed architectures introduce LRP optimization into a gamut of applications that the original ISNet model cannot feasibly handle. Code for the Faster ISNet and LRP-Flex is available at https://github.com/PedroRASB/FasterISNet.
2024
Bassi, P.R.A.S., Decherchi, S., Cavalli, A. (2024). Faster ISNet for Background Bias Mitigation on Deep Neural Networks. IEEE ACCESS, 12, 155151-155167 [10.1109/access.2024.3461773].
Bassi, Pedro R. A. S.; Decherchi, Sergio; Cavalli, Andrea
File in questo prodotto:
File Dimensione Formato  
Faster_ISNet_for_Background_Bias_Mitigation_on_Deep_Neural_Networks.pdf

accesso aperto

Tipo: Versione (PDF) editoriale / Version Of Record
Licenza: Licenza per Accesso Aperto. Creative Commons Attribuzione (CCBY)
Dimensione 2 MB
Formato Adobe PDF
2 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/1009141
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact