Machine learning (ML) may be oblivious to human bias but it is not immune to its perpetuation. Marginalization and iniquitous group representation are often traceable in the very data used for training and may be reflected or even enhanced by the learning models. In the present work, we aim to clarify the role played by data geometry in the emergence of ML bias. We introduce an exactly solvable high-dimensional model of data imbalance, where parametric control over the many bias-inducing factors allows for an extensive exploration of the bias inheritance mechanism. Through the tools of statistical physics, we analytically characterize the typical properties of learning models trained in this synthetic framework and obtain exact predictions for the observables that are commonly employed for fairness assessment. Simplifying the nature of the problem to its minimal components, we can retrace and unpack typical unfairness behavior observed on real-world datasets. Finally, we focus on the effectiveness of bias mitigation strategies, first by considering a loss-reweighing scheme that allows for an implicit minimization of different unfairness metrics and a quantification of the incompatibilities between existing fairness criteria. Then, we propose a mitigation strategy based on a matched inference setting that entails the introduction of coupled learning models. Our theoretical analysis of this approach shows that the coupled strategy can strike superior fairness-accuracy trade-offs.

Sarao Mannelli, S., Gerace, F., Rostamzadeh, N., Saglietti, L. (2025). Bias-inducing geometries: An exactly solvable data model with fairness implications. PHYSICAL REVIEW. E, 112(2-2), 1-10 [10.1103/nlfl-35t6].

Bias-inducing geometries: An exactly solvable data model with fairness implications

Gerace F.;
2025

Abstract

Machine learning (ML) may be oblivious to human bias but it is not immune to its perpetuation. Marginalization and iniquitous group representation are often traceable in the very data used for training and may be reflected or even enhanced by the learning models. In the present work, we aim to clarify the role played by data geometry in the emergence of ML bias. We introduce an exactly solvable high-dimensional model of data imbalance, where parametric control over the many bias-inducing factors allows for an extensive exploration of the bias inheritance mechanism. Through the tools of statistical physics, we analytically characterize the typical properties of learning models trained in this synthetic framework and obtain exact predictions for the observables that are commonly employed for fairness assessment. Simplifying the nature of the problem to its minimal components, we can retrace and unpack typical unfairness behavior observed on real-world datasets. Finally, we focus on the effectiveness of bias mitigation strategies, first by considering a loss-reweighing scheme that allows for an implicit minimization of different unfairness metrics and a quantification of the incompatibilities between existing fairness criteria. Then, we propose a mitigation strategy based on a matched inference setting that entails the introduction of coupled learning models. Our theoretical analysis of this approach shows that the coupled strategy can strike superior fairness-accuracy trade-offs.
2025
Sarao Mannelli, S., Gerace, F., Rostamzadeh, N., Saglietti, L. (2025). Bias-inducing geometries: An exactly solvable data model with fairness implications. PHYSICAL REVIEW. E, 112(2-2), 1-10 [10.1103/nlfl-35t6].
Sarao Mannelli, S.; Gerace, F.; Rostamzadeh, N.; Saglietti, L.
File in questo prodotto:
Eventuali allegati, non sono esposti

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/1028254
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact