Machine unlearning (MU) is often analyzed in terms of how it can facilitate the “right to be forgotten.” In this commentary, we show that MU can support the OECD’s five principles for trustworthy AI, which are influencing AI development and regulation worldwide. This makes it a promising tool to translate AI principles into practice. We also argue that the implementation of MU is not without ethical risks. To address these concerns and amplify the positive impact of MU, we offer policy recommendations across six categories to encourage the research and uptake of this potentially highly influential new technology.
Hine, E., Novelli, C., Taddeo, M., Floridi, L. (2024). Supporting Trustworthy AI Through Machine Unlearning. SCIENCE AND ENGINEERING ETHICS, 30(5), 1-13 [10.1007/s11948-024-00500-5].
Supporting Trustworthy AI Through Machine Unlearning
Emmie Hine
Primo
;Claudio NovelliSecondo
;Luciano FloridiUltimo
2024
Abstract
Machine unlearning (MU) is often analyzed in terms of how it can facilitate the “right to be forgotten.” In this commentary, we show that MU can support the OECD’s five principles for trustworthy AI, which are influencing AI development and regulation worldwide. This makes it a promising tool to translate AI principles into practice. We also argue that the implementation of MU is not without ethical risks. To address these concerns and amplify the positive impact of MU, we offer policy recommendations across six categories to encourage the research and uptake of this potentially highly influential new technology.File | Dimensione | Formato | |
---|---|---|---|
s11948-024-00500-5.pdf
accesso aperto
Tipo:
Versione (PDF) editoriale
Licenza:
Licenza per Accesso Aperto. Creative Commons Attribuzione (CCBY)
Dimensione
700.64 kB
Formato
Adobe PDF
|
700.64 kB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.