The rapid integration of machine learning (ML) predictors into in silico medicine has revolutionized the estimation of quantities of interest that are otherwise challenging to measure directly. However, the credibility of these predictors is critical, especially when they inform high-stakes healthcare decisions. This position paper presents a consensus statement developed by experts within the In Silico World Community of Practice. We outline 12 key statements forming the theoretical foundation for evaluating the credibility of ML predictors, emphasizing the necessity of causal knowledge, rigorous error quantification, and robustness to biases. By comparing ML predictors with biophysical models, we highlight unique challenges associated with implicit causal knowledge and propose strategies to ensure reliability and applicability. Our recommendations aim to guide researchers, developers, and regulators in the rigorous assessment and deployment of ML predictors in clinical and biomedical contexts.
Aldieri, A., Gamage, T.P.B., Mattina, A.A.L., Loewe, A., Pappalardo, F., Viceconti, M. (2025). Consensus statement on the credibility assessment of machine learning predictors. BRIEFINGS IN BIOINFORMATICS, 26(2), 1-7 [10.1093/bib/bbaf100].
Consensus statement on the credibility assessment of machine learning predictors
Viceconti M.
2025
Abstract
The rapid integration of machine learning (ML) predictors into in silico medicine has revolutionized the estimation of quantities of interest that are otherwise challenging to measure directly. However, the credibility of these predictors is critical, especially when they inform high-stakes healthcare decisions. This position paper presents a consensus statement developed by experts within the In Silico World Community of Practice. We outline 12 key statements forming the theoretical foundation for evaluating the credibility of ML predictors, emphasizing the necessity of causal knowledge, rigorous error quantification, and robustness to biases. By comparing ML predictors with biophysical models, we highlight unique challenges associated with implicit causal knowledge and propose strategies to ensure reliability and applicability. Our recommendations aim to guide researchers, developers, and regulators in the rigorous assessment and deployment of ML predictors in clinical and biomedical contexts.| File | Dimensione | Formato | |
|---|---|---|---|
|
bbaf100.pdf
accesso aperto
Tipo:
Versione (PDF) editoriale / Version Of Record
Licenza:
Licenza per Accesso Aperto. Creative Commons Attribuzione - Non commerciale (CCBYNC)
Dimensione
306.6 kB
Formato
Adobe PDF
|
306.6 kB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


