Background: Prognostic machine learning research in multiple sclerosis has been mainly focusing on black-box models predicting whether a patients’ disability will progress in a fixed number of years. However, as this is a binary yes/no question, it cannot take individual disease severity into account. Therefore, in this work we propose to model the time to disease progression instead. Additionally, we use explainable machine learning techniques to make the model outputs more interpretable. Methods: A preprocessed subset of 29,201 patients of the international data registry MSBase was used. Disability was assessed in terms of the Expanded Disability Status Scale (EDSS). We predict the time to significant and confirmed disability progression using random survival forests, a machine learning model for survival analysis. Performance is evaluated on a time-dependent area under the receiver operating characteristic and the precision–recall curves. Importantly, predictions are then explained using SHAP and Bellatrex, two explainability toolboxes, and lead to both global (population-wide) as well as local (patient visit-specific) insights. Results: On the task of predicting progression in 2 years, the random survival forest achieves state-of-the-art performance, comparable to previous work employing a random forest. However, here the random survival forest has the added advantage of being able to predict progression over a longer time horizon, with AUROC >60% for the first 10 years after baseline. Explainability techniques further validated the model by extracting clinically valid insights from the predictions made by the model. For example, a clear decline in the per-visit probability of progression is observed in more recent years since 2012, likely reflecting globally increasing use of more effective MS therapies. Conclusion: The binary classification models found in the literature can be extended to a time-to-event setting without loss of performance, thus allowing a more comprehensive prediction of patient prognosis. Furthermore, explainability techniques proved to be key to reach a better understanding of the model and increase validation of its behaviour.

D'Hondt, R., Dedja, K., Aerts, S., Van Wijmeersch, B., Kalincik, T., Reddel, S., et al. (2025). Explainable time-to-progression predictions in multiple sclerosis. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 263(025 May:263:108624.), 8624-8646 [10.1016/j.cmpb.2025.108624].

Explainable time-to-progression predictions in multiple sclerosis

Lugaresi, Alessandra
Writing – Review & Editing
;
2025

Abstract

Background: Prognostic machine learning research in multiple sclerosis has been mainly focusing on black-box models predicting whether a patients’ disability will progress in a fixed number of years. However, as this is a binary yes/no question, it cannot take individual disease severity into account. Therefore, in this work we propose to model the time to disease progression instead. Additionally, we use explainable machine learning techniques to make the model outputs more interpretable. Methods: A preprocessed subset of 29,201 patients of the international data registry MSBase was used. Disability was assessed in terms of the Expanded Disability Status Scale (EDSS). We predict the time to significant and confirmed disability progression using random survival forests, a machine learning model for survival analysis. Performance is evaluated on a time-dependent area under the receiver operating characteristic and the precision–recall curves. Importantly, predictions are then explained using SHAP and Bellatrex, two explainability toolboxes, and lead to both global (population-wide) as well as local (patient visit-specific) insights. Results: On the task of predicting progression in 2 years, the random survival forest achieves state-of-the-art performance, comparable to previous work employing a random forest. However, here the random survival forest has the added advantage of being able to predict progression over a longer time horizon, with AUROC >60% for the first 10 years after baseline. Explainability techniques further validated the model by extracting clinically valid insights from the predictions made by the model. For example, a clear decline in the per-visit probability of progression is observed in more recent years since 2012, likely reflecting globally increasing use of more effective MS therapies. Conclusion: The binary classification models found in the literature can be extended to a time-to-event setting without loss of performance, thus allowing a more comprehensive prediction of patient prognosis. Furthermore, explainability techniques proved to be key to reach a better understanding of the model and increase validation of its behaviour.
2025
D'Hondt, R., Dedja, K., Aerts, S., Van Wijmeersch, B., Kalincik, T., Reddel, S., et al. (2025). Explainable time-to-progression predictions in multiple sclerosis. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 263(025 May:263:108624.), 8624-8646 [10.1016/j.cmpb.2025.108624].
D'Hondt, Robbe; Dedja, Klest; Aerts, Sofie; Van Wijmeersch, Bart; Kalincik, Tomas; Reddel, Stephen; Havrdova, Eva Kubala; Lugaresi, Alessandra; Weinst...espandi
File in questo prodotto:
Eventuali allegati, non sono esposti

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/1011766
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact