Content area

Abstract

Gaining clinicians’ trust will unleash the full potential of artificial intelligence (AI) in medicine, and explaining AI decisions is seen as the way to build trustworthy systems. However, explainable artificial intelligence (XAI) methods in medicine often lack a proper evaluation. In this paper, we present our evaluation methodology for XAI methods using forward simulatability. We define the Forward Simulatability Score (FSS) and analyze its limitations in the context of clinical predictors. Then, we applied FSS to our XAI approach defined over an ML-RO, a machine learning clinical predictor based on random optimization over a multiple kernel support vector machine (SVM) algorithm. To Compare FSS values before and after the explanation phase, we test our evaluation methodology for XAI methods on three clinical datasets, namely breast cancer, VTE, and migraine. The ML-RO system is a good model on which to test our XAI evaluation strategy based on the FSS. Indeed, ML-RO outperforms two other base models—a decision tree (DT) and a plain SVM—in the three datasets and gives the possibility of defining different XAI models: TOPK, MIGF, and F4G. The FSS evaluation score suggests that the explanation method F4G for the ML-RO is the most effective in two datasets out of the three tested, and it shows the limits of the learned model for one dataset. Our study aims to introduce a standard practice for evaluating XAI methods in medicine. By establishing a rigorous evaluation framework, we seek to provide healthcare professionals with reliable tools for assessing the performance of XAI methods to enhance the adoption of AI systems in clinical practice.

Details

Business indexing term
Title
Evaluating Explainable Machine Learning Models for Clinicians
Author
Scarpato, Noemi 1   VIAFID ORCID Logo  ; Nourbakhsh, Aria 2 ; Ferroni, Patrizia 1 ; Riondino, Silvia 3 ; Roselli, Mario 3 ; Fallucchi, Francesca 4 ; Barbanti, Piero 5 ; Guadagni, Fiorella 1 ; Zanzotto, Fabio Massimo 2   VIAFID ORCID Logo 

 San Raffaele Roma Open University, Rome, Italy (GRID:grid.466134.2) (ISNI:0000 0004 4912 5648); Interinstitutional Multidisciplinary Biobank (BioBIM), IRCCS San Raffaele Roma, Rome, Italy (GRID:grid.18887.3e) (ISNI:0000000417581884) 
 University of Rome Tor Vergata, Department of Enterprise Engineering, Rome, Italy (GRID:grid.6530.0) (ISNI:0000 0001 2300 0941) 
 University of Rome Tor Vergata, Department of Systems Medicine, Rome, Italy (GRID:grid.6530.0) (ISNI:0000 0001 2300 0941) 
 Guglielmo Marconi University, Rome, Italy (GRID:grid.440899.8) (ISNI:0000 0004 1780 761X) 
 San Raffaele Roma Open University, Rome, Italy (GRID:grid.466134.2) (ISNI:0000 0004 4912 5648); IRCCS San Raffaele Roma, Headache and Pain Unit, Rome, Italy (GRID:grid.18887.3e) (ISNI:0000000417581884) 
Publication title
Volume
16
Issue
4
Pages
1436-1446
Publication year
2024
Publication date
Jul 2024
Publisher
Springer Nature B.V.
Place of publication
New York
Country of publication
Netherlands
ISSN
18669956
e-ISSN
18669964
Source type
Scholarly Journal
Language of publication
English
Document type
Journal Article
Publication history
 
 
Online publication date
2024-05-31
Milestone dates
2024-05-06 (Registration); 2023-07-19 (Received); 2024-05-04 (Accepted)
Publication history
 
 
   First posting date
31 May 2024
ProQuest document ID
3076132393
Document URL
https://www.proquest.com/scholarly-journals/evaluating-explainable-machine-learning-models/docview/3076132393/se-2?accountid=208611
Copyright
© The Author(s) 2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Last updated
2024-07-06
Database
ProQuest One Academic