Estudios
I+D+I
Institución
Internacional
Vida Universitaria
Repositorio Institucional de Documentos
Buscar
Enviar
Personalizar
Sus alertas
Sus carpetas
Sus búsquedas
Ayuda
EN
/
ES
Página principal
>
Artículos
> Explainable AI toward understanding the performance of the top three TADPOLE Challenge methods in the forecast of Alzheimer’s disease diagnosis
Estadísticas de uso
Gráficos
Explainable AI toward understanding the performance of the top three TADPOLE Challenge methods in the forecast of Alzheimer’s disease diagnosis
Hernandez, Mónica
(Universidad de Zaragoza)
;
Ramón-Julvez, Ubaldo
(Universidad de Zaragoza)
;
Ferraz, Francisco
Resumen:
The Alzheimer0s Disease Prediction Of Longitudinal Evolution (TADPOLE) Challenge is the most comprehensive challenge to date with regard to the number of subjects, considered features, and challenge participants. The initial objective of TADPOLE was the identification of the most predictive data, features, and methods for the progression of subjects at risk of developing Alzheimer0s. The challenge was successful in recognizing tree-based ensemble methods such as gradient boosting and random forest as the best methods for the prognosis of the clinical status in Alzheimer’s disease (AD). However, the challenge outcome was limited to which combination of data processing and methods exhibits the best accuracy; hence, it is difficult to determine the contribution of the methods to the accuracy. The quantification of feature importance was globally approached by all the challenge participant methods. In addition, TADPOLE provided general answers that focused on improving performance while ignoring important issues such as interpretability. The purpose of this study is to intensively explore the models of the top three TADPOLE Challenge methods in a common framework for fair comparison. In addition, for these models, the most meaningful features for the prognosis of the clinical status of AD are studied and the contribution of each feature to the accuracy of the methods is quantified. We provide plausible explanations as to why the methods achieve such accuracy, and we investigate whether the methods use information coherent with clinical knowledge. Finally, we approach these issues through the analysis of SHapley Additive exPlanations (SHAP) values, a technique that has recently attracted increasing attention in the field of explainable artificial intelligence (XAI). Copyright: © 2022 Hernandez et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Idioma:
Inglés
DOI:
10.1371/journal.pone.0264695
Año:
2022
Publicado en:
PLoS ONE
17, 5 (2022), [32 pp.]
ISSN:
1932-6203
Factor impacto JCR:
3.7 (2022)
Categ. JCR:
MULTIDISCIPLINARY SCIENCES
rank: 26 / 73 = 0.356
(2022)
- Q2
- T2
Factor impacto CITESCORE:
6.0 -
General
(Q1)
Factor impacto SCIMAGO:
0.885 -
Multidisciplinary
(Q1)
Financiación:
info:eu-repo/grantAgreement/ES/DGA/T64-20R
Financiación:
info:eu-repo/grantAgreement/ES/MICINN/PID2019-104358RB-I00
Tipo y forma:
Artículo (Versión definitiva)
Área (Departamento):
Área Lenguajes y Sistemas Inf.
(
Dpto. Informát.Ingenie.Sistms.
)
Debe reconocer adecuadamente la autoría, proporcionar un enlace a la licencia e indicar si se han realizado cambios. Puede hacerlo de cualquier manera razonable, pero no de una manera que sugiera que tiene el apoyo del licenciador o lo recibe por el uso que hace.
Exportado de SIDERAL (2024-03-18-14:35:57)
Enlace permanente:
Copiar
Visitas y descargas
Este artículo se encuentra en las siguientes colecciones:
Artículos
Volver a la búsqueda
Registro creado el 2024-02-01, última modificación el 2024-03-19
Versión publicada:
PDF
Valore este documento:
Rate this document:
1
2
3
4
5
(Sin ninguna reseña)
Añadir a una carpeta personal
Exportar como
BibTeX
,
MARC
,
MARCXML
,
DC
,
EndNote
,
NLM
,
RefWorks