Advanced search
Start date
Betweenand


Are Explanations Helpful? A Comparative Analysis of Explainability Methods in Skin Lesion Classifiers

Full text
Author(s):
Paccotacya-Yanque, Rosa Y. G. ; Bissoto, Alceu ; Avila, Sandra
Total Authors: 3
Document type: Journal article
Source: 2024 20TH INTERNATIONAL SYMPOSIUM ON MEDICAL INFORMATION PROCESSING AND ANALYSIS, SIPAIM 2024; v. N/A, p. 5-pg., 2024-01-01.
Abstract

Deep Learning has shown outstanding results in computer vision tasks; healthcare is no exception. However, there is no straightforward way to expose the decision-making process of DL models. Good accuracy is not enough for skin cancer predictions. Understanding the model's behavior is crucial for clinical application and reliable outcomes. In this work, we identify desiderata for explanations in skin-lesion models. We analyzed seven methods, four based on pixel-attribution (Grad-CAM, Score-CAM, LIME, SHAP) and three high-level concepts (ACE, ICE, CME), for a deep neural network trained on the International Skin Imaging Collaboration Archive. Our findings indicate that while these techniques reveal biases, there is room for improving the comprehensiveness of explanations to achieve transparency in skin-lesion models. (AU)

FAPESP's process: 23/12086-9 - Araceli: Artificial Intelligence in the Fight Against Child Sexual Abuse
Grantee:Sandra Eliza Fontes de Avila
Support Opportunities: Regular Research Grants
FAPESP's process: 20/09838-0 - BI0S - Brazilian Institute of Data Science
Grantee:João Marcos Travassos Romano
Support Opportunities: Research Grants - Research Centers in Engineering Program
FAPESP's process: 13/08293-7 - CCES - Center for Computational Engineering and Sciences
Grantee:Munir Salomao Skaf
Support Opportunities: Research Grants - Research, Innovation and Dissemination Centers - RIDC