EURO 2021 Athens<br />Abstract Submission

Online presentation.

423. Exploiting Auto-Encoders for Explaining Black Box Classifiers

Invited abstract in session MD-32: Fair and explainable models 2, stream Multiple Criteria Decision Analysis.

Area: Decision support

Monday, 14:30-16:00
Room: Virtual Room 32

Authors (first author is the speaker)

1. Riccardo Guidotti
Computer Science, University of Pisa
2. Anna Monreale
Computer Science, University of Pisa


Artificial Intelligence (AI) has nowadays a tremendous socio-economic impact and a pervasive adoption in every field of modern society. Many applications in different fields, such as credit score assessment, medical diagnosis, etc., are based on AI systems. Unfortunately, these systems often reach their impressive performance through obscure machine learning models that "hide" the logic of their internal decision processes to humans. For this reason, these models are called black boxes. The missing interpretability of black boxes is a limitation to AI adoption in socially sensitive contexts. As a consequence, the research in eXplainable AI (XAI) has recently caught much attention. A promising line of research in XAI exploits auto-encoders for explaining black box classifiers working on non-tabular data (images, time series, etc.). The ability of autoencoders to compress any data in a low-dimensional tabular representation, and then reconstruct it with negligible loss, provides the great opportunity to work in the latent space for the extraction of meaningful explanations, for example, through the generation of new synthetic samples that can be fed to a black-box to understand where its decision boundary lies. We discuss recent XAI solutions based on autoencoders that enable the extraction of meaningful explanations composed by factual and counterfactual rules, and by exemplars and counter-exemplars, offering a deep understanding of the local decision of the black box.


Status: accepted

Back to the list of papers