Integration of Local and Global Features Explanation with Global Rules Extraction and Generation Tools
by Contreras Ordoñez Victor Hugo, Davide Calvaresi, and Michael I. Schumacher
Widely used in a growing number of domains, Deep Learning predictors are achieving remarkable results. However, the lack of transparency (i.e., opacity) of their inner mechanisms has raised trust and employability concerns. Nevertheless, several approaches fostering models of interpretability and explainability have been developed in the last decade. This paper combines approaches for local feature explanation (i.e., Contextual Importance and Utility – CIU) and global feature explanation (i.e., Explainable Layers) with a rule extraction system, namely ECLAIRE. The proposed pipeline has been tested in four scenarios employing a breast cancer diagnosis dataset. The results show improvements such as the production of more human-interpretable rules and adherence of the produced rules with the original model.
@inproceedings{ContrerasSC22,
author = {Victor Contreras and
Michael Schumacher and
Davide Calvaresi},
editor = {Davide Calvaresi and
Amro Najjar and
Michael Winikoff and
Kary Fr{\"{a}}mling},
title = {Integration of Local and Global Features Explanation with Global Rules
Extraction and Generation Tools},
booktitle = {Explainable and Transparent {AI} and Multi-Agent Systems - 4th International
Workshop, {EXTRAAMAS} 2022, Virtual Event, May 9-10, 2022, Revised
Selected Papers},
series = {Lecture Notes in Computer Science},
volume = {13283},
pages = {19--37},
publisher = {Springer},
year = {2022},
url = {https://doi.org/10.1007/978-3-031-15565-9\_2},
doi = {10.1007/978-3-031-15565-9\_2},
timestamp = {Tue, 18 Oct 2022 22:16:54 +0200},
biburl = {https://dblp.org/rec/conf/atal/ContrerasSC22.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}