Investigating Neighborhood Generation Methods for Explanations of Obscure Image Classifiers
Given the wide use of machine learning approaches based on opaque prediction models,understanding the reasons behind decisions of black box decision systems is nowadays a crucial topic.We address the problem of providing meaningful explanations in the widely-applied image classification tasks.In particular,we explore the impact of changing the neighborhood generation function for a local interpretable modelagnostic explanator by proposing four different variants.All the proposed methods are based on a grid-based segmentation of the images,but each of them proposes a different strategy for generating the neighborhood of the image for which an explanation is required.A deep experimentation shows both improvements and weakness of each proposed approach.
Riccardo Guidotti Anna Monreale Leonardo Cariaggi
ISTI-CNR,Pisa,Italy;University of Pisa,Pisa,Italy University of Pisa,Pisa,Italy
国际会议
澳门
英文
55-68
2019-04-14(万方平台首次上网日期,不代表论文的发表时间)