Toward Scalable and Unified Example-Based Explanation and Outlier Detection

Penny Chong, Ngai Man Cheung, Yuval Elovici, Alexander Binder

Research output: Contribution to journalArticlepeer-review

8 Scopus citations
30 Downloads (Pure)

Abstract

When neural networks are employed for high-stakes decision-making, it is desirable that they provide explanations for their prediction in order for us to understand the features that have contributed to the decision. At the same time, it is important to flag potential outliers for in-depth verification by domain experts. In this work we propose to unify two differing aspects of explainability with outlier detection. We argue for a broader adoption of prototype-based student networks capable of providing an example-based explanation for their prediction and at the same time identify regions of similarity between the predicted sample and the examples. The examples are real prototypical cases sampled from the training set via a novel iterative prototype replacement algorithm. Furthermore, we propose to use the prototype similarity scores for identifying outliers. We compare performance in terms of the classification, explanation quality and outlier detection of our proposed network with baselines. We show that our prototype-based networks extending beyond similarity kernels deliver meaningful explanations and promising outlier detection results without compromising classification accuracy.

Original languageEnglish
Pages (from-to)525-540
Number of pages16
JournalIEEE Transactions on Image Processing
Volume31
DOIs
StatePublished - 18 Nov 2021

Keywords

  • Explainability
  • Image classification
  • LRP
  • Outlier detection
  • Prototypes
  • Pruning

ASJC Scopus subject areas

  • Software
  • Computer Graphics and Computer-Aided Design

Fingerprint

Dive into the research topics of 'Toward Scalable and Unified Example-Based Explanation and Outlier Detection'. Together they form a unique fingerprint.

Cite this