Abstract

Deep neural networks are widely used for classification. These deep models often suffer from a lack of interpretability---they are particularly difficult to understand because of their non-linear nature. As a result, neural networks are often treated as "black box" models, and in the past, have been trained purely to optimize the accuracy of predictions. In this work, we create a novel network architecture for deep learning that naturally explains its own reasoning for each prediction. This architecture contains an autoencoder and a special prototype layer, where each unit of that layer stores a weight vector that resembles an encoded training input. The encoder of the autoencoder allows us to do comparisons within the latent space, while the decoder allows us to visualize the learned prototypes. The training objective has four terms: an accuracy term, a term that encourages every prototype to be similar to at least one encoded input, a term that encourages every encoded input to be close to at least one prototype, and a term that encourages faithful reconstruction by the autoencoder. The distances computed in the prototype layer are used as part of the classification process. Since the prototypes are learned during training, the learned network naturally comes with explanations for each prediction, and the explanations are loyal to what the network actually computes.

Keywords

AutoencoderInterpretabilityArtificial intelligenceComputer scienceDeep learningArtificial neural networkMachine learningTerm (time)Black boxEncoderNetwork architectureProcess (computing)Layer (electronics)ArchitectureDeep belief networkRecurrent neural network

Affiliated Institutions

Related Publications

Publication Info

Year
2018
Type
article
Volume
32
Issue
1
Citations
390
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

390
OpenAlex

Cite This

Oscar Li, Hao Liu, Chaofan Chen et al. (2018). Deep Learning for Case-Based Reasoning Through Prototypes: A Neural Network That Explains Its Predictions. Proceedings of the AAAI Conference on Artificial Intelligence , 32 (1) . https://doi.org/10.1609/aaai.v32i1.11771

Identifiers

DOI
10.1609/aaai.v32i1.11771