Interpretations are Useful: Penalizing Explanations to Align Neural Networks with Prior Knowledge

Laura Rieger*, Chandan Singh, W. James Murdoch, Bin Yu

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

56 Downloads (Pure)

Abstract

For an explanation of a deep learning model to be effective, it must both provide insight into a model and suggest a corresponding action in order to achieve an objective. Too often, the litany of proposed explainable deep learning methods stop at the first step, providing practitioners with insight into a model, but no way to act on it. In this paper we propose contextual decomposition explanation penalization (CDEP), a method that enables practitioners to leverage explanations to improve the performance of a deep learning model. In particular, CDEP enables inserting domain knowledge into a model to ignore spurious correlations, correct errors, and generalize to different types of dataset shifts. We demonstrate the ability of CDEP to increase performance on an array of toy and real datasets.
Original languageEnglish
Title of host publicationProceedings of the 37th International Conference on Machine Learning
PublisherInternational Machine Learning Society (IMLS)
Publication date2020
Pages8116-8126
Publication statusPublished - 2020
Event37th International Conference on Machine Learning - Virtual event, Virtual, Online
Duration: 13 Jul 202018 Jul 2020
https://icml.cc/Conferences/2020

Conference

Conference37th International Conference on Machine Learning
LocationVirtual event
CityVirtual, Online
Period13/07/202018/07/2020
Internet address
SeriesProceedings of Machine Learning Research
Volume119
ISSN2640-3498

Fingerprint

Dive into the research topics of 'Interpretations are Useful: Penalizing Explanations to Align Neural Networks with Prior Knowledge'. Together they form a unique fingerprint.

Cite this