Skip to content

GDPlumb/ExpO

Repository files navigation

ExpO

This is the code to reproduce the results from Regularizing Black-box Models for Improved Interpretability which was published at NeurIPS 2020. A summary of this paper is as follows:

''Most of the work on interpretable machine learning has focused on designing either inherently interpretable models, which typically trade-off accuracy for interpretability, or post-hoc explanation systems, whose explanation quality can be unpredictable. Our method, ExpO, is a hybridization of these approaches that regularizes a model for explanation quality at training time. Importantly, these regularizers are differentiable, model agnostic, and require no domain knowledge to define. We demonstrate that post-hoc explanations for ExpO-regularized models have better explanation quality, as measured by the common fidelity and stability metrics. We verify that improving these metrics leads to significantly more useful explanations with a user study on a realistic task.''

About

Explanation Optimization

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published