Skip to content

Break Down with interactions for local explanations (SHAP, BreakDown, iBreakDown)


Notifications You must be signed in to change notification settings


Repository files navigation

Model Agnostic Local Attributions

R build status Coverage Status CRAN_Status_Badge Total Downloads


The iBreakDown package is a model agnostic tool for explanation of predictions from black boxes ML models. Break Down Table shows contributions of every variable to a final prediction. Break Down Plot presents variable contributions in a concise graphical way. SHAP (Shapley Additive Attributions) values are calculated as average from random Break Down profiles. This package works for binary classifiers as well as regression models.

iBreakDown is a successor of the breakDown package. It is faster (complexity O(p) instead of O(p^2)). It supports variable interactions and interactive explanations with D3.js visualizations. It is imported and used to compute model explanations in multiple packages e.g. DALEX, modelStudio, arenar.

Methodology behind the iBreakDown package is described in the arXiv paper and Explanatory Model Analysis book. It is a part of DrWhy.AI universe.


# the easiest way to get iBreakDown is to install it from CRAN:

# Or the the development version from GitHub:
# install.packages("devtools")

Learn more

Find more examples in the EMA book:

This version also works with D3: see an example and demo.



Work on this package was financially supported by the NCN Opus grant 2016/21/B/ST6/02176.