The AI Explainbability 360 toolkit is an open-source library that supports interpretability and explainability of data and machine learning models. The AI Explainability 360 Python package includes a comprehensive set of algorithms that cover different dimensions of explanations along with proxy explainability metrics.
The AI Explainability 360 interactive experience provides a gentle introduction to the concepts and capabilities by walking through an example use case for different consumer personas. The tutorials and example notebooks offer a deeper, data scientist-oriented introduction. The complete API is also available.
There is no single approach to explainability that works best. There are many ways to explain: data vs. model, directly interpretable vs. post hoc explanation, local vs. global, etc. It may therefore be confusing to figure out which algorithms are most appropriate for a given use case. To help, we have created some guidance material and a chart that can be consulted.
We have developed the package with extensibility in mind. This library is still in development. We encourage the contribution of your explainability algorithms and metrics. To get started as a contributor, please join the AI Explainability 360 Community by requesting an invitation here. Please review the instructions to contribute code here.
Supported explainability algorithms
Local post-hoc explanation
- ProtoDash (Gurumoorthy et al., 2017)
- Contrastive Explanations Method (Dhurandhar et al., 2018)
- Contrastive Explanations Method with Monotonic Attribute Functions (Luss et al., 2019)
Local direct explanation
- Teaching AI to Explain its Decisions (Hind et al., 2019)
Global direct explanation
- Boolean Decision Rules via Column Generation (Light Edition) (Dash et al., 2018)
- Generalized Linear Rule Models (Wei et al., 2019)
Global post-hoc explanation
- ProfWeight (Dhurandhar et al., 2018)
Supported explainability metrics
(Optional) Create a virtual environment
AIX360 requires specific versions of many Python packages which may conflict with other projects on your system. A virtual environment manager is strongly recommended to ensure dependencies may be installed safely. If you have trouble installing AIX360, try this first.
Conda is recommended for all configurations though Virtualenv is generally interchangeable for our purposes. Miniconda is sufficient (see the difference between Anaconda and Miniconda if you are curious) and can be installed from here if you do not already have it.
Then, to create a new Python 3.6 environment, run:
conda create --name aix360 python=3.6 conda activate aix360
The shell should now look like
(aix360) $. To deactivate the environment, run:
(aix360)$ conda deactivate
The prompt will return back to
Note: Older versions of conda may use
source activate aix360 and
source deactivate (
activate aix360 and
deactivate on Windows).
Clone the latest version of this repository:
(aix360)$ git clone https://github.com/IBM/AIX360
If you'd like to run the examples and tutorial notebooks, download the datasets now and place them in their respective folders as described in aix360/data/README.md.
Then, navigate to the root directory of the project which contains
setup.py file and run:
(aix360)$ pip install -e .
examples directory contains a diverse collection of jupyter notebooks
that use AI Explainability 360 in various ways. Both examples and tutorial notebooks illustrate
working code using AIX360. Tutorials provide additional discussion that walks
the user through the various steps of the notebook. See the details about
tutorials and demos here
- Coming soon.