github: https://github.com/explain-llm/CafGa
CafGa can be installed through PyPI using
pip install cafga
When running CafGa from the repository run:
pip install -r requirements.txt
Note that some of the extra functionality requires further installations:
-
CafGa provides two jupyter widgets. The edit widget allows one to visually edit assignments and the display widget displays the attributions generated by the explanation. To use these please follow the instructions in the 'Demo Instructions.md' file.
-
CafGa offers a predefined ChatGPT model. To use it you need to place a .env file with your API key in the directory that contains the cafga.py file (in the site-packages directory).
The following provides an explanation of the main functions of cafga. To see an example of how to use cafga please look at the demo.
To begin using CafGa, start by importing CafGa creating a cafga object:
from cafga import CafGa
cafga = CafGa(model = 'your_model')
The model parameter is where you pass the model you want to explain. To allow for parallelization in how your model generates predictions (e.g. by batching) cafga sends lists of inputs to your model instead of single inputs. Thus, the function that implements your model should take a list of strings as input and output either a list of strings or a list of floats as output (i.e. a list containing one output for every input).
Once cafga is instantiated the typical usage of cafga runs proceeds in three steps: Explanation, Evaluation, and Visualisation.
To generate an explanation run the explain function on the instantiated cafga object:
explanation = cafga.explain(args)
There are two way of using the explain functions.
Firstly, you can pass the string you want to get an explanation for without segmenting it into the individual parts that you want to get attributions for. In this case you need to provide the name of the predefined attribution method ('word', 'sentence', 'syntax-parse') that you want to use.
Secondly, you can provide your own segmentation of the input by using the segmented_input
parameter. In this case you will also need to provide the assignments of input segment to group with the input_assignments
parameter. Specifically, the input_assignments[i] = g_i
should be the index of the group that input_segments[i]
belongs to.
Once an explanation object has been generated you can pass it on to the evaluation function:
evaluation = cafga.evaluate(explanation, args)
The two forms of evaluation currently supported are deletion (going from all features present to no features present) and insertion (going from no features present to all features present), which can be indicated by the direction
parameter. The resulting evaluation accordinlgy contains the array of difference values computed as part of the perturbation curve.
Finally, the perturbation curve generated by the evaluation can be visualised using the visualisation function:
cafga.visualize_evaluation(evaluated_explanations, args)
Since you may want to plot the aggregate over many evaluations the visualisation functions takes in a list of evaluations as input. The two forms of aggregation currently supported are equal width binning and linear interpolation.