Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
12 changes: 12 additions & 0 deletions captum/attr/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,12 @@
from ._core.input_x_gradient import InputXGradient # noqa
from ._core.saliency import Saliency # noqa
from ._core.noise_tunnel import NoiseTunnel # noqa
from ._core.layer_conductance import LayerConductance # noqa
from ._core.layer_gradient_x_activation import LayerGradientXActivation # noqa
from ._core.layer_activation import LayerActivation # noqa
from ._core.neuron_conductance import NeuronConductance # noqa
from ._core.neuron_gradient import NeuronGradient # noqa
from ._core.neuron_integrated_gradients import NeuronIntegratedGradients # noqa
from ._models.base import (
InterpretableEmbeddingBase,
TokenReferenceBase,
Expand All @@ -19,6 +25,12 @@
"DeepLift",
"InputXGradient",
"Saliency",
"LayerConductance",
"LayerGradientXActivation",
"LayerActivation",
"NeuronConductance",
"NeuronGradient",
"NeuronIntegratedGradients",
"NoiseTunnel",
"InterpretableEmbeddingBase",
"TokenReferenceBase",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@
from .._utils.gradient import compute_layer_gradients_and_eval


class Conductance(LayerAttribution):
class LayerConductance(LayerAttribution):
def __init__(self, forward_func, layer):
r"""
Args
Expand Down
2 changes: 1 addition & 1 deletion captum/attr/_core/neuron_conductance.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@ def attribute(
neuron_index,
baselines=None,
target=None,
n_steps=500,
n_steps=50,
method="riemann_trapezoid",
):
r"""
Expand Down
844 changes: 844 additions & 0 deletions notebooks/Titanic_Basic_Interpret.ipynb

Large diffs are not rendered by default.

Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
import unittest

import torch
from captum.attr._core.conductance import Conductance
from captum.attr._core.layer_conductance import LayerConductance

from .helpers.basic_models import TestModel_ConvNet, TestModel_MultiLayer
from .helpers.conductance_reference import ConductanceReference
Expand Down Expand Up @@ -67,7 +67,7 @@ def test_matching_conv_with_baseline_conductance(self):
def _conductance_test_helper(
self, model, target_layer, test_input, expected_conductance
):
cond = Conductance(model, target_layer)
cond = LayerConductance(model, target_layer)
attributions = cond.attribute(
test_input, target=0, n_steps=500, method="gausslegendre"
)
Expand All @@ -88,7 +88,7 @@ def forward_hook(module, inp, out):
final_output = model(test_input)
hook.remove()
target_index = torch.argmax(torch.sum(final_output, 0))
cond = Conductance(model, target_layer)
cond = LayerConductance(model, target_layer)
cond_ref = ConductanceReference(model, target_layer)
attributions = cond.attribute(
test_input,
Expand Down
4 changes: 2 additions & 2 deletions tests/attr/test_neuron_conductance.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
import unittest

import torch
from captum.attr._core.conductance import Conductance
from captum.attr._core.layer_conductance import LayerConductance
from captum.attr._core.neuron_conductance import NeuronConductance

from .helpers.basic_models import TestModel_ConvNet, TestModel_MultiLayer
Expand Down Expand Up @@ -59,7 +59,7 @@ def _conductance_input_test_assert(
def _conductance_input_sum_test_assert(
self, model, target_layer, test_input, test_baseline=None
):
layer_cond = Conductance(model, target_layer)
layer_cond = LayerConductance(model, target_layer)
attributions = layer_cond.attribute(
test_input,
baselines=test_baseline,
Expand Down