Skip to content

Commit

Permalink
Remove LRPZIgnoreBias and LRPEpsilonIgnoreBias
Browse files Browse the repository at this point in the history
  • Loading branch information
adrhill committed Apr 14, 2022
1 parent 60e8f8a commit 64bab68
Show file tree
Hide file tree
Showing 6 changed files with 0 additions and 53 deletions.
4 changes: 0 additions & 4 deletions src/innvestigate/analyzer/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -24,15 +24,13 @@
LRPAlpha2Beta1IgnoreBias,
LRPAlphaBeta,
LRPEpsilon,
LRPEpsilonIgnoreBias,
LRPFlat,
LRPSequentialPresetA,
LRPSequentialPresetAFlat,
LRPSequentialPresetB,
LRPSequentialPresetBFlat,
LRPSequentialPresetBFlatUntilIdx,
LRPWSquare,
LRPZIgnoreBias,
LRPZPlus,
LRPZPlusFast,
)
Expand Down Expand Up @@ -60,9 +58,7 @@
# Relevance based
"lrp": LRP,
"lrp.z": LRPZ,
"lrp.z_IB": LRPZIgnoreBias,
"lrp.epsilon": LRPEpsilon,
"lrp.epsilon_IB": LRPEpsilonIgnoreBias,
"lrp.w_square": LRPWSquare,
"lrp.flat": LRPFlat,
"lrp.alpha_beta": LRPAlphaBeta,
Expand Down
21 changes: 0 additions & 21 deletions src/innvestigate/analyzer/relevance_based/relevance_analyzer.py
Original file line number Diff line number Diff line change
Expand Up @@ -21,9 +21,7 @@
"LRP",
"LRP_RULES",
"LRPZ",
"LRPZIgnoreBias",
"LRPEpsilon",
"LRPEpsilonIgnoreBias",
"LRPWSquare",
"LRPFlat",
"LRPAlphaBeta",
Expand Down Expand Up @@ -148,9 +146,7 @@ def _create_analysis(
# Utility list enabling name mappings via string
LRP_RULES: Dict = {
"Z": rrule.ZRule,
"ZIgnoreBias": rrule.ZIgnoreBiasRule,
"Epsilon": rrule.EpsilonRule,
"EpsilonIgnoreBias": rrule.EpsilonIgnoreBiasRule,
"WSquare": rrule.WSquareRule,
"Flat": rrule.FlatRule,
"AlphaBeta": rrule.AlphaBetaRule,
Expand Down Expand Up @@ -573,16 +569,6 @@ def __init__(self, model, *args, **kwargs):
self._do_model_checks()


class LRPZIgnoreBias(_LRPFixedParams):
"""LRP-analyzer that uses the LRP-Z-ignore-bias rule"""

def __init__(self, model, *args, **kwargs):
super().__init__(
model, *args, rule="ZIgnoreBias", bn_layer_rule="ZIgnoreBias", **kwargs
)
self._do_model_checks()


class LRPEpsilon(_LRPFixedParams):
"""LRP-analyzer that uses the LRP-Epsilon rule"""

Expand Down Expand Up @@ -611,13 +597,6 @@ def __init__(self, *args, **kwargs):
self._do_model_checks()


class LRPEpsilonIgnoreBias(LRPEpsilon):
"""LRP-analyzer that uses the LRP-Epsilon-ignore-bias rule"""

def __init__(self, model, *args, epsilon=1e-7, **kwargs):
super().__init__(model, *args, epsilon=epsilon, bias=False, **kwargs)
self._do_model_checks()


class LRPWSquare(_LRPFixedParams):
"""LRP-analyzer that uses the DeepTaylor W**2 rule"""
Expand Down
16 changes: 0 additions & 16 deletions src/innvestigate/analyzer/relevance_based/relevance_rule.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,9 +19,7 @@
# dedicated treatment for special layers
# general rules
"ZRule",
"ZIgnoreBiasRule",
"EpsilonRule",
"EpsilonIgnoreBiasRule",
"WSquareRule",
"FlatRule",
"AlphaBetaRule",
Expand Down Expand Up @@ -71,14 +69,6 @@ def apply(
return [klayers.Multiply()([a, b]) for a, b in zip(Xs, grads)]


class ZIgnoreBiasRule(ZRule):
"""
Basic LRP decomposition rule, ignoring the bias neuron
"""

def __init__(self, *args, **kwargs):
super().__init__(*args, bias=False, **kwargs)


class EpsilonRule(igraph.ReverseMappingBase):
"""
Expand Down Expand Up @@ -122,12 +112,6 @@ def apply(
return [klayers.Multiply()([a, b]) for a, b in zip(Xs, grads)]


class EpsilonIgnoreBiasRule(EpsilonRule):
"""Same as EpsilonRule but ignores the bias."""

def __init__(self, *args, **kwargs):
super().__init__(*args, bias=False, **kwargs)


class WSquareRule(igraph.ReverseMappingBase):
"""W**2 rule from Deep Taylor Decomposition"""
Expand Down
4 changes: 0 additions & 4 deletions tests/analyzer/test_reference_layer.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,14 +25,12 @@
LRPAlpha2Beta1,
LRPAlpha2Beta1IgnoreBias,
LRPEpsilon,
LRPEpsilonIgnoreBias,
LRPFlat,
LRPSequentialPresetA,
LRPSequentialPresetAFlat,
LRPSequentialPresetB,
LRPSequentialPresetBFlat,
LRPWSquare,
LRPZIgnoreBias,
LRPZPlus,
LRPZPlusFast,
)
Expand All @@ -51,15 +49,13 @@
"LRPZ": (LRPZ, {}),
"LRPZ_Flat_input_layer_rule": (LRPZ, {"input_layer_rule": "Flat"}),
"LRPZ_boxed_input_layer_rule": (LRPZ, {"input_layer_rule": (-10, 10)}),
"LRPZIgnoreBias": (LRPZIgnoreBias, {}),
"LRPZPlus": (LRPZPlus, {}),
"LRPZPlusFast": (LRPZPlusFast, {}),
"LRPAlpha1Beta0": (LRPAlpha1Beta0, {}),
"LRPAlpha1Beta0IgnoreBias": (LRPAlpha1Beta0IgnoreBias, {}),
"LRPAlpha2Beta1": (LRPAlpha2Beta1, {}),
"LRPAlpha2Beta1IgnoreBias": (LRPAlpha2Beta1IgnoreBias, {}),
"LRPEpsilon": (LRPEpsilon, {}),
"LRPEpsilonIgnoreBias": (LRPEpsilonIgnoreBias, {}),
"LRPFlat": (LRPFlat, {}),
"LRPWSquare": (LRPWSquare, {}),
"LRPSequentialPresetA": (LRPSequentialPresetA, {}),
Expand Down
4 changes: 0 additions & 4 deletions tests/analyzer/test_reference_vgg16.py
Original file line number Diff line number Diff line change
Expand Up @@ -42,14 +42,12 @@
LRPAlpha2Beta1,
LRPAlpha2Beta1IgnoreBias,
LRPEpsilon,
LRPEpsilonIgnoreBias,
LRPFlat,
LRPSequentialPresetA,
LRPSequentialPresetAFlat,
LRPSequentialPresetB,
LRPSequentialPresetBFlat,
LRPWSquare,
LRPZIgnoreBias,
LRPZPlus,
LRPZPlusFast,
)
Expand All @@ -68,15 +66,13 @@
"LRPZ": (LRPZ, {}),
"LRPZ_Flat_input_layer_rule": (LRPZ, {"input_layer_rule": "Flat"}),
"LRPZ_boxed_input_layer_rule": (LRPZ, {"input_layer_rule": (-10, 10)}),
"LRPZIgnoreBias": (LRPZIgnoreBias, {}),
"LRPZPlus": (LRPZPlus, {}),
"LRPZPlusFast": (LRPZPlusFast, {}),
"LRPAlpha1Beta0": (LRPAlpha1Beta0, {}),
"LRPAlpha1Beta0IgnoreBias": (LRPAlpha1Beta0IgnoreBias, {}),
"LRPAlpha2Beta1": (LRPAlpha2Beta1, {}),
"LRPAlpha2Beta1IgnoreBias": (LRPAlpha2Beta1IgnoreBias, {}),
"LRPEpsilon": (LRPEpsilon, {}),
"LRPEpsilonIgnoreBias": (LRPEpsilonIgnoreBias, {}),
"LRPFlat": (LRPFlat, {}),
"LRPWSquare": (LRPWSquare, {}),
"LRPSequentialPresetA": (LRPSequentialPresetA, {}),
Expand Down
4 changes: 0 additions & 4 deletions tests/analyzer/test_relevance_based.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,14 +9,12 @@
LRPAlpha2Beta1,
LRPAlpha2Beta1IgnoreBias,
LRPEpsilon,
LRPEpsilonIgnoreBias,
LRPFlat,
LRPSequentialPresetA,
LRPSequentialPresetAFlat,
LRPSequentialPresetB,
LRPSequentialPresetBFlat,
LRPWSquare,
LRPZIgnoreBias,
LRPZPlus,
LRPZPlusFast,
)
Expand All @@ -27,7 +25,6 @@
"LRPZ": (LRPZ, {}),
"LRPZ_Flat_input_layer_rule": (LRPZ, {"input_layer_rule": "Flat"}),
"LRPZ_boxed_input_layer_rule": (LRPZ, {"input_layer_rule": (-10, 10)}),
"LRPZIgnoreBias": (LRPZIgnoreBias, {}),
"LRPZPlus": (LRPZPlus, {}),
"LRPZPlusFast": (LRPZPlusFast, {}),
"BaselineLRPZ": (BaselineLRPZ, {}),
Expand All @@ -36,7 +33,6 @@
"LRPAlpha2Beta1": (LRPAlpha2Beta1, {}),
"LRPAlpha2Beta1IgnoreBias": (LRPAlpha2Beta1IgnoreBias, {}),
"LRPEpsilon": (LRPEpsilon, {}),
"LRPEpsilonIgnoreBias": (LRPEpsilonIgnoreBias, {}),
"LRPFlat": (LRPFlat, {}),
"LRPWSquare": (LRPWSquare, {}),
"LRPSequentialPresetA": (LRPSequentialPresetA, {}),
Expand Down

0 comments on commit 64bab68

Please sign in to comment.