TheSequence

TheSequence

Share this post

TheSequence
TheSequence
Edge 263: Local Model-Agnostic Interpretability Methods: Counterfactual Explanations

Edge 263: Local Model-Agnostic Interpretability Methods: Counterfactual Explanations

Counterfactual explanations as an ML interpretability method, Google's StylEx and Microsoft's DiCE implementation

Jan 24, 2023
∙ Paid
25

Share this post

TheSequence
TheSequence
Edge 263: Local Model-Agnostic Interpretability Methods: Counterfactual Explanations
Share
Created by: Stable Diffusion

In this issue:

  1. An overview of local-interpretability methods based on counterfactual explanations.

  2. Google’s StylEx method for generation visual explanations in image classifiers

  3. Microsoft’s open source implementation of the DiCE method.

💡 ML Concept of the Day: : Local Model-Agnostic Interpretability Methods: Counterfactual Explanations

In the last few editions of this newsletter, we have been discussing local model-agnostic interpretability methods which focus on deriving explanations based on the outcome of a single prediction. One of the most interesting methods in this area of ML interpretability is known as counterfactual explanations. From causality theory, we know counterfactual explanations as a predicate that describes the relationship between two variables in the following form: “If X had not occurred, Y would not have occurred.  How different would the world be if the Soviet Union would have landed in the moon first. Who knows but makes for an interesting counterfactual analysis. Consider a more pragmatic example of a home loan application rejected by an ML algorithm. A counterfactual analysis can reveal than an income increase is the most likely factor to increase the chances to getting a loan next time.

Image Credit: Microsoft Research

This post is for paid subscribers

Already a paid subscriber? Sign in
© 2025 Jesus Rodriguez
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share