XAI

Providing insight, explanations, and interpretability to machine learning methods.

101 resources4 categoriesView Original

Papers(93 items)

A

Ada-SISE

Adaptive semantice inpute sampling for explanation.

Papers
A

ALE

Accumulated local effects plot.

Papers
A

ALIME

Autoencoder Based Approach for Local Interpretability.

Papers
A

Anchors

High-Precision Model-Agnostic Explanations.

Papers
A

Attention is not --not-- Explanation

This is a rebutal to the above paper. Authors argue that multiple explanations can be valid and that the and that attention can produce *a* valid explanation, if not -the- valid explanation.

Papers
A

Attention is not Explanation

Authors perform a series of NLP experiments which argue attention does not provide meaningful explanations. They also demosntrate that different attentions can generate similar model outputs.

Papers
A

Auditing

Auditing black-box models.

Papers
B

BayLIME

Bayesian local interpretable model-agnostic explanations.

Papers
B

Break Down

Break down plots for additive attributions.

Papers
C

CAM

Class activation mapping.

Papers
C

CDT

Confident interpretation of Bayesian decision tree ensembles.

Papers
C

CICE

Centered ICE plot.

Papers
C

CMM

Combined multiple models metalearner.

Papers
C

Conj Rules

Using sampling and queries to extract rules from trained neural networks.

Papers
C

CP

Contribution propogation.

Papers
D

Decision List

Like a decision tree with no branches.

Papers
D

Decision Trees

The tree provides an interpretation.

Papers
D

DecText

Extracting decision trees from trained neural networks.

Papers
D

DeepLIFT

Deep label-specific feature learning for image annotation.

Papers
D

Do Not Trust Additive Explanations

Authors argue that addditive explanations (e.g. LIME, SHAP, Break Down) fail to take feature ineractions into account and are thus unreliable.

Papers
D

DTD

Deep Taylor decomposition.

Papers
E

Explainable Boosting Machine

Method that predicts based on learned vector graphs of features.

Papers
E

Explainable Deep Learning: A Field Guide for th...

An in-depth description of XAI focused on technqiues for deep learning.

Papers
E

ExplainD

Explanations of evidence in additive classifiers.

Papers
E

Explanation in Artificial Intelligence: Insight...

This paper provides an introduction to the social science research into explanations. The author provides 4 major findings: (1) explanations are constrastive, (2) explanations are selected, (3) probabilities probably don't matter, (4) explanations are social. These fit into the general theme that explanations are -contextual-.

Papers
F

FIRM

Feature importance ranking measure.

Papers
F

Fong, et. al.

Meaninful perturbations model.

Papers
G

G-REX

Rule extraction using genetic algorithms.

Papers
G

Gibbons, et. al.

Explain random forest using decision tree.

Papers
G

GoldenEye

Exploring classifiers by randomization.

Papers
G

GPD

Gaussian process decisions.

Papers
G

GPDT

Genetic program to evolve decision trees.

Papers
G

GradCAM

Gradient-weighted Class Activation Mapping.

Papers
G

GradCAM++

Generalized gradient-based visual explanations.

Papers
H

Hara, et. al.

Making tree ensembles interpretable.

Papers
I

ICE

Individual conditional expectation plots.

Papers
I

IG

Integrated gradients.

Papers
I

inTrees

Interpreting tree ensembles with inTrees.

Papers
I

IOFP

Iterative orthoganol feature projection.

Papers
I

IP

Information plane visualization.

Papers
K

k-Nearest Neighbors

The prototypical clustering method.

Papers
K

KL-LIME

Kullback-Leibler Projections based LIME.

Papers
K

Krishnan, et. al.

Extracting decision trees from trained neural networks.

Papers
L

Lei, et. al.

Rationalizing neural predictions with generator and encoder.

Papers
L

LIME

Local Interpretable Model-Agnostic Explanations.

Papers
L

Linear Regression

Easily plottable and understandable regression.

Papers
L

LOCO

Leave-one covariate out.

Papers
L

Logistic Regression

Easily plottable and understandable classification.

Papers
L

LORE

Local rule-based explanations.

Papers
L

Lou, et. al.

Accurate intelligibile models with pairwise interactions.

Papers
L

LRP

Layer-wise relevance propogation.

Papers
M

MCR

Model class reliance.

Papers
M

MES

Model explanation system.

Papers
M

MFI

Feature importance measure for non-linear algorithms.

Papers
N

Naive Bayes

Good classification, poor estimation using conditional probabilities.

Papers
N

NID

Neural interpretation diagram.

Papers
O

OptiLIME

Optimized LIME.

Papers
P

PALM

Partition aware local model.

Papers
P

PDA

Prediction Difference Analysis: Visualize deep neural network decisions.

Papers
P

PDP

Partial dependence plots.

Papers
P

Please Stop Permuting Features An Explanation a...

Authors demonstrate why permuting features is misleading, especially where there is strong feature dependence. They offer several previously described alternatives.

Papers
P

POIMs

Positional oligomer importance matrices for understanding SVM signal detectors.

Papers
P

ProfWeight

Transfer information from deep network to simpler model.

Papers
P

Prospector

Interactive partial dependence diagnostics.

Papers
Q

QII

Quantitative input influence.

Papers
Q

Quantifying Explainability of Saliency Methods ...

An analysis of how different heatmap-based saliency methods perform based on experimentation with a generated dataset.

Papers
R

REFNE

Extracting symbolic rules from trained neural network ensembles.

Papers
R

RETAIN

Reverse time attention model.

Papers
R

RISE

Randomized input sampling for explanation.

Papers
R

RuleFit

Sparse linear model as decision rules including feature interactions.

Papers
R

RxREN

Reverse engineering neural networks for rule extraction.

Papers
S

Sanity Checks for Saliency Maps

An important read for anyone using saliency maps. This paper proposes two experiments to determine whether saliency maps are useful: (1) model parameter randomization test compares maps from trained and untrained models, (2) data randomization test compares maps from models trained on the original dataset and models trained on the same dataset with randomized labels. They find that "some widely deployed saliency methods are independent of both the data the model was trained on, and the model parameters".

Papers
S

SHAP

A unified approach to interpretting model predictions.

Papers
S

SIDU

Similarity, difference, and uniqueness input perturbation.

Papers
S

Simonynan, et. al

Visualizing CNN classes.

Papers
S

Singh, et. al

Programs as black-box explanations.

Papers
S

STA

Interpreting models via Single Tree Approximation.

Papers
S

Stop Explaining Black Box Machine Learning Mode...

Authors present a number of issues with explainable ML and challenges to interpretable ML: (1) constructing optimal logical models, (2) constructing optimal sparse scoring systems, (3) defining interpretability and creating methods for specific methods. They also offer an argument for why interpretable models might exist in many different domains.

Papers
S

Strumbelj, et. al.

Explanation of individual classifications using game theory.

Papers
S

SVM+P

Rule extraction from support vector machines.

Papers
T

TCAV

Testing with concept activation vectors.

Papers
T

The (Un)reliability of Saliency Methods

Authors demonstrate how saliency methods vary attribution when adding a constant shift to the input data. They argue that methods should fulfill *input invariance*, that a saliency method mirror the sensistivity of the model with respect to transformations of the input.

Papers
T

Tolomei, et. al.

Interpretable predictions of tree-ensembles via actionable feature tweaking.

Papers
T

Tree Metrics

Making sense of a forest of trees.

Papers
T

TreeSHAP

Consistent feature attribute for tree ensembles.

Papers
T

TreeView

Feature-space partitioning.

Papers
T

TREPAN

Extracting tree-structured representations of trained networks.

Papers
T

TSP

Tree space prototypes.

Papers
V

VBP

Visual back-propagation.

Papers
V

VEC

Variable effect characteristic curve.

Papers
V

VIN

Variable interaction network.

Papers
X

X-TREPAN

Adapted etraction of comprehensible decision tree in ANNs.

Papers
X

Xu, et. al.

Show, attend, tell attention model.

Papers