Explainable Artificial Intelligence
(Autumn 2021)
(Autumn 2021)
General Concepts / Surveys
Hidden neuron analysis methods:
Visualize, revert-map or label features that are learned by hidden neurons
GAN
Maximum Activation
Qualitatively analyzing every neuron does not provide much actionable and quantitative interpretation about the overall mechanism of the whole model
Model mimicking methods: imitate the classification function of a target model and build a transparent model that is easy to interpret and has high accuracy
Due to the reduced model complexity of a mimic model, no guarantee that a deep model with a large VC dimension can be successfully imitated by a simpler shallow model
Local interpretation methods:
Compute and visualize the important features for an input instance by analyzing the predictions of its local perturbations
LIME
Saliency map
Perceptively indistinguishable instances may not be explained consistently