Главная
Study mode:
on
1
Intro
2
Content
3
Abstract Image Saliency Methods Summary Attention Map Limited by heuristic properties and architectural constraints
4
Introduction Current Problems The interpretation for the black box predictor The intuitive visualization method is only heuristic, and the meaning remains unclear.
5
Contribution Develop principles and methods to explain any black box function By determine mapping attributes - Internal mechanisms is used to implement these attributes
6
Related Work Gradient-based method -Backpropagates the gradient for a class label to the image layer Other methods: DeConvNet, Guided Backprop
7
Related Work - CAM
8
Related Work Comparison
9
Comparison with other saliency methods
10
Principle Black bax is a mapping function
11
Explanations as meta-predictors Rules are used to explain a robin classifier
12
Advantages of Explanations as Meta-predictors The faithfulness of images can be measured as prediction accuracy To find the explanations automatically
13
Local Explanations
14
Saliency Deleting parts of image x, as the perturbations for the whole image X
15
A Meaningful Image Perturbation 11
16
Deletion and Preservation
17
Artifacts Reduction
18
Experiment-Interpretability
19
Experiment Testing hypotheses: animal part saliency
20
Experiment-Adversarial defense
21
Experiment localization and pointing
22
Conclusion
23
Questions?
Description:
Explore a 27-minute lecture on interpretable explanations of black box models using meaningful perturbation. Delve into current challenges in interpreting black box predictors and the limitations of intuitive visualization methods. Learn about principles and methods developed to explain any black box function by determining mapping attributes and internal mechanisms. Compare various saliency methods, including gradient-based approaches and Class Activation Mapping (CAM). Understand the concept of explanations as meta-predictors and their advantages in measuring image faithfulness and finding automated explanations. Examine local explanations, meaningful image perturbations, and techniques for deletion and preservation. Discover experiments on interpretability, animal part saliency, adversarial defense, and localization. Gain insights into the development of more transparent and understandable machine learning models.

Interpretable Explanations of Black Boxes by Meaningful Perturbation - CAP6412 Spring 2021

University of Central Florida
Add to list
0:00 / 0:00