Adversarial Examples and Human-ML Alignment Aleksander Madry
2
Deep Networks: Towards Human Vision?
3
A Natural View on Adversarial Examples
4
Why Are Adv. Perturbations Bad?
5
Human Perspective
6
ML Perspective
7
The Robust Features Model
8
The Simple Experiment: A Second Look
9
Human vs ML Model Priors
10
In fact, models...
11
Consequence: Interpretability
12
Consequence: Training Modifications
13
Consequence: Robustness Tradeoffs
14
Robustness + Perception Alignment
15
Robustness + Better Representations
16
Problem: Correlations can be weird
17
"Counterfactual" Analysis with Robust Models
18
Adversarial examples arise from non-robust features in the data
Description:
Explore the concept of adversarial examples and human-machine learning alignment in this lecture by Aleksander Madry from MIT. Delve into the comparison between deep networks and human vision, examining the natural perspective on adversarial examples. Investigate why adversarial perturbations are problematic from both human and machine learning viewpoints. Analyze the robust features model and its implications for interpretability, training modifications, and robustness tradeoffs. Discover how robustness relates to perception alignment and improved representations. Address the challenge of unusual correlations in data and learn about counterfactual analysis using robust models. Gain insights into the origin of adversarial examples stemming from non-robust features in datasets.