Главная
Study mode:
on
1
Intro
2
My goal interpretability
3
NON-goals
4
Investigating
5
Sanity check question.
6
Benchmarking interpretability methods (BIM)
7
Three metrics for measuring false positives
8
Model Contrast Score (MCS)
9
Defining concept activation vector (CAV) Inputs
10
TCAV core idea: Derivative with CAV to get prediction sensitivity
11
Quantitative validation: Guarding against spurious CAV
12
Recap TCAV: Testing with Concept Activation Vectors
13
Sanity check experiment setup
14
Human subject experiment: Can saliency maps communicate the same information?
15
TCAV in Two widely used image prediction models
16
Collect human doctor's knowledge
17
TCAV for Diabetic Retinopathy
18
Summary: Testing with Concept Activation Vectors
19
Responses from inside of academia
20
Limitations of TCAV
21
Things to keep in mind during our journey
Description:
Explore the frontiers of deep learning in this 48-minute talk by Been Kim from Google Brain, focusing on interpretability in machine learning. Delve into the goals and non-goals of interpretability, and learn about benchmarking interpretability methods (BIM) using three metrics for measuring false positives. Discover the Model Contrast Score (MCS) and the concept of Concept Activation Vectors (CAV). Examine the TCAV (Testing with Concept Activation Vectors) approach, including its core ideas, quantitative validation, and applications in image prediction models and medical diagnosis. Gain insights from human subject experiments comparing saliency maps, and understand the limitations and considerations of TCAV. Reflect on the broader implications and challenges in the field of interpretable machine learning.

Interpretability - Now What?

Simons Institute
Add to list