Главная
Study mode:
on
1
Announcements
2
It's hard to opt-out
3
Confidence scores
4
Explanations in plain English free-text / chain-of-thoughts
5
Input attribution gradient-based & select-then-predict
6
Feature interactions effective attention
7
Concept-based explanations TCAV
8
Data influence influence functions
9
Contrastive explanations contrastive editing
10
Explainability as a dialog
11
Taxonomy of evaluation of explanations
12
Simulatability
13
Why are application-grounded evals of explanations scarce in NLP?
14
Application-grounded evaluations
15
Trust in AI
Description:
Learn about various local explainability methods and their evaluation in this comprehensive lecture from UofU Data Science. Explore key concepts starting with the challenges of opting out and understanding confidence scores in AI systems. Dive into different explanation approaches including plain English explanations with free-text and chain-of-thoughts, input attribution methods using gradient-based and select-then-predict techniques, and feature interactions through effective attention. Examine concept-based explanations using TCAV, data influence through influence functions, and contrastive explanations via contrastive editing. Understand explainability as a dialog and explore the taxonomy of explanation evaluation, including simulatability. Address the scarcity of application-grounded evaluations in NLP and conclude with a discussion on building trust in AI systems. The lecture provides a thorough examination of how to make AI systems more interpretable and trustworthy through various explainability methods. Read more

An Overview of Local Explainability Methods and Their Evaluation in AI

UofU Data Science
Add to list
0:00 / 0:00