Главная
Study mode:
on
1
Introduction
2
Overfitting vs. generalizability
3
Pitfalls of using one-time split method
4
Pitfall #1: Non-representative test set
5
Pitfall #2: Tuning to the test set
6
Cross-validation
7
Important note: in CV we are testing pipeline, not a single model
8
K-fold, folded test set
9
K-fold, hold-out test-set
10
Nested cross-validation
11
leave-one-out
12
random sampling
13
selecting an approach: pros and cons
14
Final thoughts
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Explore cross-validation techniques for AI in this comprehensive 49-minute video lecture by Dr. Tyler Bradshaw from Molecular Imaging & Therapy. Delve into the concepts of overfitting and generalizability, and learn about the pitfalls of using one-time split methods. Understand the importance of representative test sets and avoiding tuning to the test set. Discover various cross-validation approaches, including K-fold with folded and hold-out test sets, nested cross-validation, leave-one-out, and random sampling. Gain insights on selecting the most appropriate approach by weighing their pros and cons. The lecture concludes with final thoughts and references a paper for further study, providing a solid foundation for implementing effective cross-validation techniques in AI projects.

A Guide to Cross-Validation for AI - Avoiding Overfitting and Ensuring Generalizability

Molecular Imaging & Therapy
Add to list