Important note: in CV we are testing pipeline, not a single model
8
K-fold, folded test set
9
K-fold, hold-out test-set
10
Nested cross-validation
11
leave-one-out
12
random sampling
13
selecting an approach: pros and cons
14
Final thoughts
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Grab it
Explore cross-validation techniques for AI in this comprehensive 49-minute video lecture by Dr. Tyler Bradshaw from Molecular Imaging & Therapy. Delve into the concepts of overfitting and generalizability, and learn about the pitfalls of using one-time split methods. Understand the importance of representative test sets and avoiding tuning to the test set. Discover various cross-validation approaches, including K-fold with folded and hold-out test sets, nested cross-validation, leave-one-out, and random sampling. Gain insights on selecting the most appropriate approach by weighing their pros and cons. The lecture concludes with final thoughts and references a paper for further study, providing a solid foundation for implementing effective cross-validation techniques in AI projects.
A Guide to Cross-Validation for AI - Avoiding Overfitting and Ensuring Generalizability