Главная
Study mode:
on
1
OpenAI's CLIP
2
Detailed explanation of the method
3
Comparision with SimCLR
4
How does the zero-shot part work
5
WIT dataset
6
Why this method, hint efficiency
7
Zero-shot - generalizing to new tasks
8
Prompt programming and ensembling
9
Zero-shot perf
10
Few-shot comparison with best baselines
11
How good the zero-shot classifier is?
12
Compute error correlation
13
Quality of CLIP's embedding space
14
Robustness to distribution shift
15
Limitations MNIST failure
16
A short recap
Description:
Dive into a comprehensive 53-minute video lecture exploring OpenAI's CLIP (Contrastive Language-Image Pre-training) model. Learn about the contrastive learning approach behind CLIP, its comparison with SimCLR, and the intricacies of zero-shot learning. Explore the WIT dataset, prompt programming, and embedding space quality. Analyze CLIP's performance in few-shot learning scenarios, its robustness to distribution shifts, and potential limitations. Gain insights into this innovative approach connecting text and images through natural language supervision.

OpenAI CLIP - Connecting Text and Images - Paper Explained

Aleksa Gordić - The AI Epiphany
Add to list