Главная
Study mode:
on
1
Intro
2
What is Whisper
3
Example Implementation
4
Weekly supervise
5
Finetuning
6
Mixing Tasks
7
Data Quality
8
Model
9
Pipeline
10
Generalization
11
Overfitting
12
Model size
13
Multitask performance
Description:
Explore OpenAI's Whisper, a groundbreaking speech-to-text model capable of transcribing and translating 97 languages. Learn about its weakly supervised encoder-decoder transformer architecture, trained on 680,000 hours of audio. Discover the model's implementation, fine-tuning process, and multitask capabilities. Delve into topics such as data quality, pipeline structure, generalization, overfitting prevention, and the impact of model size on performance. Gain insights into the weekly supervise technique and how mixing tasks contributes to the model's versatility.

Open AI's Whisper Is Amazing

sentdex
Add to list