Главная
Study mode:
on
1
- Intro
2
- What are sparse expert models?
3
- Start of Interview
4
- What do you mean by sparse experts?
5
- How does routing work in these models?
6
- What is the history of sparse experts?
7
- What does an individual expert learn?
8
- When are these models appropriate?
9
- How comparable are sparse to dense models?
10
- How does the pathways system connect to this?
11
- What improvements did GLAM make?
12
- The "designing sparse experts" paper
13
- Can experts be frozen during training?
14
- Can the routing function be improved?
15
- Can experts be distributed beyond data centers?
16
- Are there sparse experts for other domains than NLP?
17
- Are sparse and dense models in competition?
18
- Where do we go from here?
19
- How can people get started with this?
Description:
Explore the world of Sparse Expert Models in this comprehensive interview with Google Brain researchers Barret Zoph and William Fedus. Delve into the fundamentals, history, strengths, and weaknesses of these innovative models, including Switch Transformers and GLAM, which can scale up to trillions of parameters. Learn how sparse expert models distribute parts of Transformers across large arrays of machines, using routing functions to efficiently activate only specific parts of the model. Discover the advantages of this approach, its applications in natural language processing, and potential future developments. Gain insights into the comparison between sparse and dense models, the improvements made by GLAM, and the possibilities of distributing experts beyond data centers. Whether you're a machine learning enthusiast or a seasoned researcher, this in-depth discussion provides valuable knowledge on the current state of the art in sparse expert models and their potential impact on the field of artificial intelligence. Read more

Sparse Expert Models - Switch Transformers, GLAM, and More With the Authors

Yannic Kilcher
Add to list