Главная
Study mode:
on
1
Intro
2
Outline
3
Agenda
4
Bird Paper
5
Architecture
6
Problems
7
Adapters
8
Modularity
9
Compositions
10
Overview
11
Function Composition
12
Input Composition
13
Parameter Composition
14
Fusion
15
Hyper Networks
16
Shared Hyper Networks
17
Chad GP
18
Questions
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Learn about multi-task learning in transformer-based NLP architectures through this 31-minute conference talk that explores cost-effective alternatives to training separate models. Discover how leveraging information across multiple tasks and datasets can enhance performance through shared models, representation bias, increased data efficiency, and eavesdropping. Explore solutions to challenges like catastrophic forgetting and interference, while diving into general approaches to multi-task learning, innovative adapter-based techniques, hypernetwork methods, and strategies for task sampling and balancing. The presentation covers key topics including the Bird Paper, architecture considerations, modularity concepts, function composition, input composition, parameter composition, fusion techniques, and shared hypernetworks, concluding with insights into Chad GP implementations.

Multi-Task Learning in Transformer-Based Architectures for Natural Language Processing

Data Science Conference
Add to list
0:00 / 0:00