Input Sentence: Bias . If you have a translation dictionary, use it to bias outputs (Arthur et al. 2016)
9
Previously Generated Things
10
Various Modalities
11
Multiple Sources
12
Coverage
13
Incorporating Markov Properties (Cohn et al. 2015)
14
Supervised Training (Mi et al. 2016)
15
Hard Attention
16
Multi-headed Attention
17
Attention Tricks
18
Training Tricks
Description:
Explore attention mechanisms in neural networks for natural language processing in this comprehensive lecture from CMU's Neural Networks for NLP course. Delve into various aspects of attention, including what to attend to, improvements to attention techniques, and specialized attention varieties. Examine a case study on the "Attention is All You Need" paper, and learn about attention score functions, input sentence handling, multi-headed attention, and training tricks. Gain insights into incorporating Markov properties, supervised training for attention, and hard attention concepts. This in-depth presentation covers essential topics for understanding and implementing attention in NLP models.