Главная
Study mode:
on
1
Intro
2
Why Model Sentence Pairs?
3
Siamese Network (Bromley et al. 1993)
4
Convolutional Matching Model (Hu et al. 2014) • Concatenate sentences into a 30 tensor and perform convolution
5
Convolutional Features + Matrix-based Pooling in and Schutze 2015
6
NLP and Sequential Data
7
Long-distance Dependencies in Language
8
Can be Complicated!
9
Recurrent Neural Networks (Elman 1990)
10
Unrolling in Time • What does processing a sequence look like?
11
What Can RNNs Do?
12
Representing Sentences
13
e.g. Language Modeling
14
RNNLM Example: Loss Calculation and State Update
15
Vanishing Gradient • Gradients decrease as they get pushed back
16
LSTM Structure
17
What can LSTMs Learn? (2) (Shi et al. 2016, Radford et al. 2017) Count length of sentence
18
Handling Long Sequences
Description:
Explore recurrent neural networks for sentence and language modeling in this comprehensive lecture from CMU's Neural Networks for NLP course. Dive into the structure and capabilities of RNNs, understand the vanishing gradient problem and how LSTMs address it, and examine the strengths and weaknesses of recurrence in sentence modeling. Learn about pre-training techniques for RNNs and gain insights into handling long sequences and long-distance dependencies in language processing. Discover practical applications like language modeling and sentence representation through detailed examples and explanations.

Neural Nets for NLP - Recurrent Networks for Sentence or Language Modeling

Graham Neubig
Add to list