Arc Standard Shift-Reduce Parsing (Yamada & Matsumoto 2003, Nivre 2003)
4
Shift Reduce Example
5
Classification for Shift-reduce
6
Making Classification Decisions
7
What Features to Extract?
8
Non-linear Function: Cube Function
9
Why Tree Structure?
10
Tree-structured LSTM (Tai et al. 2015)
11
Encoding Parsing Configurations w/ RNNS
12
A Simple Approximation: Linearized Trees (Vinyals et al. 2015)
13
Recursive Neural Networks (Socher et al. 2011)
Description:
Explore transition-based dependency parsing in this comprehensive lecture from CMU's Neural Networks for NLP course. Delve into the fundamentals of transition-based parsing, including shift-reduce parsing with feed-forward networks and stack LSTMs. Learn about a simple alternative approach using linearized trees. Gain insights into various parsing techniques, feature extraction methods, and the importance of tree structures in natural language processing. Examine practical code examples and follow along with detailed slides to reinforce your understanding of these advanced NLP concepts.
Neural Nets for NLP 2017 - Transition-Based Dependency Parsing