A First Attempt: Sequence-to- sequence Models (Jia and Liang 2016)
8
A Better Attempt: Tree-based Parsing Models • Generate from top-down using hierarchical sequence- to-sequence model (Dong and Lapata 2016)
9
Meaning Representation Desiderata (Jurafsky and Martin 17.1)
10
First-order Logic
11
Abstract Meaning Representation (Banarescu et al. 2013)
12
Syntax-driven Semantic Parsing
13
CCG and CCG Parsing
14
Parsing to Graph Structures
15
Semantic Role Labeling (Gildea and Jurafsky 2002)
16
Neural Models for Semantic Role Labeling
Description:
Explore neural semantic parsing in this comprehensive lecture from CMU's Neural Networks for NLP course. Delve into graph-based parsing, minimum spanning tree parsing, and structured training techniques. Learn about dynamic programming methods for phrase structure parsing and reranking. Access accompanying slides and code examples to reinforce understanding of key concepts. Gain insights into semantic parsing approaches, including sequence-to-sequence models and tree-based parsing models. Examine various meaning representations such as first-order logic and Abstract Meaning Representation. Investigate syntax-driven semantic parsing, CCG parsing, and semantic role labeling with neural models.
Neural Nets for NLP 2017 - Neural Semantic Parsing