An Extended GHKM Algorithm for Inducing λ-SCFG Peng Li Tsinghua University.

Slides:



Advertisements
Similar presentations
Computational language: week 10 Lexical Knowledge Representation concluded Syntax-based computational language Sentence structure: syntax Context free.
Advertisements

CILC2011 A framework for structured knowledge extraction and representation from natural language via deep sentence analysis Stefania Costantini Niva Florio.
1© 2003 T. Abou-Assaleh, N. Cercone, & V. Keselj Expressing Probabilistic Context-Free Grammars in the Relaxed Unification Formalism Tony Abou-Assaleh.
Learning Accurate, Compact, and Interpretable Tree Annotation Recent Advances in Parsing Technology WS 2011/2012 Saarland University in Saarbrücken Miloš.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Learning Semantic Parsers Using Statistical.
Proceedings of the Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2007) Learning for Semantic Parsing Advisor: Hsin-His.
Adapting Discriminative Reranking to Grounded Language Learning Joohyun Kim and Raymond J. Mooney Department of Computer Science The University of Texas.
GRAMMAR & PARSING (Syntactic Analysis) NLP- WEEK 4.
1 Statistical NLP: Lecture 12 Probabilistic Context Free Grammars.
1 Unsupervised Semantic Parsing Hoifung Poon and Pedro Domingos EMNLP 2009 Best Paper Award Speaker: Hao Xiong.
The Query Compiler Varun Sud ID: 104. Agenda Parsing  Syntax analysis and Parse Trees.  Grammar for a simple subset of SQL  Base Syntactic Categories.
NLP and Speech Course Review. Morphological Analyzer Lexicon Part-of-Speech (POS) Tagging Grammar Rules Parser thethe – determiner Det NP → Det.
CS4705 Natural Language Processing.  Regular Expressions  Finite State Automata ◦ Determinism v. non-determinism ◦ (Weighted) Finite State Transducers.
C. Varela; Adapted w/permission from S. Haridi and P. Van Roy1 Declarative Computation Model Defining practical programming languages Carlos Varela RPI.
A Tree-to-Tree Alignment- based Model for Statistical Machine Translation Authors: Min ZHANG, Hongfei JIANG, Ai Ti AW, Jun SUN, Sheng LI, Chew Lim TAN.
Big Ideas in Cmput366. Search Blind Search Iterative deepening Heuristic Search A* Local and Stochastic Search Randomized algorithm Constraint satisfaction.
Sentence Classifier for Helpdesk s Anthony 6 June 2006 Supervisors: Dr. Yuval Marom Dr. David Albrecht.
Probabilistic Parsing Ling 571 Fei Xia Week 5: 10/25-10/27/05.
1 Learning to Interpret Natural Language Navigation Instructions from Observation Ray Mooney Department of Computer Science University of Texas at Austin.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
Machine Learning Group Department of Computer Sciences University of Texas at Austin Learning Language Semantics from Ambiguous Supervision Rohit J. Kate.
Week 9: resources for globalisation Finish spell checkers Machine Translation (MT) The ‘decoding’ paradigm Ambiguity Translation models Interlingua and.
A Natural Language Interface for Crime-related Spatial Queries Chengyang Zhang, Yan Huang, Rada Mihalcea, Hector Cuellar Department of Computer Science.
Develop a fast semantic decoder for dialogue systems Capability to parse 10 – 100 ASR hypotheses in real time Robust to speech recognition noise Semantic.
The Impact of Grammar Enhancement on Semantic Resources Induction Luca Dini Giampaolo Mazzini
Globalisation and machine translation Machine Translation (MT) The ‘decoding’ paradigm Ambiguity Translation models Interlingua and First Order Predicate.
1 Learning for Semantic Parsing Using Statistical Syntactic Parsing Techniques Ruifang Ge Ph.D. Final Defense Supervisor: Raymond J. Mooney Machine Learning.
Learning to Transform Natural to Formal Language Presented by Ping Zhang Rohit J. Kate, Yuk Wah Wong, and Raymond J. Mooney.
Some Probability Theory and Computational models A short overview.
Scalable Inference and Training of Context- Rich Syntactic Translation Models Michel Galley, Jonathan Graehl, Keven Knight, Daniel Marcu, Steve DeNeefe.
CS 3240: Languages and Computation Context-Free Languages.
Develop a fast semantic decoder Robust to speech recognition noise Trainable on different domains: Tourist information (TownInfo) Air travel information.
11 Chapter 14 Part 1 Statistical Parsing Based on slides by Ray Mooney.
Computing Science, University of Aberdeen1 CS4025: Logic-Based Semantics l Compositionality in practice l Producing logic-based meaning representations.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
October 2005CSA3180 NLP1 CSA3180 Natural Language Processing Introduction and Course Overview.
1 Intelligente Analyse- und Informationssysteme Frank Reichartz, Hannes Korte & Gerhard Paass Fraunhofer IAIS, Sankt Augustin, Germany Dependency Tree.
Semantic Construction lecture 2. Semantic Construction Is there a systematic way of constructing semantic representation from a sentence of English? This.
What’s in a translation rule? Paper by Galley, Hopkins, Knight & Marcu Presentation By: Behrang Mohit.
3.2 Semantics. 2 Semantics Attribute Grammars The Meanings of Programs: Semantics Sebesta Chapter 3.
The AI War LISP and Prolog Basic Concepts of Logic Programming
Artificial Intelligence: Natural Language
For Monday Read chapter 26 Last Homework –Chapter 23, exercise 7.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Learning a Compositional Semantic Parser.
For Friday Finish chapter 23 Homework –Chapter 23, exercise 15.
Probabilistic Context Free Grammars Grant Schindler 8803-MDM April 27, 2006.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
11 Project, Part 3. Outline Basics of supervised learning using Naïve Bayes (using a simpler example) Features for the project 2.
Natural Language Generation with Tree Conditional Random Fields Wei Lu, Hwee Tou Ng, Wee Sun Lee Singapore-MIT Alliance National University of Singapore.
Develop a fast semantic decoder for dialogue systems Capability to parse 10 – 100 ASR hypothesis in real time Robust to speech recognition noise Trainable.
Wei Lu, Hwee Tou Ng, Wee Sun Lee National University of Singapore
For Monday Read chapter 26 Homework: –Chapter 23, exercises 8 and 9.
PPL Syntax & Formal Semantics Lecture Notes: Chapter 2.
Natural Language Processing : Probabilistic Context Free Grammars Updated 8/07.
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
PRESENTED BY: PEAR A BHUIYAN
Semantic Parsing for Question Answering
--Mengxue Zhang, Qingyang Li
Presentation by Julie Betlach 7/02/2009
Programming Language Syntax 7
Implementing Language Extensions with Model Transformations
Knowledge Representation and Inference
Learning to Parse Database Queries Using Inductive Logic Programming
Learning to Sportscast: A Test of Grounded Language Acquisition
CS4705 Natural Language Processing
المشرف د.يــــاســـــــــر فـــــــؤاد By: ahmed badrealldeen
BASICS OF ENGLISH GRAMMAR
Owen Rambow 6 Minutes.
Presentation transcript:

An Extended GHKM Algorithm for Inducing λ-SCFG Peng Li Tsinghua University

Semantic Parsing Mapping natural language (NL) sentence to its computable meaning representation (MR) NL: Every boy likes a star MR: variable predicate

Motivation Common way: inducing probabilistic grammar PCFG: Probabilistic Context Free Grammar

Motivation Common way: inducing probabilistic grammar CCG: Combinatory Categorial Grammar

Motivation Common way: inducing probabilistic grammar SCFG: Synchronous Context Free Grammar

Motivation State of the art: SCFG + λ-calculus (λ-SCFG) Major challenge: grammar induction – It is much harder to find the correspondence between NL sentence and MR than between NL sentences SCFG rule extraction is well-studied in MT GHKM is the most widely used algorithm We want to adapt GHKM to semantic parsing Experimental results show that we get the state- of-the-art performance

Background State of the art: SCFG + λ-calculus (λ-SCFG) λ-calculus – λ-expression: – β-conversion: bound variable substitution – α-conversion: bound variable renaming

λ-SCFG Rule Extraction Outline 1.Building training examples 1.Transforming logical forms to trees 2.Aligning trees with sentences 2.Identifying frontier nodes 3.Extracting minimal rules 4.Extracting composed rules

Building Training Examples NL: Every boy likes a star MR:

Building Training Examples

boy human pop like

Building Training Examples boy human pop like Every boy likes a star

Identifying Frontier Nodes

Identifying Minimal Frontier Tree

Minimal Rule Extraction X X

X X

X X

Composed Rule Extraction

λ-SCFG Rule Extraction Outline 1.Building training examples 1.Transforming logical forms to trees 2.Aligning trees with sentences 2.Identifying frontier nodes 3.Extracting minimal rules 4.Extracting composed rules

Modeling Log-linear model + MERT training Target

Parsing Type checking (Wong and Mooney, 2007)

Experiments Dataset: G EOQUERY – 880 English questions with corresponding Prolog logical form – Metric

Experiments SCFG PCFG CCG

Experiments F-measure for different languages * en - English, ge - German, el - Greek, th - Thai

Experiments

Experiments