Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence,

Slides:



Advertisements
Similar presentations
Iowa State University Department of Computer Science Center for Computational Intelligence, Learning, and Discovery Harris T. Lin and Vasant Honavar. BigData2013.
Advertisements

Grammars, constituency and order A grammar describes the legal strings of a language in terms of constituency and order. For example, a grammar for a fragment.
Incrementally Learning Parameter of Stochastic CFG using Summary Stats Written by:Brent Heeringa Tim Oates.
Iterative Optimization and Simplification of Hierarchical Clusterings Doug Fisher Department of Computer Science, Vanderbilt University Journal of Artificial.
Simplifying CFGs There are several ways in which context-free grammars can be simplified. One natural way is to eliminate useless symbols those that cannot.
About Grammars CS 130 Theory of Computation HMU Textbook: Sec 7.1, 6.3, 5.4.
Omphalos Session. Omphalos Session Programme Design & Results 25 mins Award Ceremony 5 mins Presentation by Alexander Clark25 mins Presentation by Georgios.
10. Lexicalized and Probabilistic Parsing -Speech and Language Processing- 발표자 : 정영임 발표일 :
1 Statistical NLP: Lecture 12 Probabilistic Context Free Grammars.
Kewei Tu and Vasant Honavar
March 1, 2009 Dr. Muhammed Al-Mulhem 1 ICS 482 Natural Language Processing Probabilistic Context Free Grammars (Chapter 14) Muhammed Al-Mulhem March 1,
In Search of a More Probable Parse: Experiments with DOP* and the Penn Chinese Treebank Aaron Meyers Linguistics 490 Winter 2009.
CKY Parsing Ling 571 Deep Processing Techniques for NLP January 12, 2011.
Re-ranking for NP-Chunking: Maximum-Entropy Framework By: Mona Vajihollahi.
PCFG Parsing, Evaluation, & Improvements Ling 571 Deep Processing Techniques for NLP January 24, 2011.
Parsing with CFG Ling 571 Fei Xia Week 2: 10/4-10/6/05.
Parsing with PCFG Ling 571 Fei Xia Week 3: 10/11-10/13/05.
CS Master – Introduction to the Theory of Computation Jan Maluszynski - HT Lecture 4 Context-free grammars Jan Maluszynski, IDA, 2007
Grammar induction by Bayesian model averaging Guy Lebanon LARG meeting May 2001 Based on Andreas Stolcke’s thesis UC Berkeley 1994.
Syntactic Parsing with CFGs CMSC 723: Computational Linguistics I ― Session #7 Jimmy Lin The iSchool University of Maryland Wednesday, October 14, 2009.
Multiple Sequence alignment Chitta Baral Arizona State University.
Normal forms for Context-Free Grammars
Maximum Entropy Model LING 572 Fei Xia 02/07-02/09/06.
Parsing SLP Chapter 13. 7/2/2015 Speech and Language Processing - Jurafsky and Martin 2 Outline  Parsing with CFGs  Bottom-up, top-down  CKY parsing.
Context-Free Grammars Chapter 3. 2 Context-Free Grammars and Languages n Defn A context-free grammar is a quadruple (V, , P, S), where  V is.
More on Text Management. Context Free Grammars Context Free Grammars are a more natural model for Natural Language Syntax rules are very easy to formulate.
Probabilistic Parsing Ling 571 Fei Xia Week 5: 10/25-10/27/05.
Context-Free Grammar CSCI-GA.2590 – Lecture 3 Ralph Grishman NYU.
11 CS 388: Natural Language Processing: Syntactic Parsing Raymond J. Mooney University of Texas at Austin.
Lecture 21: Languages and Grammars. Natural Language vs. Formal Language.
Tree Kernels for Parsing: (Collins & Duffy, 2001) Advanced Statistical Methods in NLP Ling 572 February 28, 2012.
Unambiguity Regularization for Unsupervised Learning of Probabilistic Grammars Kewei TuVasant Honavar Departments of Statistics and Computer Science University.
1 Statistical Parsing Chapter 14 October 2012 Lecture #9.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 16 Nov, 3, 2011 Slide credit: C. Conati, S.
SI485i : NLP Set 8 PCFGs and the CKY Algorithm. PCFGs We saw how CFGs can model English (sort of) Probabilistic CFGs put weights on the production rules.
Some Probability Theory and Computational models A short overview.
Unsupervised learning of Natural languages Eitan Volsky Yasmine Meroz.
Context Free Grammars CIS 361. Introduction Finite Automata accept all regular languages and only regular languages Many simple languages are non regular:
Parsing I: Earley Parser CMSC Natural Language Processing May 1, 2003.
Membership problem CYK Algorithm Project presentation CS 5800 Spring 2013 Professor : Dr. Elise de Doncker Presented by : Savitha parur venkitachalam.
11 Chapter 14 Part 1 Statistical Parsing Based on slides by Ray Mooney.
Page 1 Probabilistic Parsing and Treebanks L545 Spring 2000.
Coarse-to-Fine Efficient Viterbi Parsing Nathan Bodenstab OGI RPE Presentation May 8, 2006.
CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 29– CYK; Inside Probability; Parse Tree construction) Pushpak Bhattacharyya CSE.
Chapter 23: Probabilistic Language Models April 13, 2004.
PARSING 2 David Kauchak CS159 – Spring 2011 some slides adapted from Ray Mooney.
Quick Speech Synthesis CMSC Natural Language Processing April 29, 2003.
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture-14: Probabilistic parsing; sequence labeling, PCFG.
CS 4705 Lecture 10 The Earley Algorithm. Review Top-Down vs. Bottom-Up Parsers –Both generate too many useless trees –Combine the two to avoid over-generation:
School of Computer Science 1 Information Extraction with HMM Structures Learned by Stochastic Optimization Dayne Freitag and Andrew McCallum Presented.
1Computer Sciences Department. Book: INTRODUCTION TO THE THEORY OF COMPUTATION, SECOND EDITION, by: MICHAEL SIPSER Reference 3Computer Sciences Department.
Iowa State University Department of Computer Science Center for Computational Intelligence, Learning, and Discovery Harris T. Lin, Sanghack Lee, Ngot Bui.
2003 (c) University of Pennsylvania1 Better MT Using Parallel Dependency Trees Yuan Ding University of Pennsylvania.
Probabilistic Context Free Grammars Grant Schindler 8803-MDM April 27, 2006.
CPSC 422, Lecture 27Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 27 Nov, 16, 2015.
December 2011CSA3202: PCFGs1 CSA3202: Human Language Technology Probabilistic Phrase Structure Grammars (PCFGs)
Virtual Examples for Text Classification with Support Vector Machines Manabu Sassano Proceedings of the 2003 Conference on Emprical Methods in Natural.
NLP. Introduction to NLP Time flies like an arrow –Many parses –Some (clearly) more likely than others –Need for a probabilistic ranking method.
CS : Speech, NLP and the Web/Topics in AI Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture-15: Probabilistic parsing; PCFG (contd.)
Exercises on Chomsky Normal Form and CYK parsing
CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 25– Probabilistic Parsing) Pushpak Bhattacharyya CSE Dept., IIT Bombay 14 th March,
Natural Language Processing : Probabilistic Context Free Grammars Updated 8/07.
Roadmap Probabilistic CFGs –Handling ambiguity – more likely analyses –Adding probabilities Grammar Parsing: probabilistic CYK Learning probabilities:
General Information on Context-free and Probabilistic Context-free Grammars İbrahim Hoça CENG784, Fall 2013.
1 Statistical methods in NLP Diana Trandabat
CSC 594 Topics in AI – Natural Language Processing
Boosted Augmented Naive Bayes. Efficient discriminative learning of
Parsing and More Parsing
Parsing I: CFGs & the Earley Parser
Presentation transcript:

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Unsupervised Learning of Probabilistic Context-Free Grammar Using Iterative Biclustering Kewei Tu and Vasant Honavar Artificial Intelligence Research Laboratory Department of Computer Science Iowa State University

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Unsupervised Learning of Probabilistic Context-Free Grammar Greedy search to maximize the posterior of the grammar given the corpus Iterative (distributional) biclustering Competitive experimental results

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Outline Introduction Probabilistic Context Free Grammars (PCFG) The Algorithm based on Iterative Biclustering (PCFG-BCL) Experimental results

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Outline Introduction Probabilistic Context Free Grammars (PCFG) The Algorithm based on Iterative Biclustering (PCFG-BCL) Experimental results

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Motivation Probabilistic Context-Free Grammar (PCFG) find applications in many areas including: Natural Language Processing Bioinformatics Important to learn PCFG from data (training corpus) Labeled corpus not always available Hence the need for unsupervised learning

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Task Unsupervised learning of a PCFG from a positive corpus a square is above the triangle the square rolls a triangle rolls the square rolls a triangle is above the square a circle touches a square the triangle covers the circle …… S  NP VP NP  Det N VP  Vt NP (0.3) | Vi PP (0.2) | rolls (0.2) | bounces (0.1) ……

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Outline Introduction Probabilistic Context Free Grammars (PCFG) The Algorithm based on Iterative Biclustering (PCFG-BCL) Experimental results

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar PCFG Context-free Grammar (CFG) G = (N, Σ, R, S) N: non-terminals Σ: terminals R: rules S  N : the start symbol Probabilistic CFG Probabilities on grammar rules

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar P-CNF Probabilistic Chomsky normal form (P-CNF) Two types of rules: A  BC A  a

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar The AND-OR form P-CNF in the AND-OR form Two types of non-terminals: AND, OR AND  OR 1 OR 2 OR  A 1 | A 2 | a 1 | a 2 | …… with probabilities

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar The AND-OR form P-CNF in the AND-OR form

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar The AND-OR form P-CNF in the AND-OR form can be divided into two parts Start rules S  … A set of AND-OR groups Each group: AND  OR1 OR2 Bijection between ANDs and groups An OR may appear in multiple groups

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar The AND-OR form P-CNF in the AND-OR form can be divided into two parts

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Outline Introduction Probabilistic Context Free Grammars (PCFG) The Algorithm based on Iterative Biclustering (PCFG-BCL) Experimental results

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar PCFG-BCL: Outline Start with only the terminals Repeat the two steps Learn a new AND-OR group by biclustering Attach the new AND to existing ORs Post-processing: add start rules In principle, these steps are sufficient for learning any CNF grammar

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar PCFG-BCL: Outline Find new rules that yield the greatest increase in the posterior of the grammar given the corpus Local search, with the posterior as the objective function Use a prior that favors simpler grammars to avoid overfitting

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar PCFG-BCL Repeat the two steps Learn a new AND-OR group by biclustering Attach the new AND to existing ORs Post-processing: add start rules

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Intuition Construct a table T Index the rows and columns by symbols appearing in the corpus The cell at row x and column y records the number of times the pair xy appears in the corpus

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar An AND-OR group corresponds to a bicluster

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar The bicluster is multiplicatively coherent for any two rows i,j and two columns k,l

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Expression-context matrix of a bicluster Each row: a symbol pair contained in the bicluster Each column: a context in which the symbol pairs appear in the corpus It’s also multiplicatively coherent.

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Intuition If there’s a bicluster that is multiplicatively coherent and has a multiplicatively coherent expression-context matrix Then an AND-OR group can be learned from the bicluster

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Probabilistic Justification Change in likelihood as a result of adding an AND-OR group to a PCFG Bicluster multiplicative coherence Expression-context matrix multiplicative coherence

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Prior To prevent overfitting, use a prior that favors simpler grammars P(G)  2  DL(G) DL(G) is the description length of the grammar

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Learning a new AND-OR group by biclustering find in the table T a bicluster that leads to the maximal posterior gain create a new AND-OR group from the bicluster reduce the corpus using the new rules E.g., “the circle” is rewritten to the new AND symbol update T A new row and column are added for the new AND symbol

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar PCFG-BCL Repeat the two steps Learn a new AND-OR group by biclustering Attach the new AND to existing ORs Post-processing: add start rules

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Attaching the new AND under existing ORs For the new AND symbol N … There may exist OR symbols in the learned grammar, s.t. O  N is in the target grammar Such rules can't be learned in the biclustering step When learning O, N doesn’t exist When learning N, only learn N  AB We need an additional step to find such rules Recursion is learned in this step

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Intuition Adding rule O  N = adding a new row/column to the bicluster If O  N is true, then the expanded bicluster is multiplicatively coherent the expanded expression-context matrix is multiplicatively coherent If we find an OR symbol s.t. the expanded bicluster has this property Then a new rule O  N can be added to the grammar

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Probabilistic Justification Likelihood gain is an approximation of the expanded bicluster To prevent overfitting, the prior is also considered

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Attaching the new AND under existing ORs Try to find OR symbols that lead to large posterior gain When found add the new rule O  N to the grammar do a maximal reduction of the corpus update the table T

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar PCFG-BCL Repeat the two steps Learn a new AND-OR group by biclustering Attach the new AND to existing ORs Post-processing: add start rules

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Postprocessing For each sentence in the corpus: If it’s fully reduced to a single symbol x, then add S  x If not, a few options… Return the grammar

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Outline Introduction Probabilistic Context Free Grammars (PCFG) The Algorithm based on Iterative Biclustering (PCFG-BCL) Experimental results

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Experiments Measurements weak generative capacity precision, recall, F-score Test data artificial, English-like CFGs

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Experiment results P=Precision, R=Recall, F=F-score Number in the parentheses: standard deviation PCFG-BCL outperforms EMILE and ADIOS with lower standard deviations [Adriaans, et al., 2000][Solan, et al., 2005]

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Summary An unsupervised PCFG-learning algorithm It acquires new grammar rules by iterative biclustering on a table of symbol pairs In each step it tries to maximize the increase of the posterior of the grammar Competitive experimental results

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Work in progress Alternative strategies for optimizing the objective function Evaluation on and adaptation to real world applications (e.g., natural language), wrt. both weak and strong generative capacity

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Thank you~

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Backup…

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Step 1 Bicluster multiplicative coherence E-C matrix multiplicative coherence Prior gain (bias towards large BC) Likelihood Gain Posterior gain:

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Step 2 Intuition Remember O is learned by extracting a bicluster adding rule O  N = adding a new row/column to the bicluster

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Expanding the bicluster The expanded bicluster should still be multiplicatively coherent

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Step 2 Intuition Expression-context matrix adding rule O  N = adding a set of new rows to the E-C matrix

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Expanding the expression-context matrix The expanded expression-context matrix should still be multiplicatively coherent.

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Step 2 Likelihood gain: : the expected numbers of appearance of the symbol pairs when applying the current grammar to expand the current partially reduced corpus.

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Grammar selection/averaging Run the algorithm for multiple times to get multiple grammars Use the posterior of the grammars to do model selection/averaging Experimental results: Improved the performance Decreased the standard deviations

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Time Complexity N: # of ANDs k: average # of rules headed by an OR c: average column# of Expr-Cont Matrix h: average # of ORs that produce an AND or terminal d: a recursion depth limit ω: sentence# in the corpus m: average sentence length

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar biclustering vs. distributional clustering V1  makes | likes V2  likes | is Figure from [Adriaans, et al., 2000]

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar biclustering vs. substitutability heuristic N1  tea | coffee N2  eating Figure from [Adriaans, et al., 2000]

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar A set of multiplicatively coherent biclusters, which represent a set of AND-OR groups in the grammar.

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Related work Unsupervised CFG learning EMILE [Adriaans et al., 2000] ABL [Zaanen, 2000] [Clark, 2001; 2007] ADIOS [Solan et al., 2005] Main difference Distributional biclustering A unified method for different types of rules

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Related work Unsupervised PCFG learning Inside-outside [Stolcke&Omohundro, 1994] [Chen 1995] [Kurihara&Sato, 2004; 2006] [Liang et al., 2007] Main difference Different prior Structure search method

Iowa State University Department of Computer Science, Iowa State University Artificial Intelligence Research Laboratory Center for Computational Intelligence, Learning, and Discovery CCILD Talk presented at ICGI 2008, St Malo, France, September Kewei Tu and Vasant Honavar Related work Unsupervised parsing (not CFG) [Klein&Manning, 2002; 2004] U-DOP [Bod, 2006]