Biointelligence Laboratory, Seoul National University

Slides:



Advertisements
Similar presentations
Introduction: The Chomskian Perspective on Language Study.
Advertisements

Chapter 20: Natural Language Generation Presented by: Anastasia Gorbunova LING538: Computational Linguistics, Fall 2006 Speech and Language Processing.
The dynamics of iterated learning Tom Griffiths UC Berkeley with Mike Kalish, Steve Lewandowsky, Simon Kirby, and Mike Dowman.
A Bayesian view of language evolution by iterated learning Tom Griffiths Brown University Mike Kalish University of Louisiana.
1 Introduction to Computational Natural Language Learning Linguistics (Under: Topics in Natural Language Processing ) Computer Science (Under:
1 Markov Chains Algorithms in Computational Biology Spring 2006 Slides were edited by Itai Sharon from Dan Geiger and Ydo Wexler.
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Second Language Acquisition and Real World Applications Alessandro Benati (Director of CAROLE, University of Greenwich, UK) Making.
1 Modeling Parameter Setting Performance in Domains with a Large Number of Parameters: A Hybrid Approach CUNY / SUNY / NYU Linguistics Mini-conference.
Analyzing iterated learning Tom Griffiths Brown University Mike Kalish University of Louisiana.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
The effect of New Links on Google Pagerank By Hui Xie Apr, 07.
Evolution of Universal Grammar Pia Göser Universität Tübingen Seminar: Sprachevolution Dozent: Prof. Jäger
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
A Cognitive Substrate for Natural Language Understanding Nick Cassimatis Arthi Murugesan Magdalena Bugajska.
Kernel Classifiers from a Machine Learning Perspective (sec ) Jin-San Yang Biointelligence Laboratory School of Computer Science and Engineering.
THE BIG PICTURE Basic Assumptions Linguistics is the empirical science that studies language (or linguistic behavior) Linguistics proposes theories (models)
Some Probability Theory and Computational models A short overview.
TKK | Automation Technology Laboratory Partially Observable Markov Decision Process (Chapter 15 & 16) José Luis Peralta.
Learning Automata and Grammars Peter Černo.  The problem of learning or inferring automata and grammars has been studied for decades and has connections.
Formal Models in AGI Research Pei Wang Temple University Philadelphia, USA.
LOGIC AND ONTOLOGY Both logic and ontology are important areas of philosophy covering large, diverse, and active research projects. These two areas overlap.
Copyright © 2012, SAS Institute Inc. All rights reserved. ANALYTICS IN BIG DATA ERA ANALYTICS TECHNOLOGY AND ARCHITECTURE TO MANAGE VELOCITY AND VARIETY,
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Programming Languages and Design Lecture 3 Semantic Specifications of Programming Languages Instructor: Li Ma Department of Computer Science Texas Southern.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Chapter 1. Introduction From “The Computational Nature of Language Learning and Evolution” Summarized by Seok Ho-Sik.
Probabilistic Automaton Ashish Srivastava Harshil Pathak.
Chap2. Language Acquisition: The Problem of Inductive Inference (2.1 ~ 2.2) Min Su Lee The Computational Nature of Language Learning and Evolution.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
The Computational Nature of Language Learning and Evolution 10. Variations and Case Studies Summarized by In-Hee Lee
Chapter 3 Language Acquisition: A Linguistic Treatment Jang, HaYoung Biointelligence Laborotary Seoul National University.
Ch 4. Language Acquisition: Memoryless Learning 4.1 ~ 4.3 The Computational Nature of Language Learning and Evolution Partha Niyogi 2004 Summarized by.
Ch 2. The Probably Approximately Correct Model and the VC Theorem 2.3 The Computational Nature of Language Learning and Evolution, Partha Niyogi, 2004.
Chapter 9. The PlayMate System ( 2/2 ) in Cognitive Systems Monographs. Rüdiger Dillmann et al. Course: Robots Learning from Humans Summarized by Nan Changjun.
Research Design
Ch 5. Language Change: A Preliminary Model 5.1 ~ 5.2 The Computational Nature of Language Learning and Evolution P. Niyogi 2006 Summarized by Kwonill,
Chapter 9. A Model of Cultural Evolution and Its Application to Language From “The Computational Nature of Language Learning and Evolution” Summarized.
The Monte Carlo Method/ Markov Chains/ Metropolitan Algorithm from sec in “Adaptive Cooperative Systems” -summarized by Jinsan Yang.
Chapter 14. Conclusions From “The Computational Nature of Language Learning and Evolution” Summarized by Seok Ho-Sik.
Second Language Acquisition
An Introduction to Linguistics
IB Assessments CRITERION!!!.
Summarized by In-Hee Lee
C.-S. Shieh, EC, KUAS, Taiwan
Artificial Intelligence Chapter 17 Knowledge-Based Systems
Artificial Intelligence Chapter 17 Knowledge-Based Systems
Intelligent Agents Chapter 2.
Course: Autonomous Machine Learning
What is Linguistics? The scientific study of human language
THE NATURE of LEARNER LANGUAGE
Markov chain monte carlo
SNU BioIntelligence Lab.
The Nature of Learner Language (Chapter 2 Rod Ellis, 1997) Page 15
Artificial Intelligence Chapter 20 Learning and Acting with Bayes Nets
Biointelligence Laboratory, Seoul National University
Supplement Beyond Computation
Markov Decision Problems
Language Model Approach to IR
Biointelligence Laboratory, Seoul National University
Chapter 20. Learning and Acting with Bayes Nets
CPSC 503 Computational Linguistics
Ch 6. Language Change: Multiple Languages 6.1 Multiple Languages
Chapter 5 Language Change: A Preliminary Model (2/2)
Artificial Intelligence Chapter 17 Knowledge-Based Systems
The Nature Of Learner Language
Chapter 3 Interlanguage.
Biointelligence Laboratory, Seoul National University
Sets and Probabilistic Models
Presented by Nick Janus
Presentation transcript:

Biointelligence Laboratory, Seoul National University Ch 4. Memoryless Learning 4.4 Generalization and Variation 4.5 Other Kinds of Learning Algorithms 4.6 Conclusions Summarized by J.-W. Ha Biointelligence Laboratory, Seoul National University http://bi.snu.ac.kr/

(C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/ Contents 4.4 Generalizations and Variations 4.4.1 Markov Chains and Learning Algorithms 4.4.2 Memoryless Learners 4.4.3 The Power of Memoryless Learners 4.5 Other Kinds of Learning Algorithms 4.6 Conclusions (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Markov Chains and Learning Algorithms A learning algorithm A specified by a mapping from where D is the set of all finite-length data streams and H is a class of hypothesis grammars The probability with which the following event happens: Event: where T is the set of all texts and t is a generic element of T, tn refers to the first n sentences in the text t, tn ∈ D provided that means that after n examples have been received, the probability with which the learner will conjecture g at this stage is exactly 0 In mathematics, the Kolmogorov extension theorem (also known as Kolmogorov existence theorem) is a theorem that guarantees that a suitably "consistent" collection of finite-dimensional distributions will define a stochastic process (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Markov Chains and Learning Algorithms Inhomogeneous Markov Chains Learning algorithms are possible to be represented with inhomogeneous Markov chains The state space: the set of possible grammars in H At each point in time, the learner has a grammatical hypothesis and the chain is in the state corresponding to that grammar. At the n-th time step, the transition matrix of chain: A set of positive numbers αg s.t. Tn is time dependent and characterizes the evolution of the learner’s hypothesis from example sentence to example sentence (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Markov Chains and Learning Algorithms Flows of Making a Markov Chain (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Markov Chains and Learning Algorithms Theorem 4.4 Learnability of grammars is related to the convergence of non-stationary Markov chains. In general, if the chain is ergodic, such an inhomogeneous chain converges to its limiting distribution. If the learning algorithm A is such that (1) the associated Markov chain is ergodic, and (2) for each n, Tn is such that all closed sets contain the target state, then the hypothesis generated by A will converge to the target state from all initial distributions. (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

(C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/ Memoryless Learning Memoryless algorithms Memoryless algorithms may be regarded as those that have no recollection of previous data, or previous inferences made about the target function. At any point in time, the only information upon which such an algorithm acts is the current data and the current hypothesis (state) In general, given a particular hypothesis state (h in H), and a new datum (sentence, s in Σ*), a memoryless algorithm maps onto a new hypothesis (g ∈ H) Formally Tn is independent of n and Markov Chain is a stationary one (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

The Power of Memoryless Learners Pure memoryless learners belong to the more general class of memory limited learners which develop grammatical hypothesis based on a finite memory of sentences they have heard over their lifetime. k-memory limited algorithms A(u) depends only upon the previous grammatical hypothesis (A(u-)) and the last k sentences heard (u-k) (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

The Power of Memoryless Learners Memoryless learners vs. Memory limited learners One might think that the class of language learnable by memory limited learners is larger than that learnable by memoryless. However, this turns out not to be the case. Theorem 4.5: If a class of grammars H is learnable in the limit by a k-memory limited learning algorithm Ak, then there exists some memoryless learning algorithm that is also able to learn it. Theorem 4.6: If a class of grammars H is learnable in the limit (in the classical Gold sense), the it is learnable with measure 1 by some memoryless learner. Proof : Appendix (Ch. 4.7) Consequently, the characterization of memoryless learners by first order Markov chains takes on a general significance that far exceeds the original context of the TLA. (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Other Kinds of Learning Algorithms Optimality Theory In optimality theory (Prince and Smolensky 1993), grammatical variation is treated via constraints rather than rules. In this theory, one begins with a finite number of constraints C1, C2,…, Cn. These constraints are ordered in importance and determine the ranking and relative importance of constraint violations. Various Framework in Linguistic Tradition Briscoe: LFG (2000), Yang: GB (2000), Stabler: a minimalist (1998) Fodor(1994,1998), Sakas (2000), Bertolo (2001), Clark and Roberts (1993): P&P Neumann and Flickinger (1999): HPSG (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Other Kinds of Learning Algorithms Semantic Consideration To clarify the manner in which semantic considerations enter the process of acquiring a language. Learning procedures rely on semantic feedback to acquire formal structures in a more functionalist perspective on language acquisition (Feldman et al. 1996, Regier 1996, Siskind 1992) Encode Prior Knowledge Other probabilistic accounts of language acquisition attempt to encode prior knowledge of language to varying degrees of specificity in a MDL framework (Brent 1999, Goldsmith 2001, deMarcken 1996) Connectionist and other data-driven approaches (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

(C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/ Conclusions The investigation of language acquisition within the P&P framework with central attention to the kinds of models inspired by the TLA Need not only to check for learnability in the limit but also to quantify the sample complexity of the learning problem Characterizing convergence time with the Markov analysis. The convergence rates depend upon the probability distribution according to which example sentences were presented to the learner. Any learning algorithm on any enumerable class of grammars may be characterized as an inhomogeneous Markov chain. Any memory-limited learning algorithm is ultimately a first order Markov chain The process of language acquisition of children is one of the deepest scientific problems. However, computational modeling has played an important role in helping reason through various explanatory possibilities for how such a formal system may be acquired since a natural language with its phonetic distinction, morphological patterns, and syntactic forms has a certain kind of formal structure. (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Proofs for Memoryless Algorithms Two facts The first fact (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Proofs for Memoryless Algorithms The second fact (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Proofs for Memoryless Algorithms (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Proofs for Memoryless Algorithms (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Proofs for Memoryless Algorithms The Updates Flow of Learners (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Proofs for Memoryless Algorithms (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/

Proofs for Memoryless Algorithms (C) 2009, SNU Biointelligence Lab, http://bi.snu.ac.kr/