Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.

Slides:



Advertisements
Similar presentations
Computational Learning Theory
Advertisements

Concept Learning and the General-to-Specific Ordering
2. Concept Learning 2.1 Introduction
On Complexity, Sampling, and -Nets and -Samples. Range Spaces A range space is a pair, where is a ground set, it’s elements called points and is a family.
PAC Learning adapted from Tom M.Mitchell Carnegie Mellon University.
Università di Milano-Bicocca Laurea Magistrale in Informatica
Feasibility of learning: the issues solution for infinite hypothesis sets VC generalization bound (mostly lecture 5 on AMLbook.com)
Computational Learning Theory
Northwestern University Winter 2007 Machine Learning EECS Machine Learning Lecture 13: Computational Learning Theory.
Measuring Model Complexity (Textbook, Sections ) CS 410/510 Thurs. April 27, 2007 Given two hypotheses (models) that correctly classify the training.
Adapted by Doug Downey from: Bryan Pardo, EECS 349 Fall 2007 Machine Learning Lecture 2: Concept Learning and Version Spaces 1.
Probably Approximately Correct Model (PAC)
Introduction to Machine Learning course fall 2007 Lecturer: Amnon Shashua Teaching Assistant: Yevgeny Seldin School of Computer Science and Engineering.
Introduction to Boosting Aristotelis Tsirigos SCLT seminar - NYU Computer Science.
1 Computational Learning Theory and Kernel Methods Tianyi Jiang March 8, 2004.
CS 4700: Foundations of Artificial Intelligence
PAC learning Invented by L.Valiant in 1984 L.G.ValiantA theory of the learnable, Communications of the ACM, 1984, vol 27, 11, pp
1 Machine Learning: Lecture 5 Experimental Evaluation of Learning Algorithms (Based on Chapter 5 of Mitchell T.., Machine Learning, 1997)
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, January 19, 2001.
Computing & Information Sciences Kansas State University Lecture 01 of 42 Wednesday, 24 January 2008 William H. Hsu Department of Computing and Information.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Friday, 25 January 2008 William.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Friday, 16 March 2007 William.
1 Machine Learning What is learning?. 2 Machine Learning What is learning? “That is what learning is. You suddenly understand something you've understood.
Machine Learning Chapter 11.
1 CS 391L: Machine Learning: Computational Learning Theory Raymond J. Mooney University of Texas at Austin.
1 CS 391L: Machine Learning: Computational Learning Theory Raymond J. Mooney University of Texas at Austin.
1 Concept Learning By Dong Xu State Key Lab of CAD&CG, ZJU.
Machine Learning Chapter 2. Concept Learning and The General-to-specific Ordering Tom M. Mitchell.
Chapter 2: Concept Learning and the General-to-Specific Ordering.
1 Machine Learning: Lecture 8 Computational Learning Theory (Based on Chapter 7 of Mitchell T.., Machine Learning, 1997)
CpSc 810: Machine Learning Concept Learning and General to Specific Ordering.
Concept Learning and the General-to-Specific Ordering 이 종우 자연언어처리연구실.
Outline Inductive bias General-to specific ordering of hypotheses
Overview Concept Learning Representation Inductive Learning Hypothesis
Computational Learning Theory IntroductionIntroduction The PAC Learning FrameworkThe PAC Learning Framework Finite Hypothesis SpacesFinite Hypothesis Spaces.
Machine Learning Chapter 5. Evaluating Hypotheses
1 CSI5388 Current Approaches to Evaluation (Based on Chapter 5 of Mitchell T.., Machine Learning, 1997)
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Thursday, August 26, 1999 William.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 34 of 41 Wednesday, 10.
Machine Learning Concept Learning General-to Specific Ordering
Goal of Learning Algorithms  The early learning algorithms were designed to find such an accurate fit to the data.  A classifier is said to be consistent.
Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University Today: Computational Learning Theory Probably Approximately.
Kansas State University Department of Computing and Information Sciences CIS 690: Implementation of High-Performance Data Mining Systems Thursday, 20 May.
CS 8751 ML & KDDComputational Learning Theory1 Notions of interest: efficiency, accuracy, complexity Probably, Approximately Correct (PAC) Learning Agnostic.
MACHINE LEARNING 3. Supervised Learning. Learning a Class from Examples Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Carla P. Gomes CS4700 Computational Learning Theory Slides by Carla P. Gomes and Nathalie Japkowicz (Reading: R&N AIMA 3 rd ed., Chapter 18.5)
Concept Learning and The General-To Specific Ordering
Computational Learning Theory Part 1: Preliminaries 1.
Concept learning Maria Simi, 2011/2012 Machine Learning, Tom Mitchell Mc Graw-Hill International Editions, 1997 (Cap 1, 2).
1 CS 391L: Machine Learning: Computational Learning Theory Raymond J. Mooney University of Texas at Austin.
Computational Learning Theory
Concept Learning Machine Learning by T. Mitchell (McGraw-Hill) Chp. 2
CSE543: Machine Learning Lecture 2: August 6, 2014
CS 9633 Machine Learning Concept Learning
Computational Learning Theory
Computational Learning Theory
Analytical Learning Discussion (4 of 4):
Vapnik–Chervonenkis Dimension
Computational Learning Theory
Computational Learning Theory
Computational Learning Theory Eric Xing Lecture 5, August 13, 2010
CSCI B609: “Foundations of Data Science”
Computational Learning Theory
Machine Learning: UNIT-3 CHAPTER-2
Concept Learning Berlin Chen 2005 References:
Machine Learning Chapter 2
Version Space Machine Learning Fall 2018.
Machine Learning Chapter 2
Presentation transcript:

Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell

2 Computational Learning Theory (1/2)  Computational learning theory  Setting 1: learner poses queries to teacher  Setting 2: teacher chooses examples  Setting 3: randomly generated instances, labeled by teacher  Probably approximately correct (PAC) learning  Vapnik-Chervonenkis Dimension  Mistake bounds

3 Computational Learning Theory (2/2)  What general laws constrain inductive learning?  We seek theory to relate: –Probability of successful learning –Number of training examples –Complexity of hypothesis space –Accuracy to which target concept is approximated –Manner in which training examples presented

4 Prototypical Concept Learning Task  Given: –Instances X: Possible days, each described by the attributes Sky, AirTemp, Humidity, Wind, Water, Forecast –Target function c: EnjoySport : X  {0, 1} –Hypotheses H: Conjunctions of literals. E.g.. –Training examples D: Positive and negative examples of the target function,…  Determine: –A hypothesis h in H such that h(x) = c(x) for all x in D? –A hypothesis h in H such that h(x) = c(x) for all x in X?

5 Sample Complexity  How many training examples are sufficient to learn the target concept? 1. If learner proposes instances, as queries to teacher  Learner proposes instance x, teacher provides c(x) 2. If teacher (who knows c) provides training examples  teacher provides sequence of examples of form 3. If some random process (e.g., nature) proposes instances  instance x generated randomly, teacher provides c(x)

6 Sample Complexity: 1  Learner proposes instance x, teacher provides c(x) (assume c is in learner’s hypothesis space H)  Optimal query strategy: play 20 questions –pick instance x such that half of hypotheses in V S classify x positive, half classify x negative –When this is possible, need  log 2 |H|  queries to learn c –when not possible, need even more

7 Sample Complexity: 2  Teacher (who knows c) provides training examples (assume c is in learner’s hypothesis space H)  Optimal teaching strategy: depends on H used by learner  Consider the case H = conjunctions of up to n boolean literals and their negations –e.g., (AirTemp = Warm)  (Wind = Strong), where AirTemp, Wind, …, each have 2 possible values. –if n possible boolean attributes in H, n + 1 examples suffice –why?

8 Sample Complexity: 3  Given: –set of instances X  set of hypotheses H –set of possible target concepts C –training instances generated by a fixed, unknown probability distribution D over X  Learner observes a sequence D of training examples of form, for some target concept c  C –instances x are drawn from distribution D –teacher provides target value c(x) for each  Learner must output a hypothesis h estimating c –h is evaluated by its performance on subsequent instances drawn according to D  Note: probabilistic instances, noise-free classifications

9 True Error of a Hypothesis  Definition: The true error (denoted error D (h)) of hypothesis h with respect to target concept c and distribution D is the probability that h will misclassify an instance drawn at random according to D.

10 Two Notions of Error  Training error of hypothesis h with respect to target concept c –How often h(x)  c(x) over training instances  True error of hypothesis h with respect to c –How often h(x)  c(x) over future random instances  Our concern: –Can we bound the true error of h given the training error of h? –First consider when training error of h is zero (i.e., h  V S H,D )

11 Exhausting the Version Space  Definition: The version space V S H,D is said to be  - exhausted with respect to c and D, if every hypothesis h in V S H,D has error less than  with respect to c and D. (  h  V S H,D ) error D (h) < 

12 How many examples will  -exhaust the VS?  Theorem: [Haussler, 1988]. –If the hypothesis space H is finite, and D is a sequence of m  1 independent random examples of some target concept c, then for any 0    1, the probability that the version space with respect to H and D is not  -exhausted (with respect to c) is less than |H|e -  m  Interesting! This bounds the probability that any consistent learner will output a hypothesis h with error(h)    If we want to this probability to be below  |H|e -  m   then

13 Learning Conjunctions of Boolean Literals  How many examples are sufficient to assure with probability at least (1 -  ) that every h in V S H,D satisfies error D (h)    Use our theorem:  Suppose H contains conjunctions of constraints on up to n boolean attributes (i.e., n boolean literals). Then |H| = 3 n, and or

14 How About EnjoySport?  If H is as given in EnjoySport then |H| = 973, and ... if want to assure that with probability 95%, V S contains only hypotheses with error D (h) .1, then it is sufficient to have m examples, where

15 PAC Learning  Consider a class C of possible target concepts defined over a set of instances X of length n, and a learner L using hypothesis space H. –Definition: C is PAC-learnable by L using H if for all c  C, distributions D over X,  such that 0 <  < 1/2, and  such that 0 <  < 1/2 –learner L will with probability at least (1 -  ) output a hypothesis h  H such that error D (h)  , in time that is polynomial in 1/ , 1/ , n, and size(c).

16 Agnostic Learning  So far, assumed c  H Agnostic learning setting: don’t assume c  H –What do we want then?  The hypothesis h that makes fewest errors on training data –What is sample complexity in this case?  derived from Hoeffding bounds:

17 Shattering a Set of Instances  Definition: a dichotomy of a set S is a partition of S into two disjoint subsets.  Definition: a set of instances S is shattered by hypothesis space H if and only if for every dichotomy of S there exists some hypothesis in H consistent with this dichotomy.

18 Three Instances Shattered Instance space X

19 The Vapnik-Chervonenkis Dimension  Definition: The Vapnik-Chervonenkis dimension, V C(H), of hypothesis space H defined over instance space X is the size of the largest finite subset of X shattered by H. If arbitrarily large finite sets of X can be shattered by H, then V C(H)  .

20 VC Dim. of Linear Decision Surfaces (a)(b)

21 Sample Complexity from VC Dimension  How many randomly drawn examples suffice to  -exhaust V S H,D with probability at least (1 -  ) ?

22 Mistake Bounds  So far: how many examples needed to learn?  What about: how many mistakes before convergence?  Let’s consider similar setting to PAC learning: –Instances drawn at random from X according to distribution D –Learner must classify each instance before receiving correct classification from teacher –Can we bound the number of mistakes learner makes before converging?

23 Mistake Bounds: Find-S  Consider Find-S when H = conjunction of boolean literals Find-S: –Initialize h to the most specific hypothesis l 1   l 1  l 2   l 2 … l n   l n –For each positive training instance x  Remove from h any literal that is not satisfied by x –Output hypothesis h.  How many mistakes before converging to correct h?

24 Mistake Bounds: Halving Algorithm  Consider the Halving Algorithm: –Learn concept using version space Candidate-Elimination algorithm –Classify new instances by majority vote of version space members  How many mistakes before converging to correct h? –... in worst case? –... in best case?

25 Optimal Mistake Bounds  Let M A (C) be the max number of mistakes made by algorithm A to learn concepts in C. (maximum over all possible c  C, and all possible training sequences)  Definition: Let C be an arbitrary non-empty concept class. The optimal mistake bound for C, denoted Opt(C), is the minimum over all possible learning algorithms A of M A (C).