Concept Learning and the General-to-Specific Ordering 이 종우 자연언어처리연구실.

Slides:



Advertisements
Similar presentations
Computational Learning Theory
Advertisements

Concept Learning and the General-to-Specific Ordering
2. Concept Learning 2.1 Introduction
Concept Learning DefinitionsDefinitions Search Space and General-Specific OrderingSearch Space and General-Specific Ordering The Candidate Elimination.
Combining Inductive and Analytical Learning Ch 12. in Machine Learning Tom M. Mitchell 고려대학교 자연어처리 연구실 한 경 수
CS 484 – Artificial Intelligence1 Announcements Project 1 is due Tuesday, October 16 Send me the name of your konane bot Midterm is Thursday, October 18.
Università di Milano-Bicocca Laurea Magistrale in Informatica
Decision Trees. DEFINE: Set X of Instances (of n-tuples x = ) –E.g., days decribed by attributes (or features): Sky, Temp, Humidity, Wind, Water, Forecast.
Adapted by Doug Downey from: Bryan Pardo, EECS 349 Fall 2007 Machine Learning Lecture 2: Concept Learning and Version Spaces 1.
Chapter 2 - Concept learning
Machine Learning: Symbol-Based
MACHINE LEARNING. What is learning? A computer program learns if it improves its performance at some task through experience (T. Mitchell, 1997) A computer.
Concept Learning and Version Spaces
Artificial Intelligence 6. Machine Learning, Version Space Method
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, January 19, 2001.
Computing & Information Sciences Kansas State University Lecture 01 of 42 Wednesday, 24 January 2008 William H. Hsu Department of Computing and Information.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Friday, 25 January 2008 William.
Machine Learning Chapter 3. Decision Tree Learning
CS 484 – Artificial Intelligence1 Announcements List of 5 source for research paper Homework 5 due Tuesday, October 30 Book Review due Tuesday, October.
CS 478 – Tools for Machine Learning and Data Mining The Need for and Role of Bias.
For Monday Read chapter 18, sections 5-6 Homework: –Chapter 18, exercises 1-2.
For Friday Read chapter 18, sections 3-4 Homework: –Chapter 14, exercise 12 a, b, d.
1 Machine Learning What is learning?. 2 Machine Learning What is learning? “That is what learning is. You suddenly understand something you've understood.
Machine Learning Chapter 11.
For Wednesday Read chapter 18, section 7 Homework: –Chapter 18, exercise 11.
CpSc 810: Machine Learning Decision Tree Learning.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, January 19, 2000.
General-to-Specific Ordering. 8/29/03Logic Based Classification2 SkyAirTempHumidityWindWaterForecastEnjoySport SunnyWarmNormalStrongWarmSameYes SunnyWarmHighStrongWarmSameYes.
Machine Learning Chapter 2. Concept Learning and The General-to-specific Ordering Gun Ho Lee Soongsil University, Seoul.
1 Concept Learning By Dong Xu State Key Lab of CAD&CG, ZJU.
机器学习 陈昱 北京大学计算机科学技术研究所 信息安全工程研究中心. 课程基本信息  主讲教师:陈昱 Tel :  助教:程再兴, Tel :  课程网页:
Machine Learning Chapter 2. Concept Learning and The General-to-specific Ordering Tom M. Mitchell.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, January 22, 2001 William.
Chapter 2: Concept Learning and the General-to-Specific Ordering.
CpSc 810: Machine Learning Concept Learning and General to Specific Ordering.
Outline Inductive bias General-to specific ordering of hypotheses
Overview Concept Learning Representation Inductive Learning Hypothesis
Computational Learning Theory IntroductionIntroduction The PAC Learning FrameworkThe PAC Learning Framework Finite Hypothesis SpacesFinite Hypothesis Spaces.
1 Universidad de Buenos Aires Maestría en Data Mining y Knowledge Discovery Aprendizaje Automático 2-Concept Learning (1/3) Eduardo Poggi
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Thursday, August 26, 1999 William.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 34 of 41 Wednesday, 10.
Machine Learning: Lecture 2
Machine Learning Concept Learning General-to Specific Ordering
1 Decision Trees ExampleSkyAirTempHumidityWindWaterForecastEnjoySport 1SunnyWarmNormalStrongWarmSameYes 2SunnyWarmHighStrongWarmSameYes 3RainyColdHighStrongWarmChangeNo.
Kansas State University Department of Computing and Information Sciences CIS 690: Implementation of High-Performance Data Mining Systems Thursday, 20 May.
CS 8751 ML & KDDComputational Learning Theory1 Notions of interest: efficiency, accuracy, complexity Probably, Approximately Correct (PAC) Learning Agnostic.
CS464 Introduction to Machine Learning1 Concept Learning Inducing general functions from specific training examples is a main issue of machine learning.
Concept Learning and The General-To Specific Ordering
Computational Learning Theory Part 1: Preliminaries 1.
Concept learning Maria Simi, 2011/2012 Machine Learning, Tom Mitchell Mc Graw-Hill International Editions, 1997 (Cap 1, 2).
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
Chapter 2 Concept Learning
Computational Learning Theory
Concept Learning Machine Learning by T. Mitchell (McGraw-Hill) Chp. 2
CSE543: Machine Learning Lecture 2: August 6, 2014
CS 9633 Machine Learning Concept Learning
Analytical Learning Discussion (4 of 4):
Machine Learning Chapter 2
Introduction to Machine Learning Algorithms in Bioinformatics: Part II
Ordering of Hypothesis Space
Concept Learning.
IES 511 Machine Learning Dr. Türker İnce (Lecture notes by Prof. T. M
Concept Learning Berlin Chen 2005 References:
Machine Learning Chapter 2
Implementation of Learning Systems
Version Space Machine Learning Fall 2018.
Machine Learning Chapter 2
Presentation transcript:

Concept Learning and the General-to-Specific Ordering 이 종우 자연언어처리연구실

Concept Learning Concepts or Categories –“birds” –“car” –“situations in which I should study more in order to pass the exam” –Concept some subset of objects or events defined over a larger set, or a boolean valued function defined over this larger set.

–Learning inducing general functions from specific training examples –Concept Learning acquiring the definition of a general category given a sample of positive and negative training examples of the category

A Concept Learning Task Target Concept –“days on which Aldo enjoys water sport” Hypothesis –vector of 6 constraints (Sky, AirTemp, Humidity, Wind, Water, Forecast, EnjoySport ) –Each attribute (“?”, single value or “0”) –e.g.

Instance Sky AirTemp Humidity Wind Water Forecast EnjoySport A Sunny Warm Normal Strong Warm Same No B Sumny Warm High Strong Warm Same Yes C Rainy Cold High Strong Warm Change No D Sunny Warm High Strong Cool Change Yes Training examples for the target concept EnjoySport

Given : –instances (X): set of iterms over which the concept is defined. –target concept (c) : c : X → {0, 1} –training examples (positive/negative) : –training set D: available training examples –set of all possible hypotheses: H Determine : – to find h(x) = c(x) (for all x in X)

Inductive Learning Hypothesis –Any good hypothesis over a sufficiently large set of training examples will also approximate the target function. well over unseen examples.

Concept Learning as Search Issue of Search –to find training examples hypothesis that best fits training examples Kinds of Space in EnjoySport –3*2*2*2*2 = 96: instant space –5*4*4*4*4 = 5120: syntactically distinct hypotheses within H –1+4*3*3*3*3 = 973: semantically distinct hypotheses

Search Problem –efficient search in hypothesis space(finite/infinite)

General-to-Specific Ordering of Hypotheses Hypotheses 의 General-to-Specific Ordering –x satisfies h ⇔ h(x)=1 –more_general_than_or_equal_to relations ≦ g –more_general_than_or_equal_to relations

–partial order (reflexive,antisymmetric,transitive) Concept Learning as Search

Find-S: Finding a Maximally Specific Hypothesis algorithm 1. Initialize h to the most specific hypothesis in H 2. For each positive training example x For each attribute constraint a i in h – If the constraint a i is satisfied by x – then do nothing – else replace a i in h by the next more general constraint that is satisfied by x 3. Output hypothesis h Property guaranteed to output the most specific hypothesis no way to determine unique hypothesis not cope with inconsistent errors or noises

Find-S:Finding a Maximally Specific Hypothesis(2)

Version Spaces and the Candidate- Elimination Algorithm –output all hypotheses consistent with the training examples. –perform poorly with noisy training data. Representation –Consistent(h,D) ⇔ ( ∀  D) h(x) = c(x) –VS H,D ⇔ {h  H | Consistent(h,D)} List-Then-Eliminate Algorithm –lists all hypotheses -> remove inconsistent ones. –Appliable to finite H

Version Spaces and the Candidate- Elimination Algorithm(2) More Compact Representation for Version Spaces –general boundary G –specific boundary S –Version Space redefined with S and G

Version Spaces and the Candidate- Elimination Algorithm(3)

Version Spaces and the Candidate- Elimination Algorithm(4) Condidate-Elimination Learning Algorithm Initialize G to the set of maximally general hypotheses in H Initialize S to the set of maximally specific hypotheses in H For each training example d, do If d is a positive example Remove from G any hypothesis inconsistent with d For each hypothesis s in S that is not consistent with d Remove s from S Add to S all minimal generalizations h of s such that h is consistent with d, and some member of G is more general than h Remove from S any hypothesis that is more general than another hypothesis in S

Version Spaces and the Candidate- Elimination Algorithm(5) If d is a negative example Remove from S any hypothesis inconsistent with d For each hypothesis g in G that is not consistent with d Remove g from G Add to G all minimal specializations h of g such that h is consistent with d, and some member of S is more specific than h Remove from G any hypothesis that is less general than another hypothesis in G

Version Spaces and the Candidate- Elimination Algorithm(6) Illustrative Example

Version Spaces and the Candidate- Elimination Algorithm(7)

Version Spaces and the Candidate- Elimination Algorithm(8)

Version Spaces and the Candidate- Elimination Algorithm(9)

Remarks on Version Spaces and Candidate-Elimination Will the Candidate-Elimination Algorithm Converge to the Correct Hypothesis? –Prerequisite – 1. No error in training examples – 2. Hypothesis exists which correctly describes c(x). –S and G boundary sets converge to an empty set => no hypothesis in H consistent with observed examples. What Training Example Should the Learner Request Next? –Negative one specifies G, positive one generalizes S. –optimal query satisfy half the hypotheses.

Remarks on Version Spaces and Candidate-Elimination(2) How Can Partially Learned Concepts Be Used? Instance Sky AirTemp Humidity Wind Water Forecast EnjoySport A Sunny Warm Normal Strong Cool Change ? B Rainy Cold Normal Light Warm Same ? C Sunny Warm Normal Light Warm Same ? D Sunny Cold Normal Strong Warm Same ? A : classified to positive B : classified to negative C : 3 positive, 3 negative D : 2 positive, 4 negative

Inductive Bias A Biased Hypothesis Space Example Sky AirTemp Humidity Wind Water Forecast EnjoySport 1 Sunny Warm Normal Strong Cool Change Yes 2 Cloudy Warm Normal Strong Cool Change Yes 3 Rainy Warm Normal Strong Cool Change No - zero hypothesis in the version space - caused by only conjunctive hypothesis

Inductive Bias(2) An Unbiased Learner –Power set of X : set of all subsets of a set X number of size of power set : 2 |X| –e.g. ∨ –new problem : unable to generalize beyond the observed examples. Observed examples are only unambiguously classified. Voting results in no majority or minority.

Inductive Bias(3) The Futility of Bias-Free Learning –no inductive bias => cannot classify unseen data reasonably –inductive bias of L : any minimal set of assertions B such that –inductive bias of Candidate-Elimination algorithm c ∈ H –advantage of introducing inductive bias generalizing beyond the observed data allows comparison of different learners

Inductive Bias(4) e.g –Rote-learner : no inductive bias –Candidate-Elimination algo : c ∈ H => more strong –Find-S : c ∈ H and that all are negative unless not proved positive

Inductive Bias(5)

Summary Concept learning can be cast as a problem of searching through a large predefined space of potential hypotheses. General-to-specific partial ordering of hypotheses provides a useful structure for search. Find-S algorithm performs specific-to-general search to find the most specific hypothesis. Candidate-Elimination algorithm computes version space by incrementally computing the sets of maximally specific (S) and maximally general (G) hypotheses. S and G delimit the entire set of hypotheses consistent with the data.

Version spaces and Candidate-Elimination algorithm provide a useful conceptual framework for studying concept learning. Candidate-Elimination algorithm not robust to noisy data or to situations where the unknown target concept is not expressible in the provided hypothesis space. Inductive bias in Candidate-Elimination algorithm is that target concept exists in H If hypothesis space be enriched so that there is a every possible hypothesis, that would remove the ability to classify any instance beyond the observed examples.