1 Inductive Learning of Rules MushroomEdible? SporesSpots Color YN BrownN YY GreyY NY BlackY NN BrownN YN WhiteN YY BrownY YN Brown NN Red Don’t try this.

Slides:



Advertisements
Similar presentations
Machine learning Overview
Advertisements

Board Games Draughts/Checkers Humans 0 – 1 Computers 1962 Arthur Samuels program beat state champion 1990 world champ beaten Completely solved in 2007.
1 Machine Learning: Lecture 1 Overview of Machine Learning (Based on Chapter 1 of Mitchell T.., Machine Learning, 1997)
Machine Learning: Intro and Supervised Classification
CS 4700: Foundations of Artificial Intelligence
Combining Inductive and Analytical Learning Ch 12. in Machine Learning Tom M. Mitchell 고려대학교 자연어처리 연구실 한 경 수
CS 484 – Artificial Intelligence1 Announcements Project 1 is due Tuesday, October 16 Send me the name of your konane bot Midterm is Thursday, October 18.
1er. Escuela Red ProTIC - Tandil, de Abril, Introduction How to program computers to learn? Learning: Improving automatically with experience.
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Spring 2004.
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Fall 2005.
Machine Learning II Decision Tree Induction CSE 473.
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Fall 2004.
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18.
Machine Learning CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanning.
1 Some rules  No make-up exams ! If you miss with an official excuse, you get average of your scores in the other exams – at most once.  WP only-if you.
1 Theory of Inductive Learning zSuppose our examples are drawn with a probability distribution Pr(x), and that we learned a hypothesis f to describe a.
Machine Learning: Symbol-Based
Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward.
A Brief Survey of Machine Learning
Part I: Classification and Bayesian Learning
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
1 What is learning? “Learning denotes changes in a system that... enable a system to do the same task more efficiently the next time.” –Herbert Simon “Learning.
For Friday Read chapter 18, sections 3-4 Homework: –Chapter 14, exercise 12 a, b, d.
CpSc 810: Machine Learning Design a learning system.
1 Machine Learning What is learning?. 2 Machine Learning What is learning? “That is what learning is. You suddenly understand something you've understood.
Machine Learning Chapter 11.
CS-424 Gregory Dudek Lecture 14 Learning –Probably approximately correct learning (cont’d) –Version spaces –Decision trees.
General-to-Specific Ordering. 8/29/03Logic Based Classification2 SkyAirTempHumidityWindWaterForecastEnjoySport SunnyWarmNormalStrongWarmSameYes SunnyWarmHighStrongWarmSameYes.
1 Mining in geographic data Original slides:Raymond J. Mooney University of Texas at Austin.
Lecture 10: 8/6/1435 Machine Learning Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Learning from Observations Chapter 18 Through
CHAPTER 18 SECTION 1 – 3 Learning from Observations.
Course Overview  What is AI?  What are the Major Challenges?  What are the Main Techniques?  Where are we failing, and why?  Step back and look at.
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 5.
Well Posed Learning Problems Must identify the following 3 features –Learning Task: the thing you want to learn. –Performance measure: must know when you.
1 Machine Learning 1.Where does machine learning fit in computer science? 2.What is machine learning? 3.Where can machine learning be applied? 4.Should.
Learning from observations
Machine Learning, Decision Trees, Overfitting Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 14,
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Prof. Carla P. Gomes Module: Intro Learning (Reading: Chapter.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 9 of 42 Wednesday, 14.
Chapter 1: Introduction. 2 목 차목 차 t Definition and Applications of Machine t Designing a Learning System  Choosing the Training Experience  Choosing.
Machine Learning Chapter 5. Evaluating Hypotheses
CS 5751 Machine Learning Chapter 3 Decision Tree Learning1 Decision Trees Decision tree representation ID3 learning algorithm Entropy, Information gain.
Machine Learning Introduction. Class Info Office Hours –Monday:11:30 – 1:00 –Wednesday:10:00 – 1:00 –Thursday:11:30 – 1:00 Course Text –Tom Mitchell:
Data Mining and Decision Support
1 Introduction to Machine Learning Chapter 1. cont.
CS 8751 ML & KDDComputational Learning Theory1 Notions of interest: efficiency, accuracy, complexity Probably, Approximately Correct (PAC) Learning Agnostic.
Introduction Machine Learning: Chapter 1. Contents Types of learning Applications of machine learning Disciplines related with machine learning Well-posed.
Well Posed Learning Problems Must identify the following 3 features –Learning Task: the thing you want to learn. –Performance measure: must know when you.
Outline Logistics Review Machine Learning –Induction of Decision Trees (7.2) –Version Spaces & Candidate Elimination –PAC Learning Theory (7.1) –Ensembles.
Chapter 18 Section 1 – 3 Learning from Observations.
Inductive Learning (2/2) Version Space and PAC Learning Russell and Norvig: Chapter 18, Sections 18.5 through 18.7 Chapter 18, Section 18.5 Chapter 19,
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
Learning From Observations Inductive Learning Decision Trees Ensembles.
Machine Learning & Datamining CSE 454. © Daniel S. Weld 2 Project Part 1 Feedback Serialization Java Supplied vs. Manual.
CS-424 Gregory Dudek Lecture 14 Learning –Inductive inference –Probably approximately correct learning.
1 Machine Learning Patricia J Riddle Computer Science 367 6/26/2016Machine Learning.
Supervise Learning Introduction. What is Learning Problem Learning = Improving with experience at some task – Improve over task T, – With respect to performance.
Data Mining Lecture 3.
Computational Learning Theory
Spring 2003 Dr. Susan Bridges
Chapter 11: Learning Introduction
Knowledge Representation
CSE P573 Applications of Artificial Intelligence Decision Trees
Computational Learning Theory
Why Machine Learning Flood of data
CS639: Data Management for Data Science
Lecture 14 Learning Inductive inference
Inductive Learning (2/2) Version Space and PAC Learning
Presentation transcript:

1 Inductive Learning of Rules MushroomEdible? SporesSpots Color YN BrownN YY GreyY NY BlackY NN BrownN YN WhiteN YY BrownY YN Brown NN Red Don’t try this at home...

2 Types of Learning zWhat is learning? yImproved performance over time/experience yIncreased knowledge zSpeedup learning yNo change to set of theoretically inferable facts yChange to speed with which agent can infer them zInductive learning yMore facts can be inferred

3 Mature Technology zMany Applications yDetect fraudulent credit card transactions yInformation filtering systems that learn user preferences yAutonomous vehicles that drive public highways (ALVINN) yDecision trees for diagnosing heart attacks ySpeech synthesis (correct pronunciation) (NETtalk) zData mining: huge datasets, scaling issues

4 Defining a Learning Problem zExperience: zTask: zPerformance Measure: A program is said to learn from experience E with respect to task T and performance measure P, if it’s performance at tasks in T, as measured by P, improves with experience E.

5 Example: Checkers zTask T: yPlaying checkers zPerformance Measure P: yPercent of games won against opponents zExperience E: yPlaying practice games against itself

6 Example: Handwriting Recognition zTask T: y zPerformance Measure P: y zExperience E: y Recognizing and classifying handwritten words within images

7 Example: Robot Driving zTask T: zPerformance Measure P: zExperience E: Driving on a public four-lane highway using vision sensors

8 Example: Speech Recognition zTask T: zPerformance Measure P: zExperience E: Identification of a word sequence from audio recorded from arbitrary speakers... noise

9 Issues zWhat feedback (experience) is available? zWhat kind of knowledge is being increased? zHow is that knowledge represented? zWhat prior information is available? zWhat is the right learning algorithm? zHow avoid overfitting?

10 Choosing the Training Experience zCredit assignment problem: yDirect training examples: xE.g. individual checker boards + correct move for each yIndirect training examples : xE.g. complete sequence of moves and final result zWhich examples: yRandom, teacher chooses, learner chooses Supervised learning Reinforcement learning Unsupervised learning

11 Choosing the Target Function zWhat type of knowledge will be learned? zHow will the knowledge be used by the performance program? zE.g. checkers program yAssume it knows legal moves yNeeds to choose best move ySo learn function: F: Boards -> Moves xhard to learn yAlternative: F: Boards -> R

12 The Ideal Evaluation Function zV(b) = 100 if b is a final, won board zV(b) = -100 if b is a final, lost board zV(b) = 0 if b is a final, drawn board zOtherwise, if b is not final V(b) = V(s) where s is best, reachable final board Nonoperational… Want operational approximation of V: V

13 How Represent Target Function zx 1 = number of black pieces on the board zx 2 = number of red pieces on the board zx 3 = number of black kings on the board zx 4 = number of red kings on the board zx 5 = number of black pieces threatened by red zx 6 = number of red pieces threatened by black V(b) = a + bx 1 + cx 2 + dx 3 + ex 4 + fx 5 + gx 6 Now just need to learn 7 numbers!

14 Target Function zProfound Formulation: Can express any type of inductive learning as approximating a function zE.g., Checkers yV: boards -> evaluation zE.g., Handwriting recognition yV: image -> word zE.g., Mushrooms yV: mushroom-attributes -> {E, P} zInductive bias

15 Theory of Inductive Learning

16 Theory of Inductive Learning zSuppose our examples are drawn with a probability distribution Pr(x), and that we learned a hypothesis f to describe a concept C. zWe can define Error(f) to be: zwhere D are the set of all examples on which f and C disagree.

17 PAC Learning zWe’re not perfect (in more than one way). So why should our programs be perfect? zWhat we want is:  Error(f) <  for some chosen  zBut sometimes, we’re completely clueless: (hopefully, with low probability). What we really want is:  Prob ( Error(f)  < .  As the number of examples grows,  and  should decrease. zWe call this Probably approximately correct.

18 Definition of PAC Learnability zLet C be a class of concepts. zWe say that C is PAC learnable by a hypothesis space H if: ythere is a polynomial-time algorithm A, ya polynomial function p,  such that for every C in C, every probability distribution Pr, and  and ,  if A is given at least p(1/ , 1/  ) examples,  then A returns with probability 1-  a hypothesis whose error is less than . zk-DNF, and k-CNF are PAC learnable.

19 Version Spaces: A Learning Alg. zKey idea: yMaintain most specific and most general hypotheses at every point. Update them as examples come in. zWe describe objects in the space by attributes: yfaculty, staff, student y20’s, 30’s, 40’s. ymale, female zConcepts: boolean combination of attribute- values: yfaculty, 30’s, male, yfemale, 20’s.

20 Generalization and Specializ... zA concept C1 is more general than C2 if it describes a superset of the objects: yC1={20’s, faculty} is more general than C2={20’s, faculty, female}. yC2 is a specialization of C1. zImmediate specializations (generalizations). zThe version space algorithm maintains the most specific and most general boundaries at every point of the learning.

21 Example T malefemale facultystudent 20’s30’s male, fac male,studfemale,facfemale,studfac,20’s fac, 30’s male,fac,20male,fac,30fem,fac,20male,stud,30