Artificial Intelligence 6. Machine Learning, Version Space Method

Slides:



Advertisements
Similar presentations
Concept Learning and the General-to-Specific Ordering
Advertisements

2. Concept Learning 2.1 Introduction
Introduction to Machine Learning BITS C464/BITS F464
Godfather to the Singularity
CS 484 – Artificial Intelligence1 Announcements Project 1 is due Tuesday, October 16 Send me the name of your konane bot Midterm is Thursday, October 18.
Università di Milano-Bicocca Laurea Magistrale in Informatica
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Adapted by Doug Downey from: Bryan Pardo, EECS 349 Fall 2007 Machine Learning Lecture 2: Concept Learning and Version Spaces 1.
Machine Learning: Symbol-Based
Learning Programs Danielle and Joseph Bennett (and Lorelei) 4 December 2007.
MACHINE LEARNING. What is learning? A computer program learns if it improves its performance at some task through experience (T. Mitchell, 1997) A computer.
Concept Learning and Version Spaces
Part I: Classification and Bayesian Learning
INTRODUCTION TO Machine Learning 3rd Edition
Introduction to machine learning
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, January 19, 2001.
Computing & Information Sciences Kansas State University Lecture 01 of 42 Wednesday, 24 January 2008 William H. Hsu Department of Computing and Information.
Short Introduction to Machine Learning Instructor: Rada Mihalcea.
CS 484 – Artificial Intelligence1 Announcements List of 5 source for research paper Homework 5 due Tuesday, October 30 Book Review due Tuesday, October.
For Friday Read chapter 18, sections 3-4 Homework: –Chapter 14, exercise 12 a, b, d.
1 Machine Learning What is learning?. 2 Machine Learning What is learning? “That is what learning is. You suddenly understand something you've understood.
Machine Learning Chapter 11.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Machine Learning Introduction Study on the Coursera All Right Reserved : Andrew Ng Lecturer:Much Database Lab of Xiamen University Aug 12,2014.
General-to-Specific Ordering. 8/29/03Logic Based Classification2 SkyAirTempHumidityWindWaterForecastEnjoySport SunnyWarmNormalStrongWarmSameYes SunnyWarmHighStrongWarmSameYes.
Lecture 10: 8/6/1435 Machine Learning Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
1 Concept Learning By Dong Xu State Key Lab of CAD&CG, ZJU.
Machine Learning Chapter 2. Concept Learning and The General-to-specific Ordering Tom M. Mitchell.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, January 22, 2001 William.
Chapter 2: Concept Learning and the General-to-Specific Ordering.
1 Machine Learning 1.Where does machine learning fit in computer science? 2.What is machine learning? 3.Where can machine learning be applied? 4.Should.
Learning from observations
CpSc 810: Machine Learning Concept Learning and General to Specific Ordering.
Concept Learning and the General-to-Specific Ordering 이 종우 자연언어처리연구실.
Outline Inductive bias General-to specific ordering of hypotheses
Overview Concept Learning Representation Inductive Learning Hypothesis
Artificial Intelligence 8. Supervised and unsupervised learning Japan Advanced Institute of Science and Technology (JAIST) Yoshimasa Tsuruoka.
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Thursday, August 26, 1999 William.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 34 of 41 Wednesday, 10.
Machine Learning: Lecture 2
Introduction Welcome Machine Learning.
Machine Learning Concept Learning General-to Specific Ordering
Kansas State University Department of Computing and Information Sciences CIS 690: Implementation of High-Performance Data Mining Systems Thursday, 20 May.
CS464 Introduction to Machine Learning1 Concept Learning Inducing general functions from specific training examples is a main issue of machine learning.
Concept Learning and The General-To Specific Ordering
Computational Learning Theory Part 1: Preliminaries 1.
Concept learning Maria Simi, 2011/2012 Machine Learning, Tom Mitchell Mc Graw-Hill International Editions, 1997 (Cap 1, 2).
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
Introduction to Classification & Clustering Villanova University Machine Learning Lab Module 4.
Brief Intro to Machine Learning CS539
Chapter 2 Concept Learning
Machine Learning for Computer Security
Introduction to Classification & Clustering
Machine Learning overview Chapter 18, 21
CSE543: Machine Learning Lecture 2: August 6, 2014
CS 9633 Machine Learning Concept Learning
Analytical Learning Discussion (4 of 4):
Ordering of Hypothesis Space
AV Autonomous Vehicles.
Basic Intro Tutorial on Machine Learning and Data Mining
3.1.1 Introduction to Machine Learning
Concept Learning Berlin Chen 2005 References:
Machine Learning Chapter 2
Machine Learning overview Chapter 18, 21
Version Space Machine Learning Fall 2018.
Machine Learning Chapter 2
Presentation transcript:

Artificial Intelligence 6. Machine Learning, Version Space Method Japan Advanced Institute of Science and Technology (JAIST) Yoshimasa Tsuruoka

Outline Introduction to machine learning Version space method What is machine learning? Applications of machine learning Version space method Representing hypotheses, version space Find-S algorithm Candidate-Elimination algorithm http://www.jaist.ac.jp/~tsuruoka/lectures/

Recognizing handwritten digits Hastie, Tibshirani and Friedman (2008). The Elements of Statistical Learning (2nd edition). Springer-Verlag.

Natural language processing GENIA tagger Tokenization Part-of-speech tagging Shallow parsing Named entity recognition http://www-tsujii.is.s.u-tokyo.ac.jp/GENIA/tagger/

Applications of machine learning Image/speech recognition Part-of-speech tagging, syntactic parsing, word sense disambiguation Detection of spam emails Intrusion detection Credit card fraud detection Automatic driving AI players in computer games etc.

Types of machine learning Supervised learning “correct” output is given for each instance Unsupervised learning No output is given Analyses relations between instances Reinforcement learning Supervision is given via “rewards”

Application of Unsupervised learning Search engine + clustering http://clusty.com

Reinforcement learning Autonomous helicopters and robots Inverted autonomous helicopter flight via reinforcement learning, Andrew Y. Ng, Adam Coates, Mark Diel, Varun Ganapathi, Jamie Schulte, Ben Tse, Eric Berger and Eric Liang. In International Symposium on Experimental Robotics, 2004 Quadruped Robot Obstacle Negotiation via Reinforcement Learning, Honglak Lee, Yirong Shen, Chih-Han Yu, Gurjeet Singh, and Andrew Y. Ng. In Proceedings of the IEEE International Conference on Robotics and Automation , 2006

What is machine learning? What does a machine learn? What “machine learning” can do: Classification, regression, structured prediction Clustering Machine learning involves Numerical optimization, probabilistic modeling, graphs, search, logic, etc.

Why machine learning? Why not write rules manually? Too many rules Detecting spam emails If the mail contains the word “Nigeria” then it is a spam If the mail comes from IP X.X.X.X then it is a spam If the mail contains a large image then it is a spam … Too many rules Hard to keep consistency Each rule may not be completely correct

Version space method Chapter 2 of Mitchell, T Version space method Chapter 2 of Mitchell, T., Machine Learning (1997) Concept Learning Training examples Representing hypotheses Find-S algorithm Version space Candidate-Elimination algorithm

Learning a concept with examples Training examples The concept we want to learn Days on which my friend Aldo enjoys his favorite water sports attributes Ex. Sky AirTemp Humidity Wind Water Forecast EnjoySport 1 Sunny Warm Normal Strong Same Yes 2 High 3 Rainy Cold Change No 4 Cool

Hypotheses Representing hypotheses General and Specific h1 = <Sunny, ?, ?, Strong, ?, ?> Weather = Sunny, Wind = Strong (the other attributes can take any values) h2 = <Sunny, ?, ?, ?, ?, ?> Weather = Sunny General and Specific h1 is more specific than h2 (h2 is more general than h1)

Find-S Algorithm Initialize h to the most specific hypothesis in H For each positive training instance x For each attribute constraint ai in h If the constraint ai is satisfied by x Then do nothing Else replace ai in h by the next more general constraint that is satisfied by x Output hypothesis h

Example h0 = <0, 0, 0, 0, 0, 0> x1 = <Sunny, Warm, Normal, Strong, Warm, Same>, yes h1 = <Sunny, Warm, Normal, Strong, Warm, Same> x2 = <Sunny, Warm, High, Strong, Warm, Same>, yes h2 = <Sunny, Warm, ?, Strong, Warm, Same> x3 = <Rainy, Cold, High, Strong, Warm, Change>, no h3 = <Sunny, Warm, ?, Strong, Warm, Same> x4 = <Sunny, Warm, High, Strong, Cool, Change>, yes h4 = <Sunny, Warm, ?, Strong, ?, ?>

Problems with the Find-S algorithm It is not clear whether the output hypothsis is the “correct” hypothesis There can be other hypotheses that are consistent with the training examples. Why prefer the most specific hypothesis? Cannot detect when the training data is inconsistent

Version Space Definition Hypothesis space H Training examples D The subset of hypotheses from H consistent with the training examples in D

LIST-THEN-ELIMINATE algorithm VersionSpace ← initialized with a list containing every hypothesis in H For each training example, <x, c(x)> Remove from VersionSpace any hypothesis h for which h(x) ≠ c(x) Output the list of hypothesis in VersionSpace

Version Space Specific boundary and General boundary S: { <Sunny, Warm, ?, Strong, ?, ?> } <Sunny, ?, ?, Strong, ?, ?> <Sunny, Warm, ?, ?, ?, ?> <?, Warm, ?, Strong, ?, ?> G: { <Sunny, ?, ?, ?, ?, ?>, <?, Warm, ?, ?, ?, ?> } The version space can be represented with S and G. You don’t have to list all the hypotheses.

Candidate-Elimination algorithm Initialization G: the set of maximally general hypotheses in H S: the set of maximally specific hypotheses in H For each training example d, do If d is a positive example Remove from G any hypothesis inconsistent with d For each hypothesis s in S that is not consistent with d Remove s from S Add to S all minimal generalization h of s such that h is consistent with d, and some member of G is more general than h Remove from S any hypothesis that is more general than another hypothesis in S If d is a negative example …

Example 1st training example <Sunny, Warm, Normal, Strong, Warm, Same>, yes S0: { <0, 0, 0, 0, 0, 0> } S1: { <Sunny, Warm, Normal, Strong, Warm, Same> } G0, G1: { <?, ?, ?, ?, ?, ?> }

Example 2nd training example <Sunny, Warm, High, Strong, Warm, Same>, yes S1: { <Sunny, Warm, Normal, Strong, Warm, Same> } S2: { <Sunny, Warm, ?, Strong, Warm, Same> } G0, G1 , G2 : { <?, ?, ?, ?, ?, ?> }

Example 3rd training example <Rainy, Cold, High, Strong, Warm, Change>, no S2,S3 :{ <Sunny, Warm, ?, Strong, Warm, Same> } G3: { <Sunny, ?, ?, ?, ?, ?> <?, Warm, ?, ?, ?, ?> <?, ?, ?, ?, ?, Same> } G2 : { <?, ?, ?, ?, ?, ?> }

Example 4th training example <Sunny, Warm, High, Strong, Cool, Change>, yes S3 :{ <Sunny, Warm, ?, Strong, Warm, Same> } S4 :{ <Sunny, Warm, ?, Strong, ?, ?> } G4: { <Sunny, ?, ?, ?, ?, ?> <?, Warm, ?, ?, ?, ?> } G3: { <Sunny, ?, ?, ?, ?, ?> <?, Warm, ?, ?, ?, ?> <?, ?, ?, ?, ?, Same> }

The final version space S4 :{ <Sunny, Warm, ?, Strong, ?, ?> } <Sunny, ?, ?, Strong, ?, ?> <Sunny, Warm, ?, ?, ?, ?> <?, Warm, ?, Strong, ?, ?> G4: { <Sunny, ?, ?, ?, ?, ?> <?, Warm, ?, ?, ?, ?> }