Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 34 of 42 Monday, 13 November.

Slides:



Advertisements
Similar presentations
Decision Trees Decision tree representation ID3 learning algorithm
Advertisements

ICS320-Foundations of Adaptive and Learning Systems
Classification Techniques: Decision Tree Learning
Computing & Information Sciences Kansas State University Lecture 35 of 42 CIS 530 / 730 Artificial Intelligence Lecture 35 of 42 Machine Learning: Artificial.
CS 484 – Artificial Intelligence1 Announcements Project 1 is due Tuesday, October 16 Send me the name of your konane bot Midterm is Thursday, October 18.
Università di Milano-Bicocca Laurea Magistrale in Informatica
Computing & Information Sciences Kansas State University Lecture 33 of 42 CIS 530 / 730 Artificial Intelligence Lecture 33 of 42 Machine Learning: Version.
Decision Tree Algorithm
Computing & Information Sciences Kansas State University Lecture 34 of 42 CIS 530 / 730 Artificial Intelligence Lecture 34 of 42 Machine Learning: Decision.
Adapted by Doug Downey from: Bryan Pardo, EECS 349 Fall 2007 Machine Learning Lecture 2: Concept Learning and Version Spaces 1.
Computing & Information Sciences Kansas State University Lecture 11 of 42 CIS 530 / 730 Artificial Intelligence Lecture 11 of 42 William H. Hsu Department.
Induction of Decision Trees
Data Mining with Decision Trees Lutz Hamel Dept. of Computer Science and Statistics University of Rhode Island.
MACHINE LEARNING. What is learning? A computer program learns if it improves its performance at some task through experience (T. Mitchell, 1997) A computer.
Machine Learning Lecture 10 Decision Trees G53MLE Machine Learning Dr Guoping Qiu1.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, January 19, 2001.
Computing & Information Sciences Kansas State University Lecture 01 of 42 Wednesday, 24 January 2008 William H. Hsu Department of Computing and Information.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Friday, 25 January 2008 William.
Computing & Information Sciences Kansas State University Wednesday, 15 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 35 of 42 Wednesday, 15 November.
Machine Learning Chapter 3. Decision Tree Learning
CS-424 Gregory Dudek Today’s outline Administrative issues –Assignment deadlines: 1 day = 24 hrs (holidays are special) –The project –Assignment 3 –Midterm.
1 Machine Learning What is learning?. 2 Machine Learning What is learning? “That is what learning is. You suddenly understand something you've understood.
Machine Learning Chapter 11.
Lecture 7. Outline 1. Overview of Classification and Decision Tree 2. Algorithm to build Decision Tree 3. Formula to measure information 4. Weka, data.
CpSc 810: Machine Learning Decision Tree Learning.
General-to-Specific Ordering. 8/29/03Logic Based Classification2 SkyAirTempHumidityWindWaterForecastEnjoySport SunnyWarmNormalStrongWarmSameYes SunnyWarmHighStrongWarmSameYes.
Computing & Information Sciences Kansas State University Lecture 10 of 42 CIS 530 / 730 Artificial Intelligence Lecture 10 of 42 William H. Hsu Department.
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, February 5, 2001.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.3: Decision Trees Rodney Nielsen Many of.
Machine Learning Chapter 2. Concept Learning and The General-to-specific Ordering Tom M. Mitchell.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, January 22, 2001 William.
Chapter 2: Concept Learning and the General-to-Specific Ordering.
Computing & Information Sciences Kansas State University Friday, 10 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 33 of 42 Friday, 10 November.
For Wednesday No reading Homework: –Chapter 18, exercise 6.
CpSc 810: Machine Learning Concept Learning and General to Specific Ordering.
Concept Learning and the General-to-Specific Ordering 이 종우 자연언어처리연구실.
Outline Inductive bias General-to specific ordering of hypotheses
Overview Concept Learning Representation Inductive Learning Hypothesis
CS690L Data Mining: Classification
For Monday No new reading Homework: –Chapter 18, exercises 3 and 4.
Computing & Information Sciences Kansas State University Wednesday, 15 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 35 of 42 Wednesday, 15 November.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 9 of 42 Wednesday, 14.
1 Decision Tree Learning Original slides by Raymond J. Mooney University of Texas at Austin.
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Thursday, August 26, 1999 William.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 34 of 41 Wednesday, 10.
Kansas State University Department of Computing and Information Sciences CIS 690: Implementation of High-Performance Data Mining Systems Wednesday, 21.
Machine Learning Concept Learning General-to Specific Ordering
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Wednesday, 31 January 2007.
Kansas State University Department of Computing and Information Sciences CIS 690: Implementation of High-Performance Data Mining Systems Thursday, 20 May.
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Tuesday, September 7, 1999.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Friday, Febuary 2, 2001 Presenter:Ajay.
Concept Learning and The General-To Specific Ordering
Computing & Information Sciences Kansas State University CIS 530 / 730: Artificial Intelligence Lecture 09 of 42 Wednesday, 17 September 2008 William H.
Concept learning Maria Simi, 2011/2012 Machine Learning, Tom Mitchell Mc Graw-Hill International Editions, 1997 (Cap 1, 2).
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Thursday, 25 January 2007 William.
Chapter 2 Concept Learning
CS 9633 Machine Learning Concept Learning
Artificial Intelligence
Analytical Learning Discussion (4 of 4):
Data Science Algorithms: The Basic Methods
Machine Learning Chapter 3. Decision Tree Learning
Machine Learning: Lecture 3
Machine Learning Chapter 3. Decision Tree Learning
Machine Learning Chapter 2
Machine Learning Chapter 2
Presentation transcript:

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 34 of 42 Monday, 13 November 2006 William H. Hsu Department of Computing and Information Sciences, KSU KSOL course page: Course web site: Instructor home page: Reading for Next Class: Section 20.1, Russell & Norvig 2 nd edition Introduction to Machine Learning Discussion: BNJ

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture Outline Today’s Reading: Sections 18.3, R&N 2e Wednesday’s Reading: Section 20.1, R&N 2e Machine Learning  Definition  Supervised learning and hypothesis space Finding Hypotheses  Version spaces  Candidate elimination Decision Trees  Induction  Greedy learning  Entropy

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence A Target Function for Learning to Play Checkers Possible Definition  If b is a final board state that is won, then V(b) = 100  If b is a final board state that is lost, then V(b) = -100  If b is a final board state that is drawn, then V(b) = 0  If b is not a final board state in the game, then V(b) = V(b’) where b’ is the best final board state that can be achieved starting from b and playing optimally until the end of the game  Correct values, but not operational Choosing a Representation for the Target Function  Collection of rules?  Neural network?  Polynomial function (e.g., linear, quadratic combination) of board features?  Other? A Representation for Learned Function   bp/rp = number of black/red pieces; bk/rk = number of black/red kings; bt/rt = number of black/red pieces threatened (can be taken on next turn)

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Obtaining Training Examples  the target function  the learned function  the training value One Rule For Estimating Training Values:  Choose Weight Tuning Rule  Least Mean Square (LMS) weight update rule: REPEAT Select a training example b at random Compute the error(b) for this training example For each board feature f i, update weight w i as follows: where c is a small, constant factor to adjust the learning rate A Training Procedure for Learning to Play Checkers

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Design Choices for Learning to Play Checkers Completed Design Determine Type of Training Experience Games against experts Games against self Table of correct moves Determine Target Function Board  valueBoard  move Determine Representation of Learned Function Polynomial Linear function of six features Artificial neural network Determine Learning Algorithm Gradient descent Linear programming

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Hypothesis Space Search by Find-S Instances XHypotheses H x 1 =, + x 2 =, + x 3 =, - x 4 =, + h 1 = h 2 = h 3 = h 4 = h 5 = Shortcomings of Find-S  Can’t tell whether it has learned concept  Can’t tell when training data inconsistent  Picks a maximally specific h (why?)  Depending on H, there might be several! h1h1 h0h0 h 2,3 h4h x3x3 x1x1 x2x2 x4x4

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Version Spaces Definition: Consistent Hypotheses  A hypothesis h is consistent with a set of training examples D of target concept c if and only if h(x) = c(x) for each training example in D.  Consistent (h, D)    D. h(x) = c(x) Definition: Version Space  The version space VS H,D, with respect to hypothesis space H and training examples D, is the subset of hypotheses from H consistent with all training examples in D.  VS H,D  { h  H | Consistent (h, D) }

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence 1. Initialization G  (singleton) set containing most general hypothesis in H, denoted { } S  set of most specific hypotheses in H, denoted { } 2. For each training example d If d is a positive example (Update-S) Remove from G any hypotheses inconsistent with d For each hypothesis s in S that is not consistent with d Remove s from S Add to S all minimal generalizations h of s such that 1. h is consistent with d 2. Some member of G is more general than h (These are the greatest lower bounds, or meets, s  d, in VS H,D ) Remove from S any hypothesis that is more general than another hypothesis in S (remove any dominated elements) Candidate Elimination Algorithm [1]

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Candidate Elimination Algorithm [2] (continued) If d is a negative example (Update-G) Remove from S any hypotheses inconsistent with d For each hypothesis g in G that is not consistent with d Remove g from G Add to G all minimal specializations h of g such that 1. h is consistent with d 2. Some member of S is more specific than h (These are the least upper bounds, or joins, g  d, in VS H,D ) Remove from G any hypothesis that is less general than another hypothesis in G (remove any dominating elements)

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence S2S2 = G 2 Example Trace S0S0 G0G0 d 1 : d 2 : d 3 : d 4 : = S 3 G3G3 S4S4 G4G4 S1S1 = G 1

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence An Unbiased Learner Example of A Biased H  Conjunctive concepts with don’t cares  What concepts can H not express? (Hint: what are its syntactic limitations?) Idea  Choose H’ that expresses every teachable concept  i.e., H’ is the power set of X  Recall: | A  B | = | B | | A | (A = X; B = {labels}; H’ = A  B)  {{Rainy, Sunny}  {Warm, Cold}  {Normal, High}  {None, Mild, Strong}  {Cool, Warm}  {Same, Change}}  {0, 1} An Exhaustive Hypothesis Language  Consider: H’ = disjunctions (  ), conjunctions (  ), negations (¬) over previous H  | H’ | = 2 ( ) = 2 96 ; | H | = 1 + ( ) = 973 What Are S, G For The Hypothesis Language H’?  S  disjunction of all positive examples  G  conjunction of all negated negative examples

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Decision Trees Classifiers: Instances (Unlabeled Examples) Internal Nodes: Tests for Attribute Values  Typical: equality test (e.g., “Wind = ?”)  Inequality, other tests possible Branches: Attribute Values  One-to-one correspondence (e.g., “Wind = Strong”, “Wind = Light”) Leaves: Assigned Classifications (Class Labels) Representational Power: Propositional Logic (Why?) Outlook? Humidity?Wind? Maybe SunnyOvercastRain YesNo HighNormal MaybeNo StrongLight Decision Tree for Concept PlayTennis

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Example: Decision Tree to Predict C-Section Risk Learned from Medical Records of 1000 Women Negative Examples are Cesarean Sections  Prior distribution: [833+, 167-] 0.83+,  Fetal-Presentation = 1: [822+, 116-]0.88+,  Previous-C-Section = 0: [767+, 81-] 0.90+, –Primiparous = 0: [399+, 13-]0.97+, –Primiparous = 1: [368+, 68-]0.84+, Fetal-Distress = 0: [334+, 47-]0.88+, – Birth-Weight  , – Birth-Weight < , Fetal-Distress = 1: [34+, 21-]0.62+,  Previous-C-Section = 1: [55+, 35-] 0.61+,  Fetal-Presentation = 2: [3+, 29-]0.11+,  Fetal-Presentation = 3: [8+, 22-]0.27+, 0.73-

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence [21+, 5-][8+, 30-] Decision Tree Learning: Top-Down Induction (ID3) A1A1 TrueFalse [29+, 35-] [18+, 33-][11+, 2-] A2A2 TrueFalse [29+, 35-] Algorithm Build-DT (Examples, Attributes) IF all examples have the same label THEN RETURN (leaf node with label) ELSE IF set of attributes is empty THEN RETURN (leaf with majority label) ELSE Choose best attribute A as root FOR each value v of A Create a branch out of the root for the condition A = v IF {x  Examples: x.A = v} = Ø THEN RETURN (leaf with majority label) ELSE Build-DT ({x  Examples: x.A = v}, Attributes ~ {A}) But Which Attribute Is Best?

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Choosing the “Best” Root Attribute Objective  Construct a decision tree that is a small as possible (Occam’s Razor)  Subject to: consistency with labels on training data Obstacles  Finding the minimal consistent hypothesis (i.e., decision tree) is NP-hard (D’oh!)  Recursive algorithm (Build-DT)  A greedy heuristic search for a simple tree  Cannot guarantee optimality (D’oh!) Main Decision: Next Attribute to Condition On  Want: attributes that split examples into sets that are relatively pure in one label  Result: closer to a leaf node  Most popular heuristic  Developed by J. R. Quinlan  Based on information gain  Used in ID3 algorithm

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence A Measure of Uncertainty  The Quantity  Purity: how close a set of instances is to having just one label  Impurity (disorder): how close it is to total uncertainty over labels  The Measure: Entropy  Directly proportional to impurity, uncertainty, irregularity, surprise  Inversely proportional to purity, certainty, regularity, redundancy Example  For simplicity, assume H = {0, 1}, distributed according to Pr(y)  Can have (more than 2) discrete class labels  Continuous random variables: differential entropy  Optimal purity for y: either  Pr(y = 0) = 1, Pr(y = 1) = 0  Pr(y = 1) = 1, Pr(y = 0) = 0  What is the least pure probability distribution?  Pr(y = 0) = 0.5, Pr(y = 1) = 0.5  Corresponds to maximum impurity/uncertainty/irregularity/surprise  Property of entropy: concave function (“concave downward”) Entropy: Intuitive Notion p + = Pr(y = +) 1.0 H(p) = Entropy(p)

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Entropy: Information Theoretic Definition Components  D: a set of examples {,, …, }  p + = Pr(c(x) = +), p - = Pr(c(x) = -) Definition  H is defined over a probability density function p  D contains examples whose frequency of + and - labels indicates p + and p - for the observed data  The entropy of D relative to c is: H(D)  -p + log b (p + ) - p - log b (p - ) What Units is H Measured In?  Depends on the base b of the log (bits for b = 2, nats for b = e, etc.)  A single bit is required to encode each example in the worst case (p + = 0.5)  If there is less uncertainty (e.g., p + = 0.8), we can use less than 1 bit each

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Information Gain: Information Theoretic Definition Partitioning on Attribute Values  Recall: a partition of D is a collection of disjoint subsets whose union is D  Goal: measure the uncertainty removed by splitting on the value of attribute A Definition  The information gain of D relative to attribute A is the expected reduction in entropy due to splitting (“sorting”) on A: where D v is {x  D: x.A = v}, the set of examples in D where attribute A has value v  Idea: partition on A; scale entropy to the size of each subset D v Which Attribute Is Best? [21+, 5-][8+, 30-] A1A1 TrueFalse [29+, 35-] [18+, 33-][11+, 2-] A2A2 TrueFalse [29+, 35-]

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Constructing A Decision Tree for PlayTennis using ID3 [1] Selecting The Root Attribute Prior (unconditioned) distribution: 9+, 5-  H(D) = -(9/14) lg (9/14) - (5/14) lg (5/14) bits = 0.94 bits  H(D, Humidity = High) = -(3/7) lg (3/7) - (4/7) lg (4/7) = bits  H(D, Humidity = Normal) = -(6/7) lg (6/7) - (1/7) lg (1/7) = bits  Gain(D, Humidity) = (7/14) * (7/14) * = bits  Similarly, Gain (D, Wind) = (8/14) * (6/14) * 1.0 = bits [6+, 1-][3+, 4-] Humidity HighNormal [9+, 5-] [3+, 3-][6+, 2-] Wind LightStrong [9+, 5-]

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Constructing A Decision Tree for PlayTennis using ID3 [2] Humidity?Wind? Yes NoYesNo Outlook? 1,2,3,4,5,6,7,8,9,10,11,12,13,14 [9+,5-] SunnyOvercastRain 1,2,8,9,11 [2+,3-] 3,7,12,13 [4+,0-] 4,5,6,10,14 [3+,2-] HighNormal 1,2,8 [0+,3-] 9,11 [2+,0-] StrongLight 6,14 [0+,2-] 4,5,10 [3+,0-]

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Decision Tree Overview Heuristic Search and Inductive Bias Decision Trees (DTs)  Can be boolean (c(x)  {+, -}) or range over multiple classes  When to use DT-based models Generic Algorithm Build-DT: Top Down Induction  Calculating best attribute upon which to split  Recursive partitioning Entropy and Information Gain  Goal: to measure uncertainty removed by splitting on a candidate attribute A  Calculating information gain (change in entropy)  Using information gain in construction of tree  ID3  Build-DT using Gain() ID3 as Hypothesis Space Search (in State Space of Decision Trees) Next: Artificial Neural Networks (Multilayer Perceptrons and Backprop) Tools to Try: WEKA, MLC++

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Inductive Bias (Inductive) Bias: Preference for Some h  H (Not Consistency with D Only) Decision Trees (DTs)  Boolean DTs: target concept is binary-valued (i.e., Boolean-valued)  Building DTs  Histogramming: a method of vector quantization (encoding input using bins)  Discretization: continuous input  discrete (e.g.., by histogramming) Entropy and Information Gain  Entropy H(D) for a data set D relative to an implicit concept c  Information gain Gain (D, A) for a data set partitioned by attribute A  Impurity, uncertainty, irregularity, surprise Heuristic Search  Algorithm Build-DT: greedy search (hill-climbing without backtracking)  ID3 as Build-DT using the heuristic Gain()  Heuristic : Search :: Inductive Bias : Inductive Generalization MLC++ (Machine Learning Library in C++)  Data mining libraries (e.g., MLC++) and packages (e.g., MineSet)  Irvine Database: the Machine Learning Database Repository at UCI

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Summary Points Taxonomies of Learning Definition of Learning: Task, Performance Measure, Experience Concept Learning as Search through H  Hypothesis space H as a state space  Learning: finding the correct hypothesis General-to-Specific Ordering over H  Partially-ordered set: Less-Specific-Than (More-General-Than) relation  Upper and lower bounds in H Version Space Candidate Elimination Algorithm  S and G boundaries characterize learner’s uncertainty  Version space can be used to make predictions over unseen cases Learner Can Generate Useful Queries Next Tuesday: When and Why Are Inductive Leaps Possible?

Computing & Information Sciences Kansas State University Monday, 13 Nov 2006CIS 490 / 730: Artificial Intelligence Terminology Supervised Learning  Concept – function from observations to categories (e.g., boolean-valued: +/-)  Target (function) - true function f  Hypothesis - proposed function h believed to be similar to f  Hypothesis space - space of all hypotheses that can be generated by the learning system  Example - tuples of the form  Instance space (aka example space) - space of all possible examples  Classifier - discrete-valued function whose range is a set of class labels The Version Space Algorithm  Algorithms: Find-S, List-Then-Eliminate, candidate elimination  Consistent hypothesis - one that correctly predicts observed examples  Version space - space of all currently consistent (or satisfiable) hypotheses Inductive Learning  Inductive generalization - process of generating hypotheses that describe cases not yet observed  The inductive learning hypothesis