For Monday Finish chapter 19 Take-home exam due. Program 4 Any questions?

Slides:



Advertisements
Similar presentations
Explanation-Based Learning (borrowed from mooney et al)
Advertisements

Analytical Learning.
Concept Learning and the General-to-Specific Ordering
1 CS 391L: Machine Learning: Rule Learning Raymond J. Mooney University of Texas at Austin.
Computer Science CPSC 322 Lecture 25 Top Down Proof Procedure (Ch 5.2.2)
1 Machine Learning: Lecture 3 Decision Tree Learning (Based on Chapter 3 of Mitchell T.., Machine Learning, 1997)
Data Mining and Machine Learning
1er. Escuela Red ProTIC - Tandil, de Abril, Decision Tree Learning 3.1 Introduction –Method for approximation of discrete-valued target functions.
Methods of Proof Chapter 7, second half.. Proof methods Proof methods divide into (roughly) two kinds: Application of inference rules: Legitimate (sound)
Answer Set Programming Overview Dr. Rogelio Dávila Pérez Profesor-Investigador División de Posgrado Universidad Autónoma de Guadalajara
Inductive Logic Programming: The Problem Specification Given: –Examples: first-order atoms or definite clauses, each labeled positive or negative. –Background.
For Wednesday Read chapter 19, sections 1-3 No homework.
Knowledge Representation and Reasoning Learning Sets of Rules and Analytical Learning Harris Georgiou – 4.
Combining Inductive and Analytical Learning Ch 12. in Machine Learning Tom M. Mitchell 고려대학교 자연어처리 연구실 한 경 수
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 21 Jim Martin.
Relational Data Mining in Finance Haonan Zhang CFWin /04/2003.
Machine Learning: Symbol-Based
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 22 Jim Martin.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
17.5 Rule Learning Given the importance of rule-based systems and the human effort that is required to elicit good rules from experts, it is natural to.
Issues with Data Mining
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
Mohammad Ali Keyvanrad
1 Machine Learning: Lecture 11 Analytical Learning / Explanation-Based Learning (Based on Chapter 11 of Mitchell, T., Machine Learning, 1997)
For Wednesday No new reading Homework: –Chapter 18, exercises 3, 4, 7.
For Monday Read chapter 18, sections 5-6 Homework: –Chapter 18, exercises 1-2.
Introduction to ILP ILP = Inductive Logic Programming = machine learning  logic programming = learning with logic Introduced by Muggleton in 1992.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, February 7, 2001.
1 Artificial Neural Networks Sanun Srisuk EECP0720 Expert Systems – Artificial Neural Networks.
CS 391L: Machine Learning: Ensembles
Machine Learning CSE 681 CH2 - Supervised Learning.
Theory Revision Chris Murphy. The Problem Sometimes we: – Have theories for existing data that do not match new data – Do not want to repeat learning.
For Friday No reading No homework. Program 4 Exam 2 A week from Friday Covers 10, 11, 13, 14, 18, Take home due at the exam.
November 10, Machine Learning: Lecture 9 Rule Learning / Inductive Logic Programming.
Machine Learning Chapter 2. Concept Learning and The General-to-specific Ordering Tom M. Mitchell.
Chapter 2: Concept Learning and the General-to-Specific Ordering.
For Wednesday No reading Homework: –Chapter 18, exercise 6.
Predicate Calculus Syntax Countable set of predicate symbols, each with specified arity  0. For example: clinical data with multiple tables of patient.
Machine Learning Chapter 5. Artificial IntelligenceChapter 52 Learning 1. Rote learning rote( โรท ) n. วิถีทาง, ทางเดิน, วิธีการตามปกติ, (by rote จากความทรงจำ.
CpSc 810: Machine Learning Concept Learning and General to Specific Ordering.
Concept Learning and the General-to-Specific Ordering 이 종우 자연언어처리연구실.
Outline Inductive bias General-to specific ordering of hypotheses
Overview Concept Learning Representation Inductive Learning Hypothesis
For Monday No new reading Homework: –Chapter 18, exercises 3 and 4.
For Monday Finish chapter 19 No homework. Program 4 Any questions?
Automated Reasoning Early AI explored how to automated several reasoning tasks – these were solved by what we might call weak problem solving methods as.
For Wednesday Read 20.4 Lots of interesting stuff in chapter 20, but we don’t have time to cover it all.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
CS 5751 Machine Learning Chapter 10 Learning Sets of Rules1 Learning Sets of Rules Sequential covering algorithms FOIL Induction as the inverse of deduction.
For Wednesday Read ch. 20, sections 1, 2, 5, and 7 No homework.
Machine Learning Concept Learning General-to Specific Ordering
Ensemble Methods in Machine Learning
CpSc 810: Machine Learning Analytical learning. 2 Copy Right Notice Most slides in this presentation are adopted from slides of text book and various.
Data Mining and Decision Support
CS 8751 ML & KDDComputational Learning Theory1 Notions of interest: efficiency, accuracy, complexity Probably, Approximately Correct (PAC) Learning Agnostic.
More Symbolic Learning CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
Concept Learning and The General-To Specific Ordering
Computational Learning Theory Part 1: Preliminaries 1.
Chap. 10 Learning Sets of Rules 박성배 서울대학교 컴퓨터공학과.
Outline Decision tree representation ID3 learning algorithm Entropy, Information gain Issues in decision tree learning 2.
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
Instance-Based Learning Evgueni Smirnov. Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based.
CS 9633 Machine Learning Explanation Based Learning
Chapter 2 Concept Learning
CS 9633 Machine Learning Concept Learning
CS 9633 Machine Learning Inductive-Analytical Methods
Data Mining Practical Machine Learning Tools and Techniques
Rule Learning Hankui Zhuo April 28, 2018.
Version Space Machine Learning Fall 2018.
Presentation transcript:

For Monday Finish chapter 19 Take-home exam due

Program 4 Any questions?

Beyond a Single Learner Ensembles of learners work better than individual learning algorithms Several possible ensemble approaches: –Ensembles created by using different learning methods and voting –Bagging –Boosting

Bagging Random selections of examples to learn the various members of the ensemble. Seems to work fairly well, but no real guarantees.

Boosting Most used ensemble method Based on the concept of a weighted training set. Works especially well with weak learners. Start with all weights at 1. Learn a hypothesis from the weights. Increase the weights of all misclassified examples and decrease the weights of all correctly classified examples. Learn a new hypothesis. Repeat

Approaches to Learning Maintaining a single current best hypothesis Least commitment (version space) learning

Different Ways of Incorporating Knowledge in Learning Explanation Based Learning (EBL) Theory Revision (or Theory Refinement) Knowledge Based Inductive Learning (in first-order logic - Inductive Logic Programming (ILP)

Explanation Based Learning Requires two inputs –Labeled examples (maybe very few) –Domain theory Goal –To produce operational rules that are consistent with both examples and theory –Classical EBL requires that the theory entail the resulting rules

Why Do EBL? Often utilitarian or speed-up learning Example: DOLPHIN –Uses EBL to improve planning –Both speed-up learning and improving plan quality

Theory Refinement Inputs the same as EBL –Theory –Examples Goal –Fix the theory so that it agrees with the examples Theory may be incomplete or wrong

Why Do Theory Refinement? Potentially more accurate than induction alone Able to learn from fewer examples May influence the structure of the theory to make it more comprehensible to experts

How Is Theory Refinement Done? Initial State: Initial Theory Goal State: Theory that fits training data. Operators: Atomic changes to the syntax of a theory: –Delete rule or antecedent, add rule or antecedent –Increase parameter, Decrease parameter –Delete node or link, add node or link Path cost: Number of changes made, or total cost of changes made.

Theory Refinement As Heuristic Search Finding the “closest” theory that is consistent with the data is generally intractable (NP­hard). Complete consistency with training data is not always desirable, particularly if the data is noisy. Therefore, most methods employ some form of greedy or hill­climibing search. Also, usually employ some form of over­fitting avoidance method to avoid learning an overly complex revision.

Theory Refinement As Bias Bias is to learn a theory which is syntactically similar to the initial theory. Distance can be measured in terms of the number of edit operations needed to revise the theory (edit distance). Assumes the syntax of the initial theory is “approximately correct.” A bias for minimal semantic revision would simply involve memorizing the exceptions to the theory, which is undesirable with respect to generalizing to novel data.

Inductive Logic Programming Representation is Horn clauses Builds rules using background predicates Rules are potentially much more expressive than attribute-value representations

Example Results Rules for family relations from data of primitive or related predicates. uncle(A,B) :­ brother(A,C), parent(C,B). uncle(A,B) :­ husband(A,C), sister(C,D), parent(D,B). Recursive list programs. member(X,[X | Y]). member(X, [Y | Z]) :­ member(X, Z).

ILP Goal is to induce a Horn­clause definition for some target predicate P given definitions of background predicates Q i. Goal is to find a syntactically simple definition D for P such that given background predicate definitions B –For every positive example p i : D  B |= p –For every negative example n i : D  B |/= n Background definitions are either provided –Extensionally: List of ground tuples satisfying the predicate. –Intensionally: Prolog definition of the predicate.

Sequential Covering Algorithm Let P be the set of positive examples Until P is empty do Learn a rule R that covers a large number of positives without covering any negatives. Add R to the list of learned rules. Remove positives covered by R from P

This is just an instance of the greedy algorithm for minimum set covering and does not guarantee that a minimum number of rules is learned but tends to learn a reasonably small rule set. Minimum set covering is an NP­hard problem and the greedy algorithm is a common approximation algorithm. There are several ways to learn a single rule used in various methods.

Strategies for Learning a Single Rule Top­Down (General to Specific): –Start with the most general (empty) rule. –Repeatedly add feature constraints that eliminate negatives while retaining positives. –Stop when only positives are covered. Bottom­Up (Specific to General): –Start with a most specific rule (complete description of a single instance). –Repeatedly eliminate feature constraints in order to cover more positive examples. –Stop when further generalization results in covering negatives.