Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Thursday, November 29, 2001.

Slides:



Advertisements
Similar presentations
Some Prolog Prolog is a logic programming language
Advertisements

Biointelligence Lab School of Computer Sci. & Eng.
1. An Overview of Prolog.
Standard Logical Equivalences
Logic Use mathematical deduction to derive new knowledge.
Logic.
Knowledge Representation and Reasoning Learning Sets of Rules and Analytical Learning Harris Georgiou – 4.
Chapter 10 Learning Sets Of Rules
Machine Learning Chapter 10. Learning Sets of Rules Tom M. Mitchell.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Tuesday, November 27, 2001 William.
Outline Recap Knowledge Representation I Textbook: Chapters 6, 7, 9 and 10.
CSE (c) S. Tanimoto, 2008 Propositional Logic
Inference in FOL Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 9 Spring 2004.
Relational Data Mining in Finance Haonan Zhang CFWin /04/2003.
Computing & Information Sciences Kansas State University Lecture 11 of 42 CIS 530 / 730 Artificial Intelligence Lecture 11 of 42 William H. Hsu Department.
Inductive Logic Programming Includes slides by Luis Tari CS7741L16ILP.
Notes for Chapter 12 Logic Programming The AI War Basic Concepts of Logic Programming Prolog Review questions.
Chapter 3: Methods of Inference
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Wednesday, 11 April 2007 William.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, February 7, 2001.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Wednesday, January 19, 2000.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 Wednesday, 20 October.
Pattern-directed inference systems
1 Logical Agents CS 171/271 (Chapter 7) Some text and images in these slides were drawn from Russel & Norvig’s published material.
November 10, Machine Learning: Lecture 9 Rule Learning / Inductive Logic Programming.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 16 Monday, 29 September.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Friday, 16 February 2007 William.
Machine Learning Chapter 2. Concept Learning and The General-to-specific Ordering Tom M. Mitchell.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, January 22, 2001 William.
Computing & Information Sciences Kansas State University Wednesday, 22 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 22 of 42 Wednesday, 22 October.
Computing & Information Sciences Kansas State University Wednesday, 20 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 12 of 42 Wednesday, 20 September.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 of 41 Monday, 25 October.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 11 of 41 Wednesday, 15.
Predicate Calculus Syntax Countable set of predicate symbols, each with specified arity  0. For example: clinical data with multiple tables of patient.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Thursday, 12 April 2007 William.
Overview Concept Learning Representation Inductive Learning Hypothesis
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 13 of 41 Monday, 20 September.
Computing & Information Sciences Kansas State University Lecture 13 of 42 CIS 530 / 730 Artificial Intelligence Lecture 13 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 17 Wednesday, 01 October.
The AI War LISP and Prolog Basic Concepts of Logic Programming
1 Logical Agents CS 171/271 (Chapter 7) Some text and images in these slides were drawn from Russel & Norvig’s published material.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 12 Friday, 17 September.
Computing & Information Sciences Kansas State University Lecture 14 of 42 CIS 530 / 730 Artificial Intelligence Lecture 14 of 42 William H. Hsu Department.
Logical Agents Chapter 7. Outline Knowledge-based agents Logic in general Propositional (Boolean) logic Equivalence, validity, satisfiability.
Computing & Information Sciences Kansas State University Monday, 25 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 14 of 42 Monday, 25 September.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 23 Friday, 17 October.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 14 of 41 Wednesday, 22.
CS 5751 Machine Learning Chapter 10 Learning Sets of Rules1 Learning Sets of Rules Sequential covering algorithms FOIL Induction as the inverse of deduction.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 15 of 41 Friday 24 September.
Computing & Information Sciences Kansas State University Lecture 15 of 42 CIS 530 / 730 Artificial Intelligence Lecture 15 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 18 of 41 Friday, 01 October.
Artificial Intelligence “Introduction to Formal Logic” Jennifer J. Burg Department of Mathematics and Computer Science.
CPSC 422, Lecture 21Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 21 Oct, 30, 2015 Slide credit: some slides adapted from Stuart.
Computing & Information Sciences Kansas State University Wednesday, 19 Sep 2007CIS 530 / 730: Artificial Intelligence Lecture 12 of 42 Wednesday, 19 September.
Knowledge Repn. & Reasoning Lec. #5: First-Order Logic UIUC CS 498: Section EA Professor: Eyal Amir Fall Semester 2004.
Computing & Information Sciences Kansas State University Lecture 12 of 42 CIS 530 / 730 Artificial Intelligence Lecture 12 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Thursday, 05 April 2007 William.
Computing & Information Sciences Kansas State University Wednesday, 13 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 10 of 42 Wednesday, 13 September.
First-Order Logic Semantics Reading: Chapter 8, , FOL Syntax and Semantics read: FOL Knowledge Engineering read: FOL.
Computing & Information Sciences Kansas State University CIS 530 / 730: Artificial Intelligence Lecture 09 of 42 Wednesday, 17 September 2008 William H.
Daniel Kroening and Ofer Strichman 1 Decision Procedures An Algorithmic Point of View Basic Concepts and Background.
Chap. 10 Learning Sets of Rules 박성배 서울대학교 컴퓨터공학과.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 14 of 42 Wednesday, 22.
Computing & Information Sciences Kansas State University Wednesday, 04 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 17 of 42 Wednesday, 04 October.
Computing & Information Sciences Kansas State University Monday, 22 Sep 2008CIS 530 / 730: Artificial Intelligence Lecture 11 of 42 Monday, 22 September.
Computing & Information Sciences Kansas State University Friday, 13 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 21 of 42 Friday, 13 October.
Computing & Information Sciences Kansas State University Monday, 18 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 11 of 42 Monday, 18 September.
CS 9633 Machine Learning Concept Learning
Horn Clauses and Unification
Horn Clauses and Unification
Presentation transcript:

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Thursday, November 29, 2001 William H. Hsu Department of Computing and Information Sciences, KSU Readings: Sections , Mitchell Section 21.4, Russell and Norvig Inductive Logic Programming (ILP) Lecture 27

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Lecture Outline Readings: Sections , Mitchell; Section 21.4, Russell and Norvig Suggested Exercises: 10.5, Mitchell Induction as Inverse of Deduction –Problem of inductive learning revisited –Operators for automated deductive inference Resolution rule for deduction First-order predicate calculus (FOPC) and resolution theorem proving –Inverting resolution Propositional case First-order case Inductive Logic Programming (ILP) –Cigol –Progol

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Induction as Inverted Deduction: Design Principles Recall: Definition of Induction –Induction: finding h such that   D. (B  D  x i ) |  f(x i ) A |  B means A logically entails B x i  ith target instance f(x i ) is the target function value for example x i (data set D = { }) Background knowledge B (e.g., inductive bias in inductive learning) Idea –Design inductive algorithm by inverting operators for automated deduction –Same deductive operators as used in theorem proving Theorem Prover Deductive System for Inductive Learning Training Examples New Instance Assertion { c  H } Inductive bias made explicit Classification of New Instance (or “Don’t Know”)

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Induction as Inverted Deduction: Example Deductive Query –“Pairs of people such that u is a child of v” –Relations (predicates) Child (target predicate) Father, Mother, Parent, Male, Female Learning Problem –Formulation Concept learning: target function f is Boolean-valued i.e., target predicate –Components Target function f(x i ): Child (Bob, Sharon) x i : Male (Bob), Female (Sharon), Father (Sharon, Bob) B: {Parent (x, y)  Father (x, y). Parent (x, y)  Mother (x, y).} –What satisfies   D. (B  D  x i ) |  f(x i )? h 1 : Child (u, v)  Father (v, u).- doesn’t use B h 2 : Child (u, v)  Parent (v, u).- uses B

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Perspectives on Learning and Inference Jevons (1874) –First published insight that induction can be interpreted as inverted deduction –“Induction is, in fact, the inverse operation of deduction, and cannot be conceived to exist without the corresponding operation, so that the question of relative importance cannot arise. Who thinks of asking whether addition or subtraction is the more important process in arithmetic? But at the same time much difference in difficulty may exist between a direct and inverse operation; … it must be allowed that inductive investigations are of a far higher degree of difficulty and complexity that any questions of deduction…” Aristotle (circa 330 B.C.) –Early views on learning from observations (examples) and interplay between induction and deduction –“… scientific knowledge through demonstration [i.e., deduction] is impossible unless a man knows the primary immediate premises… we must get to know the primary premises by induction; for the method by which even sense-perception implants the universal is inductive…”

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Induction as Inverted Deduction: Operators Deductive Operators –Have mechanical operators (F) for finding logically entailed conclusions (C) –F(A, B) = C where A  B |  C A, B, C: logical formulas F: deduction algorithm –Intuitive idea: apply deductive inference (aka sequent) rules to A, B to generate C Inductive Operators –Need operators O to find inductively inferred hypotheses (h, “primary premises”) –O(B, D) = h where   D. (B  D  x i ) |  f(x i ) B, D, h: logical formulas describing observations O: induction algorithm

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Induction as Inverted Deduction: Advantages and Disadvantages Advantages (Pros) –Subsumes earlier idea of finding h that “fits” training data –Domain theory B helps define meaning of “fitting” the data: B  D  x i |  f(x i ) –Suggests algorithms that search H guided by B Theory-guided constructive induction [Donoho and Rendell, 1995] aka Knowledge-guided constructive induction [Donoho, 1996] Disadvantages (Cons) –Doesn’t allow for noisy data Q: Why not? A: Consider what   D. (B  D  x i ) |  f(x i ) stipulates –First-order logic gives a huge hypothesis space H Overfitting… Intractability of calculating all acceptable h’s

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Deduction: Resolution Rule Intuitive Idea –Suppose we know P  L L  R –Can infer: P  R –Use this to reason over logical statements (propositions, first-order clauses) Resolution Rule –Sequent rule –1. Given: initial clauses C 1 and C 2, find literal L from clause C 1 such that  L occurs in clause C 2 –2. Form the resolvent C by including all literals from C 1 and C 2, except L and  L Set of literals occurring in conclusion C is C = (C 1 - {L})  (C 2 - {  L})  denotes set union, “-” denotes set difference

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Inverting Resolution: Example C: Pass-Exam   Study C 2 : Know-Material   Study C 1 : Pass-Exam   Know-Material Resolution C: Pass-Exam   Study C 2 : Know-Material   Study C 1 : Pass-Exam   Know-Material Inverse Resolution

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Inverted Resolution: Propositional Logic Problem Definition –Given: initial clauses C 1 and C –Return: clause C 2 such that C is resolvent of C 1 and C 2 Intuitive Idea –Reason from consequent and partial set of premises to unknown premises –Premise  hypothesis Inverted Resolution Procedure –1. Find literal L that occurs in C 1 but not in C –2. Form second clause C 2 by including the following literals: C 2 = (C - (C 1 - {L}))  {  L}

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Quick Review: First-Order Predicate Calculus (FOPC) Components of FOPC Formulas: Quick Intro to Terminology –Constants: e.g., John, Kansas, 42 –Variables: e.g., Name, State, x –Predicates: e.g., Father-Of, Greater-Than –Functions: e.g., age, cosine –Term: constant, variable, or function(term) –Literals (atoms): Predicate(term) or negation (e.g.,  Greater-Than (age (John), 42) –Clause: disjunction of literals with implicit universal quantification –Horn clause: at most one positive literal (H   L 1   L 2  …   L n ) FOPC: Representation Language for First-Order Resolution –aka First-Order Logic (FOL) –Applications Resolution using Horn clauses: logic programming (Prolog) Automated deduction (deductive inference), theorem proving –Goal: learn first-order rules by inverting first-order resolution

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition First-Order Resolution Intuitive Idea: Same as For Propositional Resolution Resolution Rule –Sequent rule: also same as for propositional resolution –1. Given: initial clauses C 1 and C 2, find literal L 1 from clause C 1, literal L 2 from clause C 2, and substitution  such that L 1  =  L 2  (found using unification) –2. Form the resolvent C by including all literals from C 1  and C 2 , except L 1  and  L 2  Set of literals occurring in conclusion C is C = (C 1 - {L 1 })   (C 2 - {  L 2 })   denotes set union, “-” denotes set difference Substitution  applied to sentences with matched literals removed

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Inverted Resolution: First-Order Logic Problem Definition –As for inverted propositional resolution Given: initial clauses C 1 and C Return: clause C 2 such that C is resolvent of C 1 and C 2 –Difference: must find, apply substitutions and inverse substitutions Inverted Resolution Procedure –1. Find literal L 1 that occurs in C 1 but doesn’t match any literal in C –2. Form second clause C 2 by including the following literals: C 2 = (C - (C 1 - {L 1 })  1 )  2 -1  {  L 1  1  2 -1 }

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Inverse Resolution Algorithm (Cigol): Example GrandChild (Bob, Shannon) Father (Shannon, Tom) Father (Tom, Bob) GrandChild (Bob, x)   Father (x, Tom) {Shannon / x} GrandChild (y, x)   Father (x, z)   Father (z, y) {Bob / y, Tom / z}

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Progol Problem: Searching Resolution Space Results in Combinatorial Explosion Solution Approach –Reduce explosion by generating most specific acceptable h –Conduct general-to-specific search (cf. Find-G, CN2  Learn-One-Rule) Procedure –1. User specifies H by stating predicates, functions, and forms of arguments allowed for each –2. Progol uses sequential covering algorithm FOR each DO Find most specific hypothesis h i such that B  h i  x i |  f(x i ) Actually, considers only entailment within k steps –3. Conduct general-to-specific search bounded by specific hypothesis h i, choosing hypothesis with minimum description length

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Learning First-Order Rules: Numerical versus Symbolic Approaches Numerical Approaches –Method 1: learning classifiers and extracting rules Simultaneous covering: decision trees, ANNs NB: extraction methods may not be simple enumeration of model –Method 2: learning rules directly using numerical criteria Sequential covering algorithms and search Criteria: MDL (information gain), accuracy, m-estimate, other heuristic evaluation functions Symbolic Approaches –Invert forward inference (deduction) operators Resolution rule Propositional and first-order variants –Issues Need to control search Ability to tolerate noise (contradictions): paraconsistent reasoning

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Terminology Induction and Deduction –Induction: finding h such that   D. (B  D  x i ) |  f(x i ) –Inductive learning: B  background knowledge (inductive bias, etc.) –Developing inverse deduction operators Deduction: finding entailed logical statements F(A, B) = C where A  B |  C Inverse deduction: finding hypotheses O(B, D) = h where   D. (B  D  x i ) |  f(x i ) –Resolution rule: deductive inference rule (P  L,  L  R |  P  R) Propositional logic: boolean terms, connectives ( , , ,  ) First-order predicate calculus (FOPC): well-formed formulas (WFFs), aka clauses (defined over literals, connectives, implicit quantifiers) –Inverse entailment: inverse of resolution operator Inductive Logic Programming (ILP) –Cigol: ILP algorithm that uses inverse entailment –Progol: sequential covering (general-to-specific search) algorithm for ILP

Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Summary Points Induction as Inverse of Deduction –Problem of induction revisited Definition of induction Inductive learning as specific case Role of induction, deduction in automated reasoning –Operators for automated deductive inference Resolution rule (and operator) for deduction First-order predicate calculus (FOPC) and resolution theorem proving –Inverting resolution Propositional case First-order case (inverse entailment operator) Inductive Logic Programming (ILP) –Cigol: inverse entailment (very susceptible to combinatorial explosion) –Progol: sequential covering, general-to-specific search using inverse entailment Next Week: Knowledge Discovery in Databases (KDD), Final Review