Explanation-Based Learning (borrowed from mooney et al)

Slides:



Advertisements
Similar presentations
Analytical Learning.
Advertisements

Applications of one-class classification
 2002, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci Deductive.
Knowledge Representation and Reasoning Learning Sets of Rules and Analytical Learning Harris Georgiou – 4.
Combining Inductive and Analytical Learning Ch 12. in Machine Learning Tom M. Mitchell 고려대학교 자연어처리 연구실 한 경 수
Knowledge in Learning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 19 Spring 2004.
처음 페이지로 이동 Chapter 11: Analytical Learning Inductive learning training examples n Analytical learning prior knowledge + deductive reasoning n Explanation.
Learning control knowledge and case-based planning Jim Blythe, with additional slides from presentations by Manuela Veloso.
Università di Milano-Bicocca Laurea Magistrale in Informatica
Knowledge in Learning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 19 Spring 2005.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Relational Data Mining in Finance Haonan Zhang CFWin /04/2003.
Chapter Two SCIENTIFIC METHODS IN BUSINESS
Chapter 1 Conducting & Reading Research Baumgartner et al Chapter 1 Nature and Purpose of Research.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Scientific Thinking - 1 A. It is not what the man of science believes that distinguishes him, but how and why he believes it. B. A hypothesis is scientific.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Introduction to Social Science Research
17.5 Rule Learning Given the importance of rule-based systems and the human effort that is required to elicit good rules from experts, it is natural to.
Inductive Logic Programming Includes slides by Luis Tari CS7741L16ILP.
1 Machine Learning: Lecture 11 Analytical Learning / Explanation-Based Learning (Based on Chapter 11 of Mitchell, T., Machine Learning, 1997)
Machine Learning Chapter 11. Analytical Learning
1 Abduction and Induction in Scientific Knowledge Development Peter Flach, Antonis Kakas & Oliver Ray AIAI Workshop 2006 ECAI August, 2006.
Representing variables according to the ISO/IEC standard.
CS Learning Rules1 Learning Sets of Rules. CS Learning Rules2 Learning Rules If (Color = Red) and (Shape = round) then Class is A If (Color.
CS Learning Rules1 Learning Sets of Rules. CS Learning Rules2 Learning Rules If (Color = Red) and (Shape = round) then Class is A If (Color.
ECE 8443 – Pattern Recognition Objectives: Error Bounds Complexity Theory PAC Learning PAC Bound Margin Classifiers Resources: D.M.: Simplified PAC-Bayes.
Inductive Generalizations Induction is the basis for our commonsense beliefs about the world. In the most general sense, inductive reasoning, is that in.
Scientific Inquiry.
Machine Learning Chapter 2. Concept Learning and The General-to-specific Ordering Tom M. Mitchell.
Machine Learning Chapter 5. Artificial IntelligenceChapter 52 Learning 1. Rote learning rote( โรท ) n. วิถีทาง, ทางเดิน, วิธีการตามปกติ, (by rote จากความทรงจำ.
Outline Inductive bias General-to specific ordering of hypotheses
Overview Concept Learning Representation Inductive Learning Hypothesis
Section 2 Scientific Methods Chapter 1 Bellringer Complete these two tasks: 1. Describe an advertisement that cites research results. 2. Answer this question:
CS Machine Learning 15 Jan Inductive Classification.
ARTIFICIAL INTELLIGENCE [INTELLIGENT AGENTS PARADIGM] Professor Janis Grundspenkis Riga Technical University Faculty of Computer Science and Information.
KU NLP Machine Learning1 Ch 9. Machine Learning: Symbol- based  9.0 Introduction  9.1 A Framework for Symbol-Based Learning  9.2 Version Space Search.
Explanation Based Learning (EBL) By M. Muztaba Fuad.
For Monday Finish chapter 19 No homework. Program 4 Any questions?
Generic Tasks by Ihab M. Amer Graduate Student Computer Science Dept. AUC, Cairo, Egypt.
Introduction to Scientific Research. Science Vs. Belief Belief is knowing something without needing evidence. Eg. The Jewish, Islamic and Christian belief.
For Monday Finish chapter 19 Take-home exam due. Program 4 Any questions?
CS Inductive Bias1 Inductive Bias: How to generalize on novel data.
Machine Learning Concept Learning General-to Specific Ordering
CpSc 810: Machine Learning Analytical learning. 2 Copy Right Notice Most slides in this presentation are adopted from slides of text book and various.
Copyright © by Holt, Rinehart and Winston. All rights reserved. Resources Chapter menu Section 2 Scientific Methods Chapter 1 Bellringer Complete these.
Graduate School for Social Research Autumn 2015 Research Methodology and Methods of Social Inquiry socialinquiry.wordpress.com Causality.
Introduction to Research
CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 1 Combining Inductive and Analytical Learning Why combine inductive and analytical learning?
More Symbolic Learning CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
Concept Learning and The General-To Specific Ordering
Chapter 1: Section 1 What is Science?. What Science IS and IS NOT.. The goal of Science is to investigate and understand the natural world, to explain.
Scientific Method Biology Image from:
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part One INTRODUCTION TO BUSINESS RESEARCH.
Writing a sound proposal
CS 9633 Machine Learning Explanation Based Learning
Introduction to Research Methodology
Part One INTRODUCTION TO BUSINESS RESEARCH
CS 9633 Machine Learning Concept Learning
CS 9633 Machine Learning Inductive-Analytical Methods
Research & Writing in CJ
Introduction to Research Methodology
THE NATURE OF SCIENCE Essential Questions
Introduction to Research Methodology
Scientific Method Integrated Sciences.
Knowledge in Learning Chapter 19
BBA V SEMESTER (BBA 502) DR. TABASSUM ALI
Machine Learning Chapter 2
Version Space Machine Learning Fall 2018.
Machine Learning Chapter 2
Presentation transcript:

Explanation-Based Learning (borrowed from mooney et al)

Explanation-Based Learning (EBL) One definition: Learning general problem-solving techniques by observing and analyzing solutions to specific problems.

SBL (vs. EBL) lots of data (examples) Similarity-based learning (SBL) are inductive: –generalizes from training data –empirically identifies patterns that distinguish between positive and negative examples of a target concept. Inductive results are justified empirically (e.g., by statistical arguments such as those used in establishing theoretical results in PAC learning). Generally requires significant numbers of training examples in order to produce statistically justified conclusions. Generally does not require or exploit background knowledge.

EBL (vs. SBL) lots of knowledge Explanation-based learning (EBL) is (usually) deductive: –uses prior knowledge to explain each training example –Explanation identifies what properties are relevant to the target function and which are irrelevant. Prior knowledge is used to reduce the hypothesis space and focus the learner on hypotheses that are consistent with prior knowledge about the target concept. Accurate learning is possible from very few (0) training examples (typically 1 example per learned rule).

The EBL Hypothesis By understanding why an example is a member of a target concept, one can learn the essential properties of the concept Trade-off the need to collect many examples for the ability to explain single examples (via a domain theory) This assumes the domain theory is competent: –Correct: does not entail that any negative example is positive –Complete: each positive example can be explained –Tractable: an explanation can be found for each positive example.

SBL vs. EBL entailment constraints SBL: Hypothesis & Descriptions Classifications Hypothesis is selected from restricted hypothesis space. EBL: Hypothesis & Descriptions Classifications Background Hypothesis

EBL Task In addition to a set of training examples, EBL also takes as input a domain theory, background knowledge about the target concept that is usually specified as a set of logical rules (Horn clauses) and operationality criteria. The goal is to find an efficient or operational definition of the target concept that is consistent with both the domain theory and the training examples.

EBL Task: operationality observable vs. unobservable Operationality is often imposed by restricting the hypothesis space to using only certain predicates (e.g., those that are directly used to describe the examples). Observable: predicates used to describe examples Unobservable: the target concept In classical EBL the learned definition is –logically entailed by the domain theory –a more efficient definition of the target concept –requires only look-up (pattern matching) using observable predicates rather than search (logical inference) mapping observables to unobservables.

EBL Task Given: Goal concept Training example Domain Theory Operationality Criteria Find: a generalization of the training example that is a sufficient criteria for the target concept and satisfies the operationality criteria

EBL Example Goal concept: SafeToStack(x,y) Training Examples: One example SafeToStack (Obj1,Obj2) On(Obj1,Obj2) Owner(Obj1,Molly) Type(Obj1,Box) Owner(Obj2, Muffet) Type(Obj2,Endtable) Fragile(Obj2) Color(Obj1,Red) Material(Obj1,Cardboard) Color(Obj2,Blue) Material(Obj2,Wood) Volume(Obj1, 0.1) Density(Obj1,0.1)

EBL Example Domain Theory: SafeToStack(x,y) :- not(Fragile(y)). SafeToStack(x,y) :- Lighter(x,y). Lighter(x,y) :- Weight(x,wx), Weight(y,wy), wx < wy. Weight(x,w) :- Volume(x,v), Density(x,d), w=v*d. Weight(x,5) :- Type(x,Endtable). Fragile(x) :- Material(x,Glass). Opertional predicates: Type, Color, Volume, Owner, Fragile, Material, Density, On,, =.

EBL Method For each positive example not correctly covered by an operational rule do: 1.Explain: Use the domain theory to construct a logical proof that the example is a member of the concept. 2.Analyze: Generalize the explanation to determine a rule that logically follows from the domain theory given the structure of the proof and is operational. Add the new rule to the concept definition.

EBL Example Training Example: SafeToStack (Obj1,Obj2) Type(Obj2,Endtable) Volume(Obj1, 0.1) Density(Obj1,0.1) … Domain Theory: SafeToStack(x,y) :- Lighter(x,y). Lighter(x,y) :- Weight(x,wx), Weight(y,wy), wx < wy. Weight(x,w) :- Volume(x,v), Density(x,d), w=v*d. Weight(x,5) :- Type(x,Endtable). …

Example Explanation (Proof) Volume(Obj1,2) SafeToStack(Obj1,Obj2) Lighter(Obj1,Obj2) Weight(Obj1,0.6)Weight(Obj2,5) Type(Obj2.Endtable) 0.6=2* <5 Density(Obj1,0.3)

Generalization Find the weakest preconditions A for a conclusion C such that A entails C using the given proof. The general target predicate is regressed through each rule used in the proof to produce generalized conditions at the leaves. To regress a set of literals P through a rule H :- B1,...Bn (B={B1,...Bn}) using literal L element of P Let Ф be the most general unifier of L and H apply the resulting substitution to all the literals in P and B and return: P = (PФ - LФ) U BФ Also apply the substitution to update the conclusion: C=CФ After regressing the general target concept through each rule used in the proof return: C :- P1,...Pn (P={P1...Pn})

Generalization Example Regress {SafeToStack(x,y)} through SafeToStack(x1,y1) :- Lighter(x1,y1). Unifier: Ф = {x/x1, y/y1} Result: {Lighter(x,y)} Weight(Obj2,5) Lighter(Obj1,Obj2) Weight(Obj1,0.6) 06.<5

Generalization Example Regress {Lighter(x,y)} through Lighter(x2,y2) :- Weight(x2,wx2), Weight(y2,wy2), wx2 < wy2. Unifier: Ф = {x/x2, y/y2} Result:{Weight(x,wx), Weight(y,wy), wx < wy} Weight(Obj1,0.6)Weight(Obj2,5)

Generalization Example Regress {Weight(x,wx), Weight(y,wy), wx < wy} through Weight(x3,w) :- Volume(x3,v), Density(x3,d), w=v*d. Unifeir: Ф = {x/x3, wx/w} Result: {Volume(x,v), Density(x,d), wx=v*d, Weight(y,wy), wx < wy} Weight(Obj2,5)

Generalization Example Regress {… Weight(y,wy) …} through Weight(x4,5) :- Type(x4,Endtable). Unifier: Ф = {y/x4, 5/wy} Result: {Volume(x,v), Density(x,d), wx=v*d, Type(y,Endtable), wx < 5} Learned Rule: SafeToStack(x,y) :- Volume(x,v), Density(x,d), wx=v*d, Type(y,Endtable), wx < 5.

Re Generalization Simply substituting variables for constants in the proof will not work because: –Some constants (Endtable,5) may come from the domain theory and cannot be generalized and maintain soundness. –Two instances of the same constant may or may not generalize to the same variable depending on structure of the proof (e.g. assume both the weight and density happened to be the same in the example, but they clearly dont have to be the same in general). Since generalization is basically performing a set of unifications and substitutions and these operations have linear time complexity, generalization is a quick, linear- time process.

Knowledge as Bias The hypotheses produced by EBL are obviously strongly biased by the domain theory it is given. Being able to alter the bias of a learning algorithm by supplying prior knowledge in declarative form (declarative bias) is very useful (e.g., by adding new rules and predicates). EBL assumes a complete and correct domain theory, but theory refinement and other methods can be biased by incomplete and incorrect domain theories.

Perspectives on EBL EBL as theory guided generalization of examples: Explanations are used to distinguish relevant from irrelevant features. EBL as example guided reformulation of theories: Examples are used to focus on which operational concept reformulations to learn are typical EBL as knowledge compilation: Deductive consequences that are particularly useful (e.g., for reasoning about the training examples) are compiled out to subsequently allow for more efficient reasoning.

Standard Approach to EBL

Knowledge-Level Learning (Newell, Dietterich) Knowledge closure all things that can be inferred from a collection of rules and facts Pure EBL only learns how to solve faster, not how to solve problems previously insoluble. Inductive learners make inductive leaps and hence can solve more after learning. EBL is often called Speed-up learning (not knowledge-level learning) What about considering resource-limits (e.g., time) on problem solving?

Utility of Knowledge Compilation Deductive reasoning is difficult and frequently similar conclusions must be derived repeatedly. Some domains have complete and correct theories and learning involves deriving useful consequences that make reasoning more efficient, e.g. chess, mathematics, etc.

Utility of Knowledge Compilation Different types of knowledge compilation: –Static: Not example-based, reformulate KB up front to make it more efficient for general inferences of a particular type. –Dynamic: Uses examples, perhaps, incrementally, to tune a system to improve efficiency on a particular distribution of problems. Dynamic systems like EBL make the inductive assumption that improving performance on a set of training cases will generalize to improved performance on subsequent test cases.

Utility Problem After learning many macro-operators, macro-rules, or search control rules, the time to match and search through this added knowledge may start to outweigh its benefits (Minton 1988) A learned rule must be useful in solving new problems frequently enough and save enough processing time in order to compensate for the time need to attempt to match it every time. Utility = (AvgSavings x ApplicFreq) - AvgMatchCost EBL methods can frequently result in learning a set of rules with negative overall utility resulting in slowdown rather than the intended speedup.

Addressing the Utility Problem Improve Efficiency of Matching: Preprocess learned rules to improve their match effiicency. Restrict Expressiveness: Prevent learning of rules with combinatorial match costs. Selective Acquisition: Only learn rules whose expected benefit outweighs their cost. Selective Retention: Dynamically forget expensive rules that are rarely used. Selective Utilization: Restrict the use of learned rules to avoid undue cost of application.

Imperfect Theories and EBL Incomplete Theory Problem Cannot build explanations of specific problems because of missing knowledge Intractable Theory Problem Have enough knowledge, but not enough computer time to build specific explanation Inconsistent Theory Problem Can derive inconsistent results from a theory (e.g., because of default rules)

Applications Planning (macro operators in STRIPS) Mathematics (search control in LEX)

Planning with Macro-Operators AI planning using Strips operators is search intensive. People seem to utilize canned plans to achieve everyday goals. Such pre-packaged planning sequences (macro- operators) can be learned by generalizing specific constructed or observed plans. Method is analogous to composing Horn-clause rules by generalizing proofs. A problem is solved by first trying to use learning macro- operators, falling back on general planning as a last resort.

STRIPS Original planning system which used means-ends analysis and theorem proving in robot planning Sample actions: GoThru(A,D,R1,R2) Preconditions: In(A,R1), Connects(D,R1,R2) Effects: In(A,R2), In(A,R1) PushThru(A,O,D,R1,R2) Preconditions: In(A,R1), In(O,R1) Connects(D,R1,R2) Effects: In(A,R2), In(O,R2),In(A,R1), In(O,R1)

STRIPS Sample Problem: State: In(r,room1), In(box,room2), Connects(d1,room1,room2), Connects(d2,room2,room3) Goal: In(box,room1) Sample Solution: GoThru(r,d1,room1,room2) PushThru(r,box,d1,room2,room1)

Learned Macro-Operator EBL generalizing this plan produces the following macro-operator: GoThruPushThru(A,D1,R1,R2,O,D2,R3) Preconditions: InRoom(A,R1), InRoom(O,R2), Connects(D1,R1,R2), Connects(D2,R2,R3), (A=O & R1=R2) Effects: InRoom(O,R3), InRoom(A,R3), InRoom(A,R2), InRoom(O,R2), (R3=R1) InRoom(A,R1) Extra preconditions needed to prevent precondition clobbering during execution of generalized plan. Conditional effects come from possible deletions in the generalized plan.

Representing Plan MACROPS Strips actually used a triangle table to implicitly store macros for every subsequence of the actions in the plan. Plan: [State] OP1 OP2 OP3 OP4 OP5 [Goal] Op1 Op1 Op2 Op1 Op2 Op3 Op1 Op2 Op3 Op4 Op1 Op2 Op3 Op4 Op5 The triangle table supports treating any of the 10 subsequence of the generalized plan as a macrop in future problems.

Experimental Results Planning time with and without learning (min:sec) trial No learn 3:05 9:42 7:0314:09 -- learning 3:05 3:54 6:34 4:37 9:13

Learning Search Control Search control rules are used to select operators during search. IF the state is of the form r f(x) dx, THEN apply the operator MoveConstantOutsideIntegral Such search control rules can be learned by explaining how the application of an operator in a sample problem led to a solution: 3sin(x)dx 3 sin(x)dx 3 cos(x) Positive examples of when to apply an operator are states in which applying that operator leads to a solution, negative examples are states in which applying the operator leads away from the solution (i.e. another operator leads to the solution). Induction and combinations of explanation and induction can also be used to learn search control rules.

EBL variations Generalizing to N: handling recursive rules in proofs Knowledge Deepening: explaining shallow rules Explanation-based induction and abductive generalization

Generalizing to N (Shavlik, BAGGER2) Handling recursive or iterative concepts (recursive rules in proofs). goal P PP PP Learned rules: Goal P & gen-2 P gen-3 V gen-5 V gen-6 V recursive-gen-1 V recursive-gen-2

Knowledge Deepening When two proofs, A and B, exist for a proposition, and proof A involves a single (shallow) rule, PQ, and the weakest preconditions of proof B is equivalent to P, then proof B explains rule PQ. Shallow rule: leaves are green Explanation: leaves are green because they contain mesophylls, which contain chlorophyll, which is a green pigment.

Knowledge Deepening (leaf ?x)(green ?x) Part(?x ?y) & (isa ?y Mesophyll) (green ?y) Part(?y ?z) & (isa ?z Chrorophyll) (green ?z) The weakest preconditions of both proofs are the same: (leaf ?x) Use the more complicated proof to explain the shallow rule.

Explanation-Based Induction Teleology: function suggests structure Identify a teleologic explanation Structural properties supporting physiological goal: leaf dehydration is avoided by the cutilcle covering the leafs epidermis Identify the weakest preconditions of the explanation. Separate into –Structural preconditions: epidermis covered by cuticle –Qualifying preconditions: performs transpiration Find other organs satisfying the qualifying conditions: stems, flowers, fruit. Hypothesize they also have the structural conditions: are the epidermises of stems, flowers, and fruit also covered by a cuticle?