Learning from Human Teachers: Issues and Challenges for ILP in Bootstrap Learning Sriraam Natarajan 1, Gautam Kunapuli 1, Richard Maclin 3, David Page.

Slides:



Advertisements
Similar presentations
Artificial Intelligence
Advertisements

Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Modelling with expert systems. Expert systems Modelling with expert systems Coaching modelling with expert systems Advantages and limitations of modelling.
Main Idea and Supporting Details
Session 2: Informational Text Audience: Science, Social Studies, Technical Subject Teachers.
Logic Use mathematical deduction to derive new knowledge.
Combining Inductive and Analytical Learning Ch 12. in Machine Learning Tom M. Mitchell 고려대학교 자연어처리 연구실 한 경 수
Task Analysis EDU 553 – Principles of Instructional Design Dr. Steve Broskoske.
Intelligent Agent for Designing Steel Skeleton Structures of Tall Buildings Zbigniew Skolicki Rafal Kicinger.
Session 1 Getting started with classroom research DAVID NUNAN.
CSE (c) S. Tanimoto, 2008 Propositional Logic
Cognitive Science 1 Kartik Talamadupula Subbarao Kambhampati J. Benton Dept. of Computer Science Arizona State University Paul Schermerhorn Matthias Scheutz.
Constraint Logic Programming Ryan Kinworthy. Overview Introduction Logic Programming LP as a constraint programming language Constraint Logic Programming.
NaLIX: A Generic Natural Language Search Environment for XML Data Presented by: Erik Mathisen 02/12/2008.
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 21 Jim Martin.
Relational Data Mining in Finance Haonan Zhang CFWin /04/2003.
Statistical Relational Learning for Link Prediction Alexandrin Popescul and Lyle H. Unger Presented by Ron Bjarnason 11 November 2003.
Machine Learning: Symbol-Based
Writing tips Based on Michael Kremer’s “Checklist”,
Let remember from the previous lesson what is Knowledge representation
CS 330 Programming Languages 09 / 16 / 2008 Instructor: Michael Eckmann.
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 22 Jim Martin.
© Jesse Davis 2006 View Learning Extended: Learning New Tables Jesse Davis 1, Elizabeth Burnside 1, David Page 1, Vítor Santos Costa 2 1 University of.
WISCONSIN UNIVERSITY OF WISCONSIN - MADISON Integrating Knowledge Capture and Supervised Learning through a Human-Computer Interface Trevor Walker, Gautam.
ADL Slide 1 December 15, 2009 Evidence-Centered Design and Cisco’s Packet Tracer Simulation-Based Assessment Robert J. Mislevy Professor, Measurement &
Semantic Web Technologies Lecture # 2 Faculty of Computer Science, IBA.
1/24 Learning to Extract Genic Interactions Using Gleaner LLL05 Workshop, 7 August 2005 ICML 2005, Bonn, Germany Mark Goadrich, Louis Oliphant and Jude.
Learning Objectives. Objectives Objectives: By the conclusion to this session each participant should be able to… Differentiate between a goal and objectives.
Issues with Data Mining
What Our Students Need Most The 7 Fundamental Conditions of Learning.
Chapter 11 Helping Students Construct Usable Knowledge.
Author: James Allen, Nathanael Chambers, etc. By: Rex, Linger, Xiaoyi Nov. 23, 2009.
Effective Teaching ELED Journal Topic One Chapter One, page 14 –Portfolio Activity.
© 2007 Tom Beckman Features:  Are autonomous software entities that act as a user’s assistant to perform discrete tasks, simplifying or completely automating.
Today’s Topics Chapter 2 in One Slide Chapter 18: Machine Learning (ML) Creating an ML Dataset –“Fixed-length feature vectors” –Relational/graph-based.
Theory Revision Chris Murphy. The Problem Sometimes we: – Have theories for existing data that do not match new data – Do not want to repeat learning.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Friday, February 4, 2000 Lijun.
Skill Acquisition via Transfer Learning and Advice Taking Lisa Torrey, Jude Shavlik, Trevor Walker University of Wisconsin-Madison, USA Richard Maclin.
Transfer Learning Via Advice Taking Jude Shavlik University of Wisconsin-Madison.
WISCONSIN UNIVERSITY OF WISCONSIN - MADISON Broadening the Applicability of Relational Learning Trevor Walker Ph.D. Defense 1.
Transfer in Reinforcement Learning via Markov Logic Networks Lisa Torrey, Jude Shavlik, Sriraam Natarajan, Pavan Kuppili, Trevor Walker University of Wisconsin-Madison,
Noboru Matsuda Human-Computer Interaction Institute
Speeding Up Relational Data Mining by Learning to Estimate Candidate Hypothesis Scores Frank DiMaio and Jude Shavlik UW-Madison Computer Sciences ICDM.
GRASP: Designing Objects with Responsibilities
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Relational Macros for Transfer in Reinforcement Learning Lisa Torrey, Jude Shavlik, Trevor Walker University of Wisconsin-Madison, USA Richard Maclin University.
EFFECTIVE WRITING 8 Readability. Writing - time and resource consuming, stressful process Texts have a strong tendency of using more complex, more sophisticated.
Advice Taking and Transfer Learning: Naturally-Inspired Extensions to Reinforcement Learning Lisa Torrey, Trevor Walker, Richard Maclin*, Jude Shavlik.
Summative vs. Formative Assessment. What Is Formative Assessment? Formative assessment is a systematic process to continuously gather evidence about learning.
KU NLP Machine Learning1 Ch 9. Machine Learning: Symbol- based  9.0 Introduction  9.1 A Framework for Symbol-Based Learning  9.2 Version Space Search.
For Monday Finish chapter 19 No homework. Program 4 Any questions?
© 2008 SRI International Question Asking to Inform Preference Learning: A Case Study Melinda Gervasio SRI International Karen Myers SRI International Marie.
Automated Reasoning Early AI explored how to automated several reasoning tasks – these were solved by what we might call weak problem solving methods as.
For Monday Finish chapter 19 Take-home exam due. Program 4 Any questions?
CS 5751 Machine Learning Chapter 10 Learning Sets of Rules1 Learning Sets of Rules Sequential covering algorithms FOIL Induction as the inverse of deduction.
More Symbolic Learning CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
Chap. 10 Learning Sets of Rules 박성배 서울대학교 컴퓨터공학과.
Brief Information and Strategies for Instructors.
Some Thoughts to Consider 5 Take a look at some of the sophisticated toys being offered in stores, in catalogs, or in Sunday newspaper ads. Which ones.
Feature Generation and Selection in SRL Alexandrin Popescul & Lyle H. Ungar Presented By Stef Schoenmackers.
Investigate Plan Design Create Evaluate (Test it to objective evaluation at each stage of the design cycle) state – describe - explain the problem some.
CSE573 Autumn /09/98 Machine Learning Administrative –Last topic: Decision Tree Learning Reading: 5.1, 5.4 Last time –finished NLP sample system’s.
Frank DiMaio and Jude Shavlik Computer Sciences Department
Knowledge Representation and Reasoning
Trevor Walker, Gautam Kunapuli, Noah Larsen, David Page, Jude Shavlik
Richard Maclin University of Minnesota - Duluth
Back to “Serious” Topics…
This Lecture Substitution model
Refining Rules Incorporated into Knowledge-Based Support Vector Learners via Successive Linear Programming Richard Maclin University of Minnesota - Duluth.
Implementation of Learning Systems
Presentation transcript:

Learning from Human Teachers: Issues and Challenges for ILP in Bootstrap Learning Sriraam Natarajan 1, Gautam Kunapuli 1, Richard Maclin 3, David Page 1, Ciaran O’Reilly 2, Trevor Walker 1 and Jude Shavlik 1 2 SRI International 1 University of Wisconsin-Madison 3 University of Minnesota, Duluth

Outline Bootstrap Learning (BL), Inductive Logic Programming (ILP) and Challenges The Wisconsin Relevance Interpreter: Incorporating teacher advice into an ILP agent The Onion: Automating agent’s ILP runs across tasks and domains without expert intervention Conclusions

Outline Bootstrap Learning (BL), Inductive Logic Programming (ILP) and Challenges The Wisconsin Relevance Interpreter: Incorporating teacher advice into an ILP agent The Onion: Automating agent’s ILP runs across tasks and domains without expert intervention Conclusions

Bootstrap Learning New learning paradigm by Oblinger (2006) – proposes framework to create an electronic student that learns from natural instruction – views learning as knowledge acquisition from a human teacher Image Source: Dan Oblinger, AAAI 2006 Briefing

Bootstrap Learning electronic student (agent) assumes relevant knowledge possessed by teacher – domain background teaching through human-like natural instruction methods – pedagogical examples – telling of instructions – demonstration – feedback

Bootstrap Learning Learning depends only on instruction method, is domain- independent  Unmanned Aerial Vehicle (UAV)  Armored Task Force (ATF)  International Space Station (ISS) Student learns concepts through a ladder of lessons more complex concepts bootstrap on lower-rung lessons simpler concepts; learned first lesson ladder

Bootstrap Learning From Examples We focus on instruction by examples – including teacher hints about specific examples – teacher and student communicate via Interlingua Learn models using Inductive Logic Programming (ILP, Muggleton and de Raedt, 1994) – declarative representation of examples and learned rules makes it easier for teacher and student to communicate – concepts learned earlier can be bootstrapped easily We have developed a Java-based ILP system called WILL: the Wisconsin Inductive Logic Learner

Bootstrap Learning From Examples We focus on instruction by examples – including teacher hints about specific examples – teacher and student communicate via Interlingua Learn models using Inductive Logic Programming (ILP, Muggleton and De Raedt, 1994) – declarative representation of examples and learned rules makes it easier for teacher and student to communicate – concepts learned earlier can be bootstrapped easily We have developed a Java-based ILP system called WILL: the Wisconsin Inductive Logic Learner

Bootstrap Learning From Examples We focus on instruction by examples – including teacher hints about specific examples – teacher and student communicate via Interlingua Learn models using Inductive Logic Programming (ILP, Muggleton and De Raedt, 1994) – declarative representation of examples and learned rules makes it easier for teacher and student to communicate – concepts learned earlier can be bootstrapped easily We have developed a Java-based ILP system called WILL: the Wisconsin Inductive Logic Learner

Recognize “interesting scenario” tasks UAV Domain with ByExample Lessons FullFuelTankAssessGoal Near TruckAtIntersection ReadyToFly IsStopped TruckScenario IsMoving TruckScenario ComputeStoppedTrucksIntComputeMovingTrucksInt ComputeScenarioInterestingness Fly and land UAV tasks Surveillance and Camera Tasks Goal: Train an Unmanned Aerial Vehicle agent to perform Intelligent Surveillance and Reconnaissance (ISR)

Natural Instruction of the Agent lower rung task: FullFuelTank (determine if fuel tank is full) u: LessonTeaches(WhenTrue(Full)) u: Method(byExample) p: UAV(name=uav1, altitude=4.0, fuelCapacity=500, currentFuel=400) u: GestureAt(uav1) u: Not(Full(this)) u: Relevant(Of(fuelCapacity, this)) u: Relevant(Of(currentFuel, this))... p: UAV(name=uav2, altitude=0.0, fuelCapacity=300, currentFuel=300) i: Answer(response=true) i: Grade(grade=100.0) Teacher utters information about concept being taught and the natural instruction method

Natural Instruction of the Agent lower rung task: FullFuelTank (determine if fuel tank is full) u: LessonTeaches(WhenTrue(Full)) u: Method(byExample) p: UAV(name=uav1, altitude=4.0, fuelCapacity=500, currentFuel=400) u: GestureAt(uav1) u: Not(Full(this)) u: Relevant(Of(fuelCapacity, this)) u: Relevant(Of(currentFuel, this))... p: UAV(name=uav2, altitude=0.0, fuelCapacity=300, currentFuel=300) i: Answer(response=true) i: Grade(grade=100.0) Perceptions provide information about the current state

Natural Instruction of the Agent lower rung task: FullFuelTank (determine if fuel tank is full) u: LessonTeaches(WhenTrue(Full)) u: Method(byExample) p: UAV(name=uav1, altitude=4.0, fuelCapacity=500, currentFuel=400) u: GestureAt(uav1) u: Not(Full(this)) u: Relevant(Of(fuelCapacity, this)) u: Relevant(Of(currentFuel, this))... p: UAV(name=uav2, altitude=0.0, fuelCapacity=300, currentFuel=300) i: Answer(response=true) i: Grade(grade=100.0) The teacher can gesture at perceived objects. They will be designated “this” by default.

Natural Instruction of the Agent lower rung task: FullFuelTank (determine if fuel tank is full) u: LessonTeaches(WhenTrue(Full)) u: Method(byExample) p: UAV(name=uav1, altitude=4.0, fuelCapacity=500, currentFuel=400) u: GestureAt(uav1) u: Not(Full(this)) u: Relevant(Of(fuelCapacity, this)) u: Relevant(Of(currentFuel, this))... p: UAV(name=uav2, altitude=0.0, fuelCapacity=300, currentFuel=300) i: Answer(response=true) i: Grade(grade=100.0) Teacher gives label information, identifying this as a NEGATIVE example

Natural Instruction of the Agent lower rung task: FullFuelTank (determine if fuel tank is full) u: LessonTeaches(WhenTrue(Full)) u: Method(byExample)) p: UAV(name=uav1, altitude=4.0, fuelCapacity=500, currentFuel=400) u: GestureAt(uav1) u: Not(Full(this)) u: Relevant(Of(fuelCapacity, this)) u: Relevant(Of(currentFuel, this))... p: UAV(name=uav2, altitude=0.0, fuelCapacity=300, currentFuel=300) i: Answer(response=true) i: Grade(grade=100.0) Teacher gives additional “hints” about why this is a positive or negative example. This is the relevance information.

Natural Instruction of the Agent lower rung task: FullFuelTank (determine if fuel tank is full) u: LessonTeaches(WhenTrue(Full)) u: Method(byExample) p: UAV(name=uav1, altitude=4.0, fuelCapacity=500, currentFuel=400) u: GestureAt(uav1) u: Not(Full(this)) u: Relevant(Of(fuelCapacity, this)) u: Relevant(Of(currentFuel, this))... p: UAV(name=uav2, altitude=0.0, fuelCapacity=300, currentFuel=300) i: Answer(response=true) i: Grade(grade=100.0) Some more examples follow

Natural Instruction of the Agent lower rung task: FullFuelTank (determine if fuel tank is full) u: LessonTeaches(WhenTrue(Full)) u: Method(byExample) p: UAV(name=uav1, altitude=4.0, fuelCapacity=500, currentFuel=400) u: GestureAt(uav1) u: Not(Full(this)) u: Relevant(Of(fuelCapacity, this)) u: Relevant(Of(currentFuel, this))... p: UAV(name=uav2, altitude=0.0, fuelCapacity=300, currentFuel=300) i: Answer(response=true) i: Grade(grade=100.0) Teacher tests the student and gives a grade

Representation for ILP An ILP system learns a logic program given – background knowledge as a set of first-order logic (FOL) formulae – examples expressed as facts represented in logic terms are objects in the world – constants: uav1, uav2 – variables: ?uav – functions: CurrentFuel(?uav, ?currentFuel) literals are truth-valued and represent relations among objects –sameAs(?fuelCapacity, ?currentFuel) literals can be combined into compound sentences using logical connectives AND, OR and NOT. We translate from Interlingua to an ILP representation creating background, facts and examples from messages

Representation for ILP An ILP system learns a logic program given – background knowledge as a set of first-order logic (FOL) formulae – examples expressed as facts represented in logic terms are objects in the world – constants: uav1, uav2 – variables: ?uav – functions: CurrentFuel(?uav, ?currentFuel) literals are truth-valued and represent relations among objects –sameAs(?fuelCapacity, ?currentFuel) literals can be combined into compound sentences using logical connectives AND, OR and NOT. We translate from Interlingua to an ILP representation creating background, facts and examples from messages

Representation for ILP An ILP system learns a logic program given – background knowledge as a set of first-order logic (FOL) formulae – examples expressed as facts represented in logic terms are objects in the world – constants: uav1, uav2 – variables: ?uav – functions: CurrentFuel(?uav, ?currentFuel) literals are truth-valued and represent relations among objects –sameAs(?fuelCapacity, ?currentFuel) literals can be combined into compound sentences using logical connectives AND, OR and NOT. We translate from Interlingua to an ILP representation creating background, facts and examples from messages

Representation for ILP An ILP system learns a logic program given – background knowledge as a set of first-order logic (FOL) formulae – examples expressed as facts represented in logic terms are objects in the world – constants: uav1, uav2 – variables: ?uav – functions: CurrentFuel(?uav, ?currentFuel) literals are truth-valued and represent relations among objects –sameAs(?fuelCapacity, ?currentFuel) literals can be combined into compound sentences using logical connectives AND, OR and NOT. We translate from Interlingua to an ILP representation creating background, facts and examples from messages

Representation for ILP An ILP system learns a logic program given – background knowledge as a set of first-order logic (FOL) formulae – examples expressed as facts represented in logic terms are objects in the world – constants: uav1, uav2 – variables: ?uav – functions: CurrentFuel(?uav, ?currentFuel) literals are truth-valued and represent relations among objects –sameAs(?fuelCapacity, ?currentFuel) literals can be combined into compound sentences using logical connectives AND, OR and NOT. We translate from Interlingua to an ILP representation creating background, facts and examples from messages

Learning in ILP Features: A..Z Target Concept to be learned: A,Z,R,W / Target Have background, examples and facts for learning ILP search adds one literal at a time n predicates leads to O(n!) combos ILP requires large number of examples Hypothesis space True / Target A,Z / TargetA,B / Target A / Target A, Z, R, W / Target A, Z, R / Target...

Learning in ILP with Relevance Teacher tells agent that p redicates A, Z, R are relevant to the concept being learned amount of search is greatly reduced ILP with relevance can succeed with fewer examples Hypothesis space True / Target A,Z / TargetA,B / Target A / Target A, Z, R, W / Target A, Z, R / Target... Features: A..Z Target Concept to be learned: A,Z,R,W / Target

BL/ILP Challenges Relevant statements specify partial theory – how can we use relevance effectively in ILP? No ILP expert in the loop for param selection – agent should be trainable by non-ILP experts – agent should be able to use ILP over different lessons in the same domain AND across different domains – cannot exhaustively try all possible combinations Teacher is agnostic towards learning mechanism – examples may contain a very large number of features – #examples very low (< 10 in most lessons) – some lessons have no negative examples

BL/ILP Challenges Relevant statements specify partial theory – how can we use relevance effectively in ILP? No ILP expert in the loop for param selection – agent should be trainable by non-ILP experts – agent should be able to use ILP over different lessons in the same domain AND across different domains – cannot exhaustively try all possible combinations Teacher is agnostic towards learning mechanism – examples may contain a very large number of features – #examples very low (< 10 in most lessons) – some lessons have no negative examples

BL/ILP Challenges Relevant statements specify partial theory – how can we use relevance effectively in ILP? No ILP expert in the loop for param selection – agent should be trainable by non-ILP experts – agent should be able to use ILP over different lessons in the same domain AND across different domains – cannot exhaustively try all possible combinations Teacher is agnostic towards learning mechanism – examples may contain a very large number of features – #examples very low (< 10 in most lessons) – some lessons have no negative examples

BL/ILP Challenges Relevant statements specify partial theory – Develop a robust strategy to interpret and incorporate relevance ( Wisconsin Relevance Interpreter) No ILP expert in the loop for param selection – Develop a multi-layer strategy for automating ILP runs so agent can be easily trained, generalized across domains ( Onion) Teacher is agnostic towards learning mechanism – Develop novel approaches to handle non-standard cases e.g., generation of negative examples etc.

BL/ILP Challenges Relevant statements specify partial theory – Develop a robust strategy to interpret and incorporate relevance ( Wisconsin Relevance Interpreter) No ILP expert in the loop for param selection – Develop a multi-layer strategy for automating ILP runs so agent can be easily trained, generalized across domains ( Onion) Teacher is agnostic towards learning mechanism – Develop novel approaches to handle non-standard cases e.g., generation of negative examples etc.

BL/ILP Challenges Relevant statements specify partial theory – Develop a robust strategy to interpret and incorporate relevance ( Wisconsin Relevance Interpreter) No ILP expert in the loop for param selection – Develop a multi-layer strategy for automating ILP runs so agent can be easily trained, generalized across domains ( Onion) Teacher is agnostic towards learning mechanism – Develop novel approaches to handle non-standard cases e.g., generation of negative examples etc.

Outline Bootstrap Learning (BL), Inductive Logic Programming (ILP) and Challenges The Wisconsin Relevance Interpreter: Incorporating teacher advice into an ILP agent The Onion: Automating agent’s ILP runs across tasks and domains without expert intervention Conclusions

Handling Teacher Hints relevance can be specified at different “resolutions” – certain constants are relevant (e.g.,  ) – attribute (incl. type) is relevant (e. g., UAV.currentFuel) – object is relevant (e.g., Truck3) – relationship among attributes/objects is relevant (e.g., lessThan, greaterThan, sameAs) – previously learned concepts (bootstrapping) – partial theory (e.g., combinations of the above) teacher gives hints about specific examples – key challenge: what should be generalized (to a logical variable) and what should remain constant?

Handling Teacher Hints relevance can be specified at different “resolutions” – certain constants are relevant (e.g.,  ) – attribute (incl. type) is relevant (e. g., UAV.currentFuel) – object is relevant (e.g., Truck3) – relationship among attributes/objects is relevant (e.g., lessThan, greaterThan, sameAs) – previously learned concepts (bootstrapping) – partial theory (e.g., combinations of the above) teacher gives hints about specific examples – key challenge: what should be generalized (to a logical variable) and what should remain constant?

Handling Teacher Hints relevance can be specified at different “resolutions” – certain constants are relevant (e.g.,  ) – attribute (incl. type) is relevant (e. g., UAV.currentFuel) – object is relevant (e.g., Truck3) – relationship among attributes/objects is relevant (e.g., lessThan, greaterThan, sameAs) – previously learned concepts (bootstrapping) – partial theory (e.g., combinations of the above) teacher gives hints about specific examples – key challenge: what should be generalized (to a logical variable) and what should remain constant?

Handling Teacher Hints relevance can be specified at different “resolutions” – certain constants are relevant (e.g.,  ) – attribute (incl. type) is relevant (e. g., UAV.currentFuel) – object is relevant (e.g., Truck3) – relationship among attributes/objects is relevant (e.g., lessThan, greaterThan, sameAs) – previously learned concepts (bootstrapping) – partial theory (e.g., combinations of the above) teacher gives hints about specific examples – key challenge: what should be generalized (to a logical variable) and what should remain constant?

Handling Teacher Hints relevance can be specified at different “resolutions” – certain constants are relevant (e.g.,  ) – attribute (incl. type) is relevant (e. g., UAV.currentFuel) – object is relevant (e.g., Truck3) – relationship among attributes/objects is relevant (e.g., lessThan, greaterThan, sameAs) – previously learned concepts (bootstrapping) – partial theory (e.g., combinations of the above) teacher gives hints about specific examples – key challenge: what should be generalized (to a logical variable) and what should remain constant?

Handling Relevance: One Example identify “interesting” candidate constants isInteresting(Truck17) isInteresting(Stopped) identify “relevant” predicates for the target concept relevant: Truck.moveStatus relevant: Truck relevant: sameAs (these get lower cost during search) GestureAt(Truck(name=Truck17, latitude=-10, longitude=10, moveStatus=“Stopped”)) RelevantRelationship(InstanceOf(this, Truck)) RelevantRelationship(SameAs(this.moveStatus, “Stopped”))

construct rules for each relevance statement pred1(?T) IF Truck(?T) pred2(?T) IF moveStatus(?T, ?S), sameAs(?S, stopped) create combo rule from all the relevance for this example relevantFromPosEx1(?T) IF pred1(?T), pred2(?T) Handling Relevance: One Example GestureAt(Truck(name=Truck17, latitude=-10, longitude=10, moveStatus=“Stopped”)) RelevantRelationship(InstanceOf(this, Truck)) RelevantRelationship(SameAs(this.moveStatus, “Stopped”))

construct combinations of rules from all pos examples posCombo(?T) IF relevantFromPosEx1(?T), relevantFromPosEx2(?T),..., relevantFromPosExN(?T) and neg examples negCombo(?T) IF ~relevantFromNegEx1(?T), ~relevantFromNegEx2(?T),..., ~relevantFromNegExN(?T) GestureAt(Car12) // Negative Example relevance RelevantRelationship(arg1 = Not(InstanceOf(arg1 = this, arg2 = Truck))) Handling Relevance: Across All Examples

construct all comboRules: cross products across different combos allCombo(?T) IF posCombo(?T), negCombo(?T) all predicates (relevantFrom, posCombo, negCombo, allCombo) added to background during search also combine advice about examples using OR Handling Relevance: Across All Examples We use both AND and OR because a human teacher might teach – a conjunctive concept with each example illustrating only one piece of the concept, or – a disjunctive concept with several alternatives, with each alternative illustrated via a different example. – Added bonus: robustness to teacher errors

construct all comboRules: cross products across different combos allCombo(?T) IF posCombo(?T), negCombo(?T) all predicates (relevantFrom, posCombo, negCombo, allCombo) added to background during search also combine advice about examples using OR Handling Relevance: Across All Examples We use both AND and OR because a human teacher might teach – a conjunctive concept with each example illustrating only one piece of the concept, or – a disjunctive concept with several alternatives, with each alternative illustrated via a different example. – Added bonus: robustness to teacher errors

Outline Bootstrap Learning (BL), Inductive Logic Programming (ILP) and Challenges The Wisconsin Relevance Interpreter: Incorporating teacher advice into an ILP agent The Onion: Automating agent’s ILP runs across tasks and domains without expert intervention Conclusions

The Need for Automation supervised learning methods require – intervention by domain experts for feature selection, predicate design – intervention by machine learning experts for parameter selection, guiding search automation of ILP setup is required for – usability by non-expert human teachers – task and domain independence – standard automation methods (such as CV) not easily applicable because of small number of examples

The Need for Automation supervised learning methods require – intervention by domain experts for feature selection, predicate design – intervention by machine learning experts for parameter selection, guiding search automation of ILP setup is required for – use by non-expert human teachers – task and domain independence – standard automation methods (such as CV) not easily applicable because of small number of examples

ONION: A Multi-Layer Strategy Automating the ILP Setup Problem start with a small search space and incrementally expand until a “good theory” is found during the search – prefer simple concepts over complex concepts – prefer highly relevant sub-concepts (exploit relevance) – try various parameter settings for best theory a “good theory” covers most of positive examples and very few negative examples level of coverage can be adjusted to find a theory of acceptable accuracy

ONION: A Multi-Layer Strategy Automating the ILP Setup Problem start with a small search space and incrementally expand until a “good theory” is found during the search – prefer simple concepts over complex concepts – prefer highly relevant sub-concepts (exploit relevance) – try various parameter settings for best theory a “good theory” covers most of positive examples and very few negative examples level of coverage can be adjusted to find a theory of acceptable accuracy

ONION: A Multi-Layer Strategy Automating the ILP Setup Problem start with a small search space and incrementally expand until a “good theory” is found during the search – prefer simple concepts over complex concepts – prefer highly relevant sub-concepts (exploit relevance) – try various parameter settings for best theory a “good theory” covers most of positive examples and very few negative examples level of coverage can be adjusted to find a theory of acceptable accuracy

ONION: A Multi-Layer Strategy short rules; highly relevant combo-rules medium-length rules; all relevant features flip example labels (learn negation of concept) consider small subset of parameter settings long rules; all features More details in longer technical paper (Walker et al., 2010)

Outline Bootstrap Learning (BL), Inductive Logic Programming (ILP) and Challenges The Wisconsin Relevance Interpreter: Incorporating teacher advice into an ILP agent The Onion: Automating agent’s ILP runs across tasks and domains without expert intervention Results and Conclusions

Experimental Results Domains and data for agent – 3 domains, 56 lessons (concepts) – 7.6 labeled examples per concept on average – mixture of positives and negatives – teacher gives relevant information for each concept Results – without relevance: average accuracy = 63.9% – with relevance: average accuracy = 100% – no expert intervention – further robustness experiments currently in progress

Experimental Results Domains and data for agent – 3 domains, 56 lessons (concepts) – 7.6 labeled examples per concept on average – mixture of positives and negatives – teacher gives relevant information for each concept Results – without relevance: average test-set accuracy = 63.9% – with relevance: average test-set accuracy = 100% – “mystery domain”: average test-set accuracy = 100% – further robustness experiments currently in progress

Conclusions Bootstrap Learning agent interacts with teacher using natural instruction methods – gets examples, background, relevance information Teacher provided relevance can dramatically reduce – ILP search needed – number of examples needed Challenge #1: Generalizing advice about specific examples (Solution: Wisconsin Relevance Interpreter) Challenge #2: Automating ILP runs for wider usability, and task- and domain-independence (Solution: Onion)

Conclusions Bootstrap Learning agent interacts with teacher using natural instruction methods – gets examples, background, relevance information Teacher provided relevance can dramatically reduce – ILP search needed – number of examples needed Challenge #1: Generalizing advice about specific examples (Solution: Wisconsin Relevance Interpreter) Challenge #2: Automating ILP runs for wider usability, and task- and domain-independence (Solution: Onion)

Questions? Acknowledgements: The authors gratefully acknowledge support of the Defense Advanced Research Projects Agency under DARPA grant FA C Views and conclusions contained in this document are those of the authors and do not necessarily represent the official opinion or policies, either expressed or implied of the US government or of DARPA. References: (Oblinger, 2006) Oblinger, D. Bootstrap learning - external materials (Muggleton and De Raedt, 1994) Muggleton, S. and De Raedt, L. Inductive logic programming: Theory and methods. Journal of Logic Programming, 19/20:629–679, (Walker et al., 2010) Walker, T., Kunapuli, G., Natarajan, S., Shavlik, J. W. and Page, D. Automating the ILP Setup Task: Converting User Advice about Specific Examples into General Background Knowledge. International Conference on Inductive Logic Programming, (to appear)