WISCONSIN UNIVERSITY OF WISCONSIN - MADISON Broadening the Applicability of Relational Learning Trevor Walker Ph.D. Defense 1.

Slides:



Advertisements
Similar presentations
Artificial Intelligence
Advertisements

Computer Science CPSC 322 Lecture 25 Top Down Proof Procedure (Ch 5.2.2)
1 Update on Learning By Observation Learning from Positive Examples Only Tolga Konik University of Michigan.
Inductive Logic Programming: The Problem Specification Given: –Examples: first-order atoms or definite clauses, each labeled positive or negative. –Background.
The Assembly Language Level
Combining Inductive and Analytical Learning Ch 12. in Machine Learning Tom M. Mitchell 고려대학교 자연어처리 연구실 한 경 수
CSE (c) S. Tanimoto, 2008 Propositional Logic
Constraint Logic Programming Ryan Kinworthy. Overview Introduction Logic Programming LP as a constraint programming language Constraint Logic Programming.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Relational Data Mining in Finance Haonan Zhang CFWin /04/2003.
Algorithms and Problem Solving-1 Algorithms and Problem Solving.
1 Learning from Behavior Performances vs Abstract Behavior Descriptions Tolga Konik University of Michigan.
Inference and Resolution for Problem Solving
Statistical Relational Learning for Link Prediction Alexandrin Popescul and Lyle H. Unger Presented by Ron Bjarnason 11 November 2003.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Knowledge Representation. 2 Outline: Output - Knowledge representation  Decision tables  Decision trees  Decision rules  Rules involving relations.
Let remember from the previous lesson what is Knowledge representation
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 22 Jim Martin.
© Jesse Davis 2006 View Learning Extended: Learning New Tables Jesse Davis 1, Elizabeth Burnside 1, David Page 1, Vítor Santos Costa 2 1 University of.
WISCONSIN UNIVERSITY OF WISCONSIN - MADISON Integrating Knowledge Capture and Supervised Learning through a Human-Computer Interface Trevor Walker, Gautam.
RL via Practice and Critique Advice Kshitij Judah, Saikat Roy, Alan Fern and Tom Dietterich PROBLEM: RL takes a long time to learn a good policy. Teacher.
1 First order theories. 2 Satisfiability The classic SAT problem: given a propositional formula , is  satisfiable ? Example:  Let x 1,x 2 be propositional.
Knowledge & Reasoning Logical Reasoning: to have a computer automatically perform deduction or prove theorems Knowledge Representations: modern ways of.
31 st October, 2012 CSE-435 Tashwin Kaur Khurana.
Katanosh Morovat.   This concept is a formal approach for identifying the rules that encapsulate the structure, constraint, and control of the operation.
Inductive Logic Programming Includes slides by Luis Tari CS7741L16ILP.
Learning from Human Teachers: Issues and Challenges for ILP in Bootstrap Learning Sriraam Natarajan 1, Gautam Kunapuli 1, Richard Maclin 3, David Page.
Slides for “Data Mining” by I. H. Witten and E. Frank.
Author: James Allen, Nathanael Chambers, etc. By: Rex, Linger, Xiaoyi Nov. 23, 2009.
Requirements Engineering
DECIDABILITY OF PRESBURGER ARITHMETIC USING FINITE AUTOMATA Presented by : Shubha Jain Reference : Paper by Alexandre Boudet and Hubert Comon.
Conjunctive normal form: any formula of the predicate calculus can be transformed into a conjunctive normal form. Def. A formula is said to be in conjunctive.
Treatment Learning: Implementation and Application Ying Hu Electrical & Computer Engineering University of British Columbia.
Skill Acquisition via Transfer Learning and Advice Taking Lisa Torrey, Jude Shavlik, Trevor Walker University of Wisconsin-Madison, USA Richard Maclin.
Transfer Learning Via Advice Taking Jude Shavlik University of Wisconsin-Madison.
Declarative vs Procedural Programming  Procedural programming requires that – the programmer tell the computer what to do. That is, how to get the output.
Transfer in Reinforcement Learning via Markov Logic Networks Lisa Torrey, Jude Shavlik, Sriraam Natarajan, Pavan Kuppili, Trevor Walker University of Wisconsin-Madison,
Modern Topics in Multivariate Methods for Data Analysis.
Fuzzy Inference (Expert) System
Introduction to Machine Learning Supervised Learning 姓名 : 李政軒.
Speeding Up Relational Data Mining by Learning to Estimate Candidate Hypothesis Scores Frank DiMaio and Jude Shavlik UW-Madison Computer Sciences ICDM.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Relational Macros for Transfer in Reinforcement Learning Lisa Torrey, Jude Shavlik, Trevor Walker University of Wisconsin-Madison, USA Richard Maclin University.
Predicate Calculus Syntax Countable set of predicate symbols, each with specified arity  0. For example: clinical data with multiple tables of patient.
Machine Learning Chapter 5. Artificial IntelligenceChapter 52 Learning 1. Rote learning rote( โรท ) n. วิถีทาง, ทางเดิน, วิธีการตามปกติ, (by rote จากความทรงจำ.
CS Introduction to AI Tutorial 8 Resolution Tutorial 8 Resolution.
Fuzzy Systems Michael J. Watts
Advice Taking and Transfer Learning: Naturally-Inspired Extensions to Reinforcement Learning Lisa Torrey, Trevor Walker, Richard Maclin*, Jude Shavlik.
For Monday Finish chapter 19 No homework. Program 4 Any questions?
Automated Reasoning Early AI explored how to automated several reasoning tasks – these were solved by what we might call weak problem solving methods as.
Ch. 13 Ch. 131 jcmt CSE 3302 Programming Languages CSE3302 Programming Languages (notes?) Dr. Carter Tiernan.
“Principles of Soft Computing, 2 nd Edition” by S.N. Sivanandam & SN Deepa Copyright  2011 Wiley India Pvt. Ltd. All rights reserved. CHAPTER 12 FUZZY.
For Monday Finish chapter 19 Take-home exam due. Program 4 Any questions?
Automated Reasoning Systems For first order Predicate Logic.
Learning to Share Meaning in a Multi-Agent System (Part I) Ganesh Padmanabhan.
CS 5751 Machine Learning Chapter 10 Learning Sets of Rules1 Learning Sets of Rules Sequential covering algorithms FOIL Induction as the inverse of deduction.
Computer Science CPSC 322 Lecture 22 Logical Consequences, Proof Procedures (Ch 5.2.2)
DeepDive Model Dongfang Xu Ph.D student, School of Information, University of Arizona Dec 13, 2015.
CS104:Discrete Structures Chapter 2: Proof Techniques.
Chap. 10 Learning Sets of Rules 박성배 서울대학교 컴퓨터공학과.
Interface specifications At the core of each Larch interface language is a model of the state manipulated by the associated programming language. Each.
Learning Procedural Knowledge through Observation -Michael van Lent, John E. Laird – 인터넷 기술 전공 022ITI02 성유진.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Frank DiMaio and Jude Shavlik Computer Sciences Department
Chapter 7. Propositional and Predicate Logic
Trevor Walker, Gautam Kunapuli, Noah Larsen, David Page, Jude Shavlik
Richard Maclin University of Minnesota - Duluth
Algorithms and Problem Solving
Topics Introduction to Functions Defining and Calling a Function
Refining Rules Incorporated into Knowledge-Based Support Vector Learners via Successive Linear Programming Richard Maclin University of Minnesota - Duluth.
Presentation transcript:

WISCONSIN UNIVERSITY OF WISCONSIN - MADISON Broadening the Applicability of Relational Learning Trevor Walker Ph.D. Defense 1

Introduction – Relational Learning car(train1, car1) train1 car(train1, car2) cargo(car1, penguin) count(car1, 2) cargo(car2, giraffe) count(car2, 3) Relations between “things” “things” – constants, values, etc. 2

Relational Supervised Learning Specifically Inductive Logic Programming (ILP) Learner Model 3

Advantages and Disadvantages Better representation of world More expressive models using logical expressions Hard for user without ILP expertise 4

Thesis Statement Providing automatic advice-taking algorithms, human-computer interfaces, and automatic parameter tuning greatly increases the applicability of ILP, enabling use of ILP by users whose expertise lies outside of ILP. 5

ILP Task Definition Given Positive examples Negative examples Background knowledge Do Learn a model that covers (i.e., implies) As many positive examples as possible As few negative examples as possible 6

Here Comes the Circus Learn rule(s) distinguishing circus trains from non-circus trains circus(Train) ← car(Train, C1), cargo(C1, penguin), count(C1, 2), car(Train, C2), cargo(C2, giraffe), count(C2, 3). 7

ILP Task Definition Positives and negatives Background knowledge – Facts – Rules Search space specification Parameters 8

ILP : Positive and Negatives Positive example: circus(train1) Negative example: circus(train2) 9

ILP : Defining the Search Space Determinations: What literals are allowed Modes: Constraints on arguments of literals – circus(-train) – car(+train, -car) – cargo(+car, #animal) Example Rule: circus(Train ) ← car(Train, Car), cargo(Car, penguin) Introduces Train Variable Uses Train Variable penguin constant 10

ILP : Background Knowledge Logical statements expressing what is known – Facts – Ground positive literals – Rules – Universally qualified implications Facts are easy Rules are hard 11

ILP : Specifying Parameters MaximumClausesInTheory MinimumRecallPerClause MinimumPrecisionPerClause MinimumTheoryPrecision MaximumClauseLength MaximumSearchNodes F1EarlyStoppingScore LearnNegatedConcept 12

Contributions 1.Background Knowledge Acquisition via Advice-Taking [ Walker et al., 2010 ] 2.Advice-Taking via Human-Computer Interface (HCI) [ Walker et al., 2011 ] 3.Parameter Tuning via the O NION [ Walker et al., 2010 ] 4.ILP Applied to Reinforcement Learning Tasks [ Torrey et al., 2007; Walker et al., 2007 ] 13

Outline 1.Introduction 2.Advice-Taking 3.Human-Computer Interfaces 4.Parameter Tuning (Time Permitting) 5.Conclusions 14

What is Advice? Teacher-provided statements Hints about the concept being learned Advice pertains to specific positive or negative examples [ Walker et al., 2010 ] 15

Advice about Specific Examples Specific Advice train1 is a circus train if car1 is carrying penguins car1 is carrying 2 objects 16 General Advice A Train is a circus train if the Train has a Car carrying penguins and the same Car is carrying 2 objects

Automated Advice-Taking Goals 1.Inclusion of most teacher- mentioned advice 2.Robustness to teacher errors 3.Create both background knowledge and setup ILP search space 17

Overview of Algorithm Teacher provides advice Convert to grounded implications Generalize implications Create background knowledge by combining all advice Setup ILP search space based on generated knowledge 18

Sample Teacher Advice Teacher says – Example ex(a) is positive instance because p(a) – Example ex(b) is negative instance because q(b, b) – Example ex(c) is positive instance because r(c) Possible meanings of advice ( p(X) ∧ ¬ q(X, Z) ) ∨ r(X) p(X) ∨ ¬ q(X, Y) ∨ r(X) 19

Assumptions Atomic features of domain are known Teacher talks about specific examples Teacher specifies ground advice Teacher understands basic logic 20

Utilizing Teacher-Provided Advice Example ex(a) is positive instance because p(a) Example ex(b) is negative instance because q(b, b) Example ex(c) is positive instance because r(c) q(b,b) may seem like an odd statement 21

Utilizing Teacher-Provided Advice Example ex(a) is positive instance because p(a) Example ex(b) is negative instance because q(b, f(x)) Example ex(c) is positive instance because r(c) q(b,b) may seem like an odd statement For instance: length(train, 3), equals(3, 3) becomes: length(train, X), equals(X, 3) 22

Utilizing Teacher-Provided Advice Example ex(a) is positive instance because p(a) Example ex(b) is negative instance because q(b, b) Example ex(c) is positive instance because r(c) Clauses may contain multiple literals 23

Utilizing Teacher-Provided Advice Example ex(a) is positive instance because p(a) Example ex(b) is negative instance because q(b, b) Example ex(c) is positive instance because r(c), s(d) Clauses may contain multiple literals 24

Utilizing Teacher-Provided Advice Example ex(a) is positive instance because p(a) Example ex(b) is negative instance because q(b, b) Example ex(c) is positive instance because r(c) Separate pieces of advice for a single example allowed 25

Utilizing Teacher-Provided Advice Example ex(a) is positive instance because p(a) Example ex(b) is negative instance because q(b, b) Example ex(a) is positive instance because r(c) Separate pieces of advice for a single example allowed 26

Utilizing Teacher-Provided Advice Example ex(a) is positive instance because p(a) Example ex(b) is negative instance because q(b, b) Example ex(c) is positive instance because r(c) 27

Step 1: Form Implications Form the implication of the advice ground body to the ground example Negate advice attached to negative examples ex(a) p(a) ex(b) ¬q(b, b) ex(c) r(c) Negative advice is negated 28

Step 2: Generalize Advice Perform anti-substitution of constants to variables Direct – Identical constant = same variable Indirect – Identical constants split into separate variables ex(X) p(X) ex(Y) ¬q(Y, Y)ex(Y) ¬q(Y, Z) ex(Z) r(Z) 29

Step 3: Standardize Advice Formula Unify heads and apply substitution to antecedents After unification, remove head leaving only body ex(X) p(X) ex(X) ¬q(X, X)ex(X) ¬q(X, Z) ex(X) r(X) 30

Step 3: Standardize Advice Formula Unify head and apply substitution to antecedents After unification, remove head leaving only body p(X) ¬q(X, X)¬q(X, Z) r(X) 31

Step 4: Create Logical Combinations Combine body via ∧ or ∨ connectives Combinations may connect only subset of formula Combinations at most one formula derived from any single piece of teacher advice p(X)p(X) ∧ ¬q(X, X) ∧ r(X) ¬q(X, X), ¬q(X, Z) ¬q(X, Z) ∨ r(X) r(X) ¬q(X, X) ∧ ¬q(X, Z) 32

Step 5: Generate Background Clauses, Determinations, & Modes Create anonymous background clauses Create modes – Variables derived from example constants assigned + or # modes – New variables in head predicate become outputs with - or # modes pred1(X) ← p(X) ∧ ¬q(X, X) ∧ r(X) pred2(X, Z) ← ¬q(X, Z) ∨ r(X) Modes: pred1(+), pred1(#), pred2(+,-), pred2(#,#),... 33

Final Results Example ex(a) is positive instance because p(a) Example ex(b) is negative instance because q(b, b) Example ex(c) is positive instance because r(c) pred1(X) ← p(X) ∧ ¬q(X, X) ∧ r(X) pred2(X, Z) ← p(X) ∧ q(X, Z) ∨ r(X) pred3(X, Z) ← q(X, Z) ∨ r(X) 34

Clause Priorities Clauses searched according to priority High All advice Medium Some advice or variables split, not both Low Some advice and variables split pred1(X) ← p(X) ∧ ¬q(X, X) ∧ r(X)HIGH pred2(X, Z) ← p(X) ∧ q(X, Z) ∨ r(X)MEDIUM pred3(X, Z) ← q(X, Z) ∨ r(X)LOW 35

Difficulties - Multiple generalizations Example ex(a) is positive instance because “p(a) ∧ q(a)” pred1(X) ← p(X) ∧ q(X)pred1(X) ← p(X) ∧ q(Y) pred1(X) ← p(Y) ∧ q(X)pred1(X) ← p(Y) ∧ q(Y) Add multiple background rules with different priorities 36

Difficulties - Constants Example ex(a) is positive instance because “equals(a, 1)” pred1(X) ← equals(X, Y)pred1(X) ← equals(X, 1) pred1(X) ← equals(a, Y)pred1(X) ← equals(a, 1) Evaluate on the training examples to determine most likely 37

Difficulties – Exposing Variables Example ex(a) is positive instance because “width(a, 7), height(a, 5), multiply(5, 7, 35)” pred1(X) ← width(X, W), height(X, H), multiply(W, H, A) pred1(X, A) ← width(X, W), height(X, H), multiply(W, H, A) Expose last variable created during generalization 38

Experiments : Methodology Bootstrap Learning domain 14 separate tasks designed by 3 rd party with examples and advice Up to 100 training examples (depending on the experiment) 100 testing examples 39

Experiment – Advice Efficacy 40

Experiment – Errors of Omission 41

Jumping into the Search Space 42

Experiment – Example Noise 43

Advice Taking: Conclusions Automates both background knowledge creation and ILP setup Effective during testing on independent domains Enables use of ILP by users without expertise in ILP 44

Outline 1.Introduction 2.Advice Taking 3.Human-Computer Interfaces 4.Parameter Tuning (Time Permitting) 5.Conclusions 45

Learning with Domain Knowledge Domain-expert knowledge LearnerModel += Approach: Advice-taking Some learning algorithms can take advice 46

Domain Knowledge Difficulties Learner + Model = Domain-expert knowledge 47

Advice Taking: Difficulties 1.Articulating knowledge requires understanding of underlying algorithm – requires AI expert 2.Identifying important knowledge is difficult Learner ? ? Teacher advice 48

Domain Knowledge Solutions 1.Use advice-taking Human-Computer Interface (HCI) 2.Iteratively acquire knowledge Learner Teacher advice 49

Learning Scenario Specifics Supervised Learning – Training examples & labels are given – Relational domains Knowledge in the form of advice – Hints to learner, possibly incorrect – Written in first-order logic – Grounded in a specific example 50

Advice Processing [Walker et al., 2010] train1 car1 Positive Example: circus(train1) giraffe Advice: train(train1), has_car(train1, car1), cargo(car1, giraffe) Generated knowledge: bk1(Train) ← train(Train), has_car(Train, Car), cargo(Car, giraffe) Advice: train(train1), has_car(train1, car1), cargo(car1, giraffe) Generated knowledge: bk1(Train) ← train(Train), has_car(Train, Car), cargo(Car, giraffe) Bad: First-Order Logic Good: Grounded Advice 51

Grounded Advice Easier for teacher to specify Easier to design HCI for grounded advice Can use previous techniques 52

Domain Knowledge Solutions 1.Iteratively acquire knowledge 2.Use advice-taking Human-Computer Interface (HCI) Knowledge Learner 53

Our Knowledge-Acquisition Loop Teacher Advice Advice HCI Review HCI ModelLearner 54

Case Study: WillTowerFall? Our HCI advice-taking case study WillTowerFall? – Subset of Wargus 55

WillTowerFall? Sample Scenario 56

Giving Advice via the HCI User selects example to provide advice for HCI presents example information User specifies advice by 1. Selecting objects 2. Specifying relations 57

Giving Advice via the HCI Wargus GUI 58

Giving Advice via the HCI Wargus GUI 59

Giving Advice via the HCI Wargus GUI 60

Giving Advice via the HCI Wargus GUI 61

What Would Users Like to Say? 1.First pass – collect advice in natural language 2.Second pass – provide GUI that supports most of natural language sentences 62

Iterative Knowledge Acquisition Advice HCI Review HCI ModelLearner Domain-expert knowledge 63

Reviewing / Correcting Results 64

Wargus Natural Language Advice Tower will fall – Three or more footmen can take down a tower if there is no moat – Five or more archers can take down a tower Tower will not fall – If there are only peasants, the tower stands – Four archers or less cannot take down a tower 90% representable by GUI 65

Experiments Wargus WillTowerFall? Use boosted relational dependency network (bRDN) [Natarajan et al., 2010] Advice-generalization algorithm [Walker et al., 2010] 66

Experiments Compare – No Advice – User-provide advice via HCI – Hand-written ILP expert advice 100 training examples 900 testing examples 67

Testset Results 68

HCI Conclusions Algorithms exist that take advice but high barrier to entry for non-expert users HCI - Allow input of advice about specific examples Enables use of ILP by users without expertise in ILP 69

Future Work on Advice HCI Evaluate iterative learning approach in more complex domains Develop domain- independent GUI for relational advice taking 70

Outline 1.Introduction 2.Advice-Taking 3.Human-Computer Interfaces 4.Parameter Tuning (Time Permitting) 5.Conclusions 71

Onion [Walker et al., 2010] Automated parameter turning framework Iterative-deepening style search of parameter space No input from user necessary 72

Experiment – Why is it Needed Used 14 Bootstrap learning tasks Used 24 training examples with no noise along with advice Evaluated cross-product of parameters MaximumLiteralsInClause and MaximumClausesInTheory 73

Experiment – Why is it Needed? 74

Experiment – Comparison to Grid-Search Used 3 standard ILP domains – Advised-By – Carcinogenesis – Mutagenesis Compared Onion with grid-search over all possible parameter combinations Evaluated on multiple folds 75

Experiments – Advised By 76

Experiments – Carcinogenesis 77

Experiments – Mutagenesis 78

Onion Conclusions Automated parameter tuning via the Onion is effective Grid-search can be equally effective but may result in overfitting Enables use of ILP by users without expertise in ILP 79

Outline 1.Introduction 2.Advice-Taking 3.Human-Computer Interfaces 4.Parameter Tuning (Time Permitting) 5.Conclusions 80

Overall Conclusions Created background knowledge and search- space specification from advice Extended advice-taking approach by examining use of HCI Designed an automated method for tuning the ILP parameters Enables use of ILP by users without expertise in ILP 81

Contributions 1.Background Knowledge Acquisition via Advice-Taking [ Walker et al., 2010 ] 2.Advice-Taking via Human-Computer Interface (HCI) [ Walker et al., 2011 ] 3.Parameter Tuning via the O NION [ Walker et al., 2010 ] 4.ILP Applied to Reinforcement Learning Tasks [ Torrey et al., 2007; Walker et al., 2007 ] 82

Acknowledgements Advisor: Jude Shavlik Committee: Mike Coen, Mark Craven, Rich Maclin, David Page Collaborators: Lisa Torrey, Sriraam Natarajan, Gautam Kunapuli, Noah Larsen Supported by DARPA grants HR , HR C-0060, and FA C

Questions Thank You 84