Presentation is loading. Please wait.

Presentation is loading. Please wait.

Kshitij Judah EECS, OSU Dissertation Proposal Presentation.

Similar presentations


Presentation on theme: "Kshitij Judah EECS, OSU Dissertation Proposal Presentation."— Presentation transcript:

1 Kshitij Judah EECS, OSU Dissertation Proposal Presentation

2  PART I: User-Initiated Learning  PART II: RL via Practice and Critique Advice  PART III: Proposed future directions for the PhD program  Extending RL via practice and critique advice  Active Learning for sequential decision making

3

4 All of CALO’s learning components can perform Learning In The Wild (LITW) But the learning tasks are all pre-defined by CALO’s engineers: – What to learn – What information is relevant for learning – How to acquire training examples – How to apply the learned knowledge UIL Goal: Make it possible for the user to define new learning tasks after the system is deployed

5 TIMELINETIMELINE Scientist Sets sensitivity to confidential Sends email to team Sends email to a colleague “Lunch today?” Does not set sensitivity to confidential Collaborates on a Classified project Sends email to team Forgets to set sensitivity to confidential Research Team

6 “Please do not forget to set sensitivity when sending email” Scientist Research Team Teaches CALO to learn to predict whether user has forgot to set sensitivity Sends email to team CALO reminds user to set sensitivity TIMELINETIMELINE

7 SAT Based Reasoning System SPARK Procedure Instrumented Outlook Instrumented Outlook Events user Integrated Task Learning user Modify Procedure Procedure Demonstration and Learning Task Creation User Interface for Feature Guidance User Selected Features user Feature Guidance Email + Related Objects CALO Ontology Trained Classifier Feature Guidance Machine Learner Knowledge Base Training Examples Learning Legal Features SAT Based Reasoning System Class Labels Compose new email

8 SAT Based Reasoning System SPARK Procedure Instrumented Outlook Instrumented Outlook Events user Integrated Task Learning user Compose new email Modify Procedure Procedure Demonstration and Learning Task Creation User Interface for Feature Guidance User Selected Features user Feature Guidance Email + Related Objects CALO Ontology Trained Classifier Feature Guidance Machine Learner Knowledge Base Training Examples Learning Legal Features SAT Based Reasoning System Class Labels

9 LAPDOG: Transforms an observed sequence of instrumented events into a SPARK procedure SPARK representation generalizes the dataflow between the actions of the workflow

10 TAILOR: Supports procedure editing – For UIL, it allows adding a condition to one or more steps in a procedure

11 – The condition becomes the new predicate to be learned

12 SAT Based Reasoning System SPARK Procedure Instrumented Outlook Instrumented Outlook Events user Integrated Task Learning user Compose new email Modify Procedure Procedure Demonstration and Learning Task Creation User Interface for Feature Guidance User Selected Features user Feature Guidance Email + Related Objects CALO Ontology Trained Classifier Feature Guidance Machine Learner Knowledge Base Training Examples Learning Legal Features SAT Based Reasoning System Class Labels

13

14 SAT Based Reasoning System SPARK Procedure Instrumented Outlook Instrumented Outlook Events user Integrated Task Learning user Compose new email Modify Procedure Procedure Demonstration and Learning Task Creation User Interface for Feature Guidance User Selected Features user Feature Guidance Email + Related Objects CALO Ontology Trained Classifier Feature Guidance Machine Learner Knowledge Base Training Examples Learning Legal Features SAT Based Reasoning System Class Labels

15 Goal: autonomously generate labeled training instances for the learning component from stored user emails Problem: actions used to create emails are not stored in the CALO knowledge base, so we need to infer how email was created {defprocedure do_rememberSensitivity.... [do: (openComposeEmailWindow $newEmail)] [do: (changeEmailField $newEmail "to")] [do: (changeEmailField $newEmail "subject")] [do: (changeEmailField $newEmail "body")] [if: (learnBranchPoint $newEmail) [do: (changeEmailField $newEmail "sensitivity")]] [do: (sendEmailInitial $newEmail)].... } Specifically, we want to know: Whether an email is an instance of the procedure? Which branch was taken during creation of the email? No such inference can be drawn

16 Domain Axioms Domain Axioms NewComposition  ComposeNewMail ReplyComposition  ReplyToMail HasAttachment  (AttachFile  ForwardMail) … SPARK Axioms SPARK Axioms ProcInstance  (U 1  U 2  …  U n ) (  forget  label)  (C 1  C 2  …  C n ) {defprocedure do_rememberSensitivity.... [do: (openComposeEmailWindow $newEmail)] [do: (changeEmailField $newEmail "to")] [do: (changeEmailField $newEmail "subject")] [do: (changeEmailField $newEmail "body")] [if: (learnBranchPoint $newEmail) [do: (changeEmailField $newEmail "sensitivity")]] [do: (sendEmailInitial $newEmail)].... } Label Analysis Formula (LAF) Label Analysis Formula (LAF) Knowledge Base NewComposition ReplyComposition HasToField HasSubject HasBody HasAttachment … Reasoning Engine Reasoning Engine E   forget ╞ (ProcInstance  Label) E   forget ╞ (ProcInstance   Label) otherwise Positive Example Negative Example Discard email

17

18 Logistic Regression is used as the core learning algorithm Features – Relational features extracted from ontology Incorporate User Advice on Features – Apply large prior variance on user selected features – Select prior variance on rest of the features through cross-validation Automated Model Selection – Parameters: Prior variance on weights, classification threshold – Technique: Maximization of leave-one-out cross-validation estimate of kappa (  )

19 Problems: – Attachment Prediction – Importance Prediction Learning Configurations Compared: – No User Advice + Fixed Model Parameters – User Advice + Fixed Model Parameters – No User Advice + Automatic parameter Tuning – User Advice + Automatic parameter Tuning User Advice: 18 keywords in the body text for each problem

20 Set of 340 emails obtained from a real desktop user 256 training set + 84 test set For each training set size, compute mean kappa (  ) using test set to generate learning curves  is a statistical measure of inter-rater agreement for discrete classes  is a common evaluation metric in cases when the classes have a skewed distribution

21 Attachment Prediction

22 Importance Prediction

23 We intended to test the robustness of the system to bad advice Bad advice was generated as follows: – Use SVM based feature selection in WEKA to produce a ranking of user provided keywords – Replace top three words in the ranking with randomly selected words from the vocabulary

24 Attachment Prediction

25 Importance Prediction

26 We want to evaluate the utility of the system for the user We use a new metric called Critical Cost Ratio (CCR) Intuition: A measure of how high cost of forgetting should be compared to cost of interruption for the system to be useful Intuition : Hence, if CCR is low, the system is useful more often For example, if CCR=10, then cost of forgetting should be 10 times more than cost of interruption for net benefit

27 Attachment Prediction At size 256, cost of forgetting should be at least 5 times of cost of interruption to gain net benefit from the system

28 Importance Prediction

29 User interfaces should support rich instrumentation, automation, and intervention User interfaces should come with models of their behavior User advice is helpful but not critical Self-tuning learning algorithms are critical for success

30 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:

31 PROBLEM:  Usually RL takes a long time to learn a good policy. Teacher behavior advice RESEARCH QUESTION: Can we make RL perform better with some outside help, such as critique/advice from teacher and how? GOALS:  Non-technical users as teachers  Natural interaction methods state action reward Environment

32 ? Policy Parameters  Trajectory Data Practice Session Advice Interface In a state s i action a i is bad, whereas action a j is good. Teacher Critique Session Critique Data

33 Trajectory Data Practice Session Advice Interface In a state s i action a i is bad, whereas action a j is good. Teacher Critique Session Policy Parameters  Critique Data Estimate Expected Utility using Importance Sampling. (Peshkin & Shelton, ICML 2002)

34 Imagine: Our teacher is an Ideal Teacher (Provides All Good Actions) Set of all Good actions Any action not in O(s i ) is suboptimal according to Ideal Teacher All actions are equally good Advice Interface Ideal Teacher Advice Interface Some good actions Some bad actions Some actions unlabeled

35 Imagine: Our teacher is an Ideal Teacher (Provides All Good Actions) Set of all Good actions Any action not in O(s i ) is suboptimal according to Ideal Teacher All actions are equally good Advice Interface Ideal Teacher  Learning Goal: Find a probabilistic policy, or classifier, that has a high probability of returning an action in O(s) when applied to s.  ALL Likelihood (L ALL ( ,C)) : Probability of selecting an action in O(S i ) given state s i

36  Coming back to reality: Not All Teachers are Ideal ! and provide partial evidence about O(s i ) Advice Interface  What about the naïve approach of treating as the true set O(s i ) ?  Difficulties:  When there are actions outside of that are equally good compared to those in, the learning problem becomes even harder.  We want a principled way of handling the situation where either or can be empty.

37 The Gradient of Expected Loss has a compact closed form. … and provide partial evidence about O(s i ) User Model Assume independence among different states. From corresponding for all states, we can get: Expected ALL Loss:

38 Map 1Map 2  Our Domain: Micro-management in tactical battles in the Real Time Strategy (RTS) game of Wargus.  5 friendly footmen against a group of 5 enemy footmen (Wargus AI).  Two battle maps, which differed only in the initial placement of the units.  Both maps had winning strategies for the friendly team and are of roughly the same difficulty.

39  Difficulty:  Fast pace and multiple units acting in parallel  Our setup:  Provide end-users with an Advice Interface that allows to watch a battle and pause at any moment.

40  Goal is to evaluate two systems 1.Supervised System = no practice session 2.Combined System = includes practice and critique  The user study involved 10 end-users  6 with CS background  4 no CS background  Each user trained both the supervised and combined systems  30 minutes total for supervised  60 minutes for combined due to additional time for practice  Since repeated runs are not practical results are qualitative  To provide statistical results we first present simulated experiments

41  After user study, selected the worst and best performing users on each map when training the combined system  Total Critique data: User#1: 36, User#2: 91, User#3: 115, User#4: 33.  For each user: divide critique data into 4 equal sized segments creating four data-sets per user containing 25%, 50%, 75%, and 100% of their respective critique data.  We provided the combined system with each of these data sets and allowed it to practice for 100 episodes. All results are averaged over 5 runs.

42  RL is unable to learn a winning policy (i.e. achieve a positive value).

43  With more critiques performance increases a little bit.

44  As the amount of critique data increases, the performance improves for a fixed number of practice episodes.  RL did not go past 12 health difference on any map even after 500 trajectories.

45  Even with no practice, the critique data was sufficient to outperform RL.  RL did not go past 12 health difference.

46  With more practice performance increases too.

47  Our approach is able to leverage practice episodes in order to improve the effectiveness on a given amount of critique data.

48  Goal is to evaluate two systems 1.Supervised System = no practice session 2.Combined System = includes practice and critique  The user study involved 10 end-users  6 with CS background  4 no CS background  Each user trained both the supervised and combined systems  30 minutes total for supervised  60 minutes for combined due to additional time for practice

49

50  Comparing to RL:  9 out of 10 users achieved 50 or more performance using Supervised System  6 out of 10 users achieved 50 or more performance using Combined System  Users effectively performed better than RL using either the Supervised or Combined Systems.  RL did not go past 12 health difference on any map even after 500 trajectories.

51  Frustrating Problems for Users   Large delay experience. (not an issue in many realistic settings)  Policy returned after practice was sometimes poor, seemed to be ignoring advice. (perhaps practice sessions were too short)  Comparing Combined and Supervised:  The end-users had slightly greater success with the supervised system v/s the combined system.  More users were able to achieve performance levels of 50 and 80 using the supervised system.

52

53  Understanding the effects of user models:  Study sensitivity of our algorithm to various settings of model parameters.  Robustness of our algorithm against inaccurate parameter settings.  Study benefits of using more elaborate user models.  Understanding the effects of mixing advice from multiple teachers:  Pros: addresses incompleteness and lack of quality of advice from a single teacher.  Cons: introduces variations and more complex patterns that are hard to generalize.  Study benefits and harms of mixing advice.  Understanding the effects of advice types:  Study the effects of feedback only versus mixed advice.

54  Current advice collection mechanism is very basic:  An entire episode is played before the teacher.  Teacher scans the episode to locate places where critique is needed.  Only one episode is played.  Problems with current advice collection mechanism:  Teacher is fully responsible for locating places where critique is needed.  Scanning an entire episode is very cognitively demanding.  Good possibility of missing places where advice is critical.  Showing only one episode is a limitation, especially in stochastic domains.  GOAL: Learner should itself discover places where it needs advice and query teacher at those places.

55 Trajectory Data Practice Session Advice Interface In a state s i action a i is bad, whereas action a j is good. Teacher Critique Session Policy parameters  Critique Data Current policy Full Episode:

56 Trajectory Data Practice Session Advice Interface In a state s i action a i is bad, whereas action a j is good. Teacher Critique Session Policy parameters  Critique Data Active Learning Module Best sequence: Current policy Cost Model ($$) Problem: How to select that best optimizes benefit-to-cost tradeoff?

57  Few techniques exist for the problem of ‘active learning’ in sequential decision making with an external teacher  All techniques make some assumptions that work only for certain applications  Some techniques request full demonstration from start state  Some techniques assume teacher is always available and request a single or multi-step demonstration when needed  Some techniques removes assumption of teacher being present at all times but they pause till the request for demonstration is satisfied

58  We feel such assumptions are unnecessary in general  Providing demonstration is quite labor intensive and sometimes not even practical  We instead seek feedback and guidance on potential execution traces of our policy  Pausing and waiting for the teacher is also inefficient  We never want to pause but keep generating execution traces from our policy for teacher to critique later when he/she is available

59  Active learning is well developed for supervised setting  All instances come from single distribution of interest  Best instance is selected based on some criterion and queried for its label  In our setting, the distribution of interest is the distribution of states along the teacher's policy (or a good policy)  Asking queries about states that are far off of the teacher's policy is likely to not produce any useful feedback (losing states in Chess or Wargus)  Learner faces additional challenge to identify states that occur along the teacher's policy and query in those states

60  Define new performance metric: Expected Common Prefix Length (ECPL): is # of time steps and agree up to first disagreement, starting from

61 State with first disagreement Common Prefix Length Unimportant States

62  Ideally we should select sequences that directly maximize ECPL  Heuristic: Identify sequence that contains states with high probability of first disagreement (similar to uncertainty sampling) Common Prefix Length Unimportant States States where first disagreement is likely High Confidence Execution (most likely we are on right path) Low Confidence Execution (most likely we are going to make a wrong turn) (most likely we should not be here)

63  Fixed length sequences  Variable length sequences: compute optimal length that trades off benefit from critique versus cost of critique. We plan to use return on investment heuristic proposed by Haertel et. Al. (Return on Investment for Active Learning, NIPS Workshop on Cost Sensitive Learning, 2009) Common Prefix Length Unimportant States States where first disagreement is likely High Confidence Execution (most likely we are on right path) Low Confidence Execution (most likely we are going to make a wrong turn) (most likely we should not be here)

64  Practice followed by active learning: let the agent practice, when critique session starts use heuristics from supervised setting to select best sequence for the current policy  Modify heuristic from supervised setting: Can we better discover sequences with first disagreement using information from practice  Based on self assessment during practice

65  We will use RTS game Wargus as our domain  Extend our current tactical battle micromanagement scenarios with more and different types of combat units  Teach control policies for attack and maneuver behavior of individual units  Teach policies to control low-level actions in resource gathering tasks  User Studies: Conduct further user studies

66 PeriodResearch Topic Aug 2010Get proposal approved. Sept-Oct 2010Active learning in supervised setting, teacher models (for journal). Nov-Dec 2010Active learning in practice setting, advice patterns (for journal). Jan 2011 Submit a paper to IJCAI/AAAI. Feb-Mar 2011 Start writing journal paper, mixing advice (for journal). Apr-May 2011 Finish writing journal paper, start writing dissertation. June-Aug 2011 Finish writing dissertation, submit journal paper. Sept 2011 defense.

67  Presented our completed work on:  User-Initiated Learning  RL via Practice and Critique Advice  Proposed future directions for the PhD program:  Extending RL via practice and critique advice  Active Learning for sequential decision making  Presented potential approaches and evaluation plans in order to carry out proposed work  Presented a timeline for completing the proposed work

68

69 Figure: Fraction of positive, negative and mixed advice. Supervised Combined  Positive (or Negative) advice is where the user only gives feedback on the action taken by the agent.  Mixed is where the user not only gives feedback on the agent's action but also suggests alternative actions to the agent.


Download ppt "Kshitij Judah EECS, OSU Dissertation Proposal Presentation."

Similar presentations


Ads by Google