A Hierarchical Framework for Composing Nested Web Processes Haibo Zhao, Prashant Doshi LSDIS Lab, Dept. of Computer Science, University of Georgia 4 th.

Slides:



Advertisements
Similar presentations
A Decision-Theoretic Model of Assistance - Evaluation, Extension and Open Problems Sriraam Natarajan, Kshitij Judah, Prasad Tadepalli and Alan Fern School.
Advertisements

CSSSIA Workshop – WWW 2008 Speeding up Web Service Composition with Volatile External Information John Harney, Prashant Doshi LSDIS Lab, Dept. of Computer.
Reinforcement Learning
Dialogue Policy Optimisation
Markov Decision Processes (MDPs) read Ch utility-based agents –goals encoded in utility function U(s), or U:S  effects of actions encoded in.
Markov Decision Process
SA-1 Probabilistic Robotics Planning and Control: Partially Observable Markov Decision Processes.
CSE-573 Artificial Intelligence Partially-Observable MDPS (POMDPs)
Experimental Design, Response Surface Analysis, and Optimization
LCSLCS 18 September 2002DARPA MARS PI Meeting Intelligent Adaptive Mobile Robots Georgios Theocharous MIT AI Laboratory with Terran Lane and Leslie Pack.
A Hybridized Planner for Stochastic Domains Mausam and Daniel S. Weld University of Washington, Seattle Piergiorgio Bertoli ITC-IRST, Trento.
COSC 878 Seminar on Large Scale Statistical Machine Learning 1.
Markov Decision Processes
Planning under Uncertainty
1 Simulation Modeling and Analysis Session 13 Simulation Optimization.
KI Kunstmatige Intelligentie / RuG Markov Decision Processes AIMA, Chapter 17.
Markov Decision Models for Order Acceptance/Rejection Problems Florian Defregger and Heinrich Kuhn Florian Defregger and Heinrich Kuhn Catholic University.
Bayesian Reinforcement Learning with Gaussian Processes Huanren Zhang Electrical and Computer Engineering Purdue University.
Nov 14 th  Homework 4 due  Project 4 due 11/26.
Integrating POMDP and RL for a Two Layer Simulated Robot Architecture Presented by Alp Sardağ.
1 Hybrid Agent-Based Modeling: Architectures,Analyses and Applications (Stage One) Li, Hailin.
Discretization Pieter Abbeel UC Berkeley EECS
Markov Decision Processes
Department of Computer Science Undergraduate Events More
Reinforcement Learning Yishay Mansour Tel-Aviv University.
Search and Planning for Inference and Learning in Computer Vision
1 Endgame Logistics  Final Project Presentations  Tuesday, March 19, 3-5, KEC2057  Powerpoint suggested ( to me before class)  Can use your own.
Planning and Verification for Stochastic Processes with Asynchronous Events Håkan L. S. Younes Carnegie Mellon University.
Generalized and Bounded Policy Iteration for Finitely Nested Interactive POMDPs: Scaling Up Ekhlas Sonu, Prashant Doshi Dept. of Computer Science University.
Real-Time Systems Hierarchical Real-Time Systems for Imprecise Computation Model The 5th EuroSys Doctoral Workshop (EuroDW 2011) Guy Martin.
Toward Optimal and Efficient Adaptation in Web Processes Prashant Doshi LSDIS Lab., Dept. of Computer Science, University of Georgia Joint work with: Kunal.
1 ECE-517 Reinforcement Learning in Artificial Intelligence Lecture 7: Finite Horizon MDPs, Dynamic Programming Dr. Itamar Arel College of Engineering.
Stochastic DAG Scheduling using Monte Carlo Approach Heterogeneous Computing Workshop (at IPDPS) 2012 Extended version: Elsevier JPDC (accepted July 2013,
Solving Large Markov Decision Processes Yilan Gu Dept. of Computer Science University of Toronto April 12, 2004.
An Ontological Framework for Web Service Processes By Claus Pahl and Ronan Barrett.
Simulation is the process of studying the behavior of a real system by using a model that replicates the behavior of the system under different scenarios.
Haley: A Hierarchical Framework for Logical Composition of Web Services Haibo Zhao, Prashant Doshi LSDIS Lab, Dept. of Computer Science, University of.
Department of Computer Science Undergraduate Events More
1 Qualitative Reasoning of Distributed Object Design Nima Kaveh & Wolfgang Emmerich Software Systems Engineering Dept. Computer Science University College.
Solving POMDPs through Macro Decomposition
Carnegie Mellon University
Reinforcement Learning Yishay Mansour Tel-Aviv University.
Decision-Theoretic Planning with Asynchronous Events Håkan L. S. Younes Carnegie Mellon University.
Haley: An End-to-End, Scalable Web Service Composition Tool Haibo Zhao, Prashant Doshi LSDIS Lab., The University of Georgia 17th International World Wide.
CUHK Learning-Based Power Management for Multi-Core Processors YE Rong Nov 15, 2011.
WSP: A Network Coordinate based Web Service Positioning Framework for Response Time Prediction Jieming Zhu, Yu Kang, Zibin Zheng and Michael R. Lyu The.
Twenty Second Conference on Artificial Intelligence AAAI 2007 Improved State Estimation in Multiagent Settings with Continuous or Large Discrete State.
An Optimal Distributed Call Admission control for Adaptive Multimedia in Wireless/Mobile Networks Reporter: 電機所 鄭志川.
16/11/ Semantic Web Services Language Requirements Presenter: Emilia Cimpian
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
Department of Computer Science Undergraduate Events More
1 Chapter 17 2 nd Part Making Complex Decisions --- Decision-theoretic Agent Design Xin Lu 11/04/2002.
A Formalism for Stochastic Decision Processes with Asynchronous Events Håkan L. S. YounesReid G. Simmons Carnegie Mellon University.
4 th International Conference on Service Oriented Computing Adaptive Web Processes Using Value of Changed Information John Harney, Prashant Doshi LSDIS.
Learning for Physically Diverse Robot Teams Robot Teams - Chapter 7 CS8803 Autonomous Multi-Robot Systems 10/3/02.
Generalized Point Based Value Iteration for Interactive POMDPs Prashant Doshi Dept. of Computer Science and AI Institute University of Georgia
Markov Decision Processes AIMA: 17.1, 17.2 (excluding ), 17.3.
Marilyn Wolf1 With contributions from:
OPERATING SYSTEMS CS 3502 Fall 2017
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 3
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 3
16th International World Wide Web Conference Speeding up Adaptation of Web Service Compositions Using Expiration Times John Harney, Prashant Doshi LSDIS.
Markov Decision Processes
Scheduling Design and Verification for Open Soft Real-time Systems
Markov Decision Processes
Hidden Markov Models (cont.) Markov Decision Processes
Presented By: Darlene Banta
Reinforcement Learning Dealing with Partial Observability
CS 416 Artificial Intelligence
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 3
Presentation transcript:

A Hierarchical Framework for Composing Nested Web Processes Haibo Zhao, Prashant Doshi LSDIS Lab, Dept. of Computer Science, University of Georgia 4 th International Conference on Service Oriented Computing

Introduction Web process  Business processes with Web services as components Existing approaches to Composition: AI planning  Classical planning techniques  Golog (McIlraith2001), MBP-based planning (Traverso2003), HTN planning (Wu2003), Synthy (Srivastava2005)  Decision-theoretic planning  MDP (Doshi2004) Limitations  Classical planning assumes deterministic behavior of Web services  Guarantees correctness but not optimality  Existing approaches do not scale well to large processes Our approach  Hierarchy in Web processes to address the scalability problem  A hierarchical semi-MDP to model the hierarchy  Stochastic SMDP model to handle uncertainties and optimality

Motivating Scenario 1 Order handling in supply chain singly nested

Motivating Scenario 2 Patient Transfer Pathway doubly nested

Our Approach Level 0 Composition using primitive SMDP Level 0 Composition using primitive SMDP Level 1 Composition using composite SMDP Abstract actions

Semi-MDP (SMDP) SMDP = S: set of all states Capture feature-based state space (factored into variables) E.g. isOrderValid  Yes/No/Unknown A: set of actions, some temporally extended Model Web service invocations or operations  Level 0 – action is the invocation of a primitive WS E.g. receiveOrder  Level > 1 – Abstract action is the invocation of a lower-level Web process E.g. verifyOrder T: transition function, T: S X A  Δ(S) Uncertain effects with probabilities E.g. T( inventoryAvail = Yes | Check Inventory Status, inventoryAvail = Unknown) = 0.3 K: lump sum reward/cost, K: S X A  R WS invocation or usage reward/cost F: sojourn time distribution, F: S X A  Δ(t)  Level 0 – uncertain response time of Web services  Level > 1 – uncertain runtime of lower level Web process C: reward/cost accumulating rate, C: S X A  R Reward/cost rate (cost per time unit) of using Web services s0 : initial state

Solving SMDPs Expected Utility of state s: Quantitative measure of immediate effects and long-term effects of actions Where: Solving a SMDP is maximizing the expected utility of each state, Solution techniques include: Value Iteration, Policy Iteration and Linear Programming The solution is a policy: Mapping from states of process environment to actions –More robust than a sequence of Web service invocations Execution of a policy: 1. Determine the current state s 2. Invoke WS given by action a based on the policy 3. Repeat 1-2 until the goal is achieved

Elicitation of Model Parameters Level 0: Model parameters may be obtained from WSDL\SAWSDL, OWL- S descriptions of Web services, and WS-Agreements Level >1: Derive model parameters related to abstract actions from lower level Web process Specifically, we want to know transition function T, lump sum reward K, sojourn time distribution F, and accumulating rate C

Abstract Actions and Variable Correspondence Abstract action “Verify Order”( ) is composed of “Check Customer”(a cc ), “Verify Payment”(a vp ), and “Charge Money”(a cm ) Actions affect only certain variables Correspondence between high-level and lower-level preconditions and effects  For example: High-level variable OrderValid=? Corresponding low-level variable values OrderValid = UCustomerValid=U and PaymentVerified=U and AccountCharged=U OrderValid = YCustomerValid=Y and PaymentVerified=Y and AccountCharged=Y OrderValid = NCustomerValid=N or PaymentVerified=N or AccountCharged=N

Deriving Transition Function T for Abstract Actions Correspondence: E.g. The transition from (OV=U) to (OV=Y) High-level variable OV=?Corresponding low-level variables OrderValid = UCustomerValid=U and PaymentVerified=U and AccountCharged=U OrderValid = YCustomerValid=Y and PaymentVerified=Y and AccountCharged=Y OrderValid = NCustomerValid=N or PaymentVerified=N or AccountCharged=N

Deriving Model Parameters for Abstract Actions Lump sum cost K –lump sum cost of the abstract action is the total of lump sum costs of the corresponding primitive actions

Deriving Model Parameters for Abstract Actions Sojourn time distribution F –Assume the sojourn time of all primitive actions follows Gaussian distribution: F CC : N(t; µ cc, σ cc ), F vo : N(t; µ vo, σ vo ) and F cm : N(t; µ cm, σ cm ) Linear combination of Gaussian distributions is a Gaussian distribution The abstract action VerifyOrder also follows Gaussian F vo : N (t; µ vo, σ vo ) where:

Deriving Model Parameters for Abstract Actions Accumulating Cost Rate C –Accumulated cost of an abstract action is the total accumulated cost of all corresponding primitive actions –Where: E cc (F), E vp (F) and E cm (F) are expected sojourn time Given model parameters for abstract actions, composite SMDP can be solved analogous to a primitive SMDP

System Architecture

BPEL Snippet

Interleaved Generation and Execution of Nested Web Process

Performance Evaluation Methodology − Comparison with HTNs (Wu 2003) on two scenarios − Run the processes generated by two approaches in a simulated environment 1000 times − Measure average reward and standard deviation The performance of HTN approaches ours as the environment becomes less uncertain

Discussion Many AI planning approaches  AI classical planning is not designed to handle WS composition  Assumes deterministic behavior of Web services  Does not scale well to large problems Our hierarchical framework  Stochastic optimization manages uncertainty and delivers optimality  Exploits hierarchy  scalability  Better performance in uncertain environments Future work  Integrate first-order logic to manage state space explosion

Thank You! Questions?

Outline Introduction Motivating scenarios Semi-Markov decision process (SMDP) Composing nested web processes using Hierarchical SMDP System architecture Experiment & Discussion

Composing Web Processes Using H-SMDP