Stochastic Routing Routing Area Meeting IETF 82 (Taipei) Nov.15, 2011.

Slides:



Advertisements
Similar presentations
Research Issues in Web Services CS 4244 Lecture Zaki Malik Department of Computer Science Virginia Tech
Advertisements

Reinforcement Learning
Dialogue Policy Optimisation
Data and Computer Communications
An Introduction to Stochastic Vehicle Routing
Markov Decision Process
CROWN “Thales” project Optimal ContRol of self-Organized Wireless Networks WP1 Understanding and influencing uncoordinated interactions of autonomic wireless.
Eick: Reinforcement Learning. Reinforcement Learning Introduction Passive Reinforcement Learning Temporal Difference Learning Active Reinforcement Learning.
Computational Stochastic Optimization:
Sogang University ICC Lab Using Game Theory to Analyze Wireless Ad Hoc networks.
Decision Making: An Introduction 1. 2 Decision Making Decision Making is a process of choosing among two or more alternative courses of action for the.
Common approach 1. Define space: assign random ID (160-bit) to each node and key 2. Define a metric topology in this space,  that is, the space of keys.
Planning under Uncertainty
Traffic Engineering With Traditional IP Routing Protocols
Reinforcement Learning
December 20, 2004MPLS: TE and Restoration1 MPLS: Traffic Engineering and Restoration Routing Zartash Afzal Uzmi Computer Science and Engineering Lahore.
Quality of Service in IN-home digital networks Alina Albu 23 October 2003.
Optimal Tuning of Continual Online Exploration in Reinforcement Learning Youssef Achbany, Francois Fouss, Luh Yen, Alain Pirotte & Marco Saerens Information.
A General approach to MPLS Path Protection using Segments Ashish Gupta Ashish Gupta.
1 Hybrid Agent-Based Modeling: Architectures,Analyses and Applications (Stage One) Li, Hailin.
Hierarchical Reinforcement Learning Ersin Basaran 19/03/2005.
Planning operation start times for the manufacture of capital products with uncertain processing times and resource constraints D.P. Song, Dr. C.Hicks.
A General approach to MPLS Path Protection using Segments Ashish Gupta Ashish Gupta.
Collaborative Reinforcement Learning Presented by Dr. Ying Lu.
Making Decisions CSE 592 Winter 2003 Henry Kautz.
MATE: MPLS Adaptive Traffic Engineering Anwar Elwalid, et. al. IEEE INFOCOM 2001.
Machine Learning and Optimization For Traffic and Emergency Resource Management. Milos Hauskrecht Department of Computer Science University of Pittsburgh.
Data Communications & Computer Networks
Distributed Quality-of-Service Routing of Best Constrained Shortest Paths. Abdelhamid MELLOUK, Said HOCEINI, Farid BAGUENINE, Mustapha CHEURFA Computers.
1 Capacity Planning under Uncertainty Capacity Planning under Uncertainty CHE: 5480 Economic Decision Making in the Process Industry Prof. Miguel Bagajewicz.
Distributed Constraint Optimization Michal Jakob Agent Technology Center, Dept. of Computer Science and Engineering, FEE, Czech Technical University A4M33MAS.
Reinforcement Learning Evaluative Feedback and Bandit Problems Subramanian Ramamoorthy School of Informatics 20 January 2012.
ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 1 Chapter 1: The DP Algorithm To do:  sequential decision-making  state.
Robot Autonomous Perception Model For Internet-Based Intelligent Robotic System By Sriram Sunnam.
Reinforcement Learning
REINFORCEMENT LEARNING LEARNING TO PERFORM BEST ACTIONS BY REWARDS Tayfun Gürel.
Swarm Intelligence 虞台文.
1 ECE-517: Reinforcement Learning in Artificial Intelligence Lecture 6: Optimality Criterion in MDPs Dr. Itamar Arel College of Engineering Department.
Computational Modeling Lab Wednesday 18 June 2003 Reinforcement Learning an introduction Ann Nowé By Sutton and.
Learning Theory Reza Shadmehr & Jörn Diedrichsen Reinforcement Learning 1: Generalized policy iteration.
Reinforcement Learning 主講人:虞台文 Content Introduction Main Elements Markov Decision Process (MDP) Value Functions.
Energy-Efficient Signal Processing and Communication Algorithms for Scalable Distributed Fusion.
Exact and heuristics algorithms
© D. Weld and D. Fox 1 Reinforcement Learning CSE 473.
Reinforcement Learning 主講人:虞台文 大同大學資工所 智慧型多媒體研究室.
TELE202 Lecture 6 Routing in WAN 1 Lecturer Dr Z. Huang Overview ¥Last Lecture »Packet switching in Wide Area Networks »Source: chapter 10 ¥This Lecture.
Eick: Reinforcement Learning. Reinforcement Learning Introduction Passive Reinforcement Learning Temporal Difference Learning Active Reinforcement Learning.
MDPs (cont) & Reinforcement Learning
Distributed Models for Decision Support Jose Cuena & Sascha Ossowski Pesented by: Gal Moshitch & Rica Gonen.
Reinforcement Learning with Laser Cats! Marshall Wang Maria Jahja DTR Group Meeting October 5, 2015.
به نام خدا سيد عليرضا كارداني مجتبي اميرخاني Path Set Selection in Mobile Ad Hoc Networks زمستان 1382.
Reinforcement Learning
Stochastic Optimization
Reinforcement Learning AI – Week 22 Sub-symbolic AI Two: An Introduction to Reinforcement Learning Lee McCluskey, room 3/10
Energy-Efficient Signal Processing and Communication Algorithms for Scalable Distributed Fusion.
MITRE 7 April 2009 CS 5214 Presenter: Phu-Gui Feng Performance Analysis of Distributed IDS Protocols for Mobile GCS Dr. Jin-Hee Cho, Dr. Ing-Ray Chen MITRE.
Chance Constrained Robust Energy Efficiency in Cognitive Radio Networks with Channel Uncertainty Yongjun Xu and Xiaohui Zhao College of Communication Engineering,
Reinforcement Learning for Mapping Instructions to Actions S.R.K. Branavan, Harr Chen, Luke S. Zettlemoyer, Regina Barzilay Computer Science and Artificial.
Optimization-based Cross-Layer Design in Networked Control Systems Jia Bai, Emeka P. Eyisi Yuan Xue and Xenofon D. Koutsoukos.
Presentation By SANJOG BHATTA Student ID : July 28’ 2009.
COmbining Probable TRAjectories — COPTRA
Learning to Generate Networks
Multi-Agent Exploration
Clearing the Jungle of Stochastic Optimization
Joelle Pineau: General info
Announcements Homework 3 due today (grace period through Friday)
Dr. Unnikrishnan P.C. Professor, EEE
traveling salesman problem
Dr. Arslan Ornek MATHEMATICAL MODELS
Presentation transcript:

Stochastic Routing Routing Area Meeting IETF 82 (Taipei) Nov.15, 2011

Routing Topology modeled as graph G = (V,E,A) –V: vertices and E: edges –A: set of attributes associated to each edge e  E, e.g., residual capacity c, delay d, loss l, etc. Consider set (s 1,t 1 ),...,(s k,t k ) of i src-dst pairs –Associated to each pair (s i,t i ): demand with known non- negative value v i and size r i Routing problem –Find for each unrouted demand (s i,t i ) a routing path from s i to t i for it that maximizes the value of these demands without violating edge attributes –Adaptive routing: routing decisions depend on the instantiated sizes of the previously routed demands

Stochastic Routing Stochastic routing problem in which one or several of the parameters are not deterministic –Demands size are stochastic: probability distribution is specified for the demands –Delay to move between nodes are random variables –(Simultaneous) failure are randomly distributed according to time and space

Key Challenge: routing information and decision-making As in any other stochastic problem, a key issue is: "How do the revelation of information on the uncertain parameters and decision-making (optimization) interact ?" –When do values taken by the uncertain parameters become known ? –What changes can each router (must each router) make on prior-routing decisions on basis of newly obtained information ? => How to make correct local decisions? Each router must know something about global state (inherently large, dynamic, and costly to collect) A routing protocol must intelligently summarize relevant information

Modeling Paradigms (1) Real-time optimization (re-optimization) –Assumption: information is revealed over time as traffic follow their assigned routes/paths (also referred to as dynamic stochastic routing) –Operation: routes are created piece by piece on the basis of the information currently available (at each node) –Approach: dynamic programming

Modeling Paradigms (2) A priori optimization –A solution must be determined beforehand –This solution is “confronted” to the realization of the stochastic parameters in a second step Approaches –Chance-constrained programming: relies on the introduction of probabilistic constraints Pr{total demand assigned to route r ≤ capacity } ≥ 1-α –(Two-stage) stochastic programming with recourse –Robust optimization: uncertainty is represented by an uncertain parameter vector that must belong to a given polyhedral set (without any probability defined) together with, e.g., lower/upper bound for each demand and upper bound on total demand –“Ad hoc” approaches

Learning-based Stochastic Adaptive Routing Reinforcement learning (RL) Objective –Learn what to do--how to map situations (deduced from feedback from the environment) to actions--so as to maximize a numerical reward signal –Learner is not told which actions to take, it must discover which actions yield the most reward by trying them (note: actions may affect not only the immediate reward but also the next situation and, through that, all subsequent rewards) Characteristics –Trial-and-error search Learn from interactions: obtain examples of desired behavior that are both correct and representative Trade-off between exploration and exploitation –Delayed reward

Learning-based Stochastic Adaptive Routing Routing problem multi-agent RL problem –Individual router  (learning) agent which adapts its routing decisions according to rewards/penalty based on Global parameters Non-local parameters (distribution) Local parameters (determined by local observations) c f c = 2 c = 1 d a b e reward/penalty c = 3 s t

Routing Space Distance/Path or Topology (and associated attributes) Detection- Identification- Analysis Rules- Decisions- Parameters (and ext.attributes) Datagrams Call/Session Distributed files/ Information IGP/BGP RSVP ? Routed entities Routing info ?? DHT

Conway's Law is an adage named after computer programmer Melvin Conway, who introduced the idea in 1968:adage computerprogrammerMelvin Conway "...organizations which design systems... are constrained to produce designs which are copies of the communication structures of these organizations."