Flows and Networks (158052) Richard Boucherie Stochastische Operations Research -- TW wwwhome.math.utwente.nl/~boucherierj/onderwijs/158052/158052.html.

Slides:



Advertisements
Similar presentations
Discrete time Markov Chain
Advertisements

Lecture 6  Calculating P n – how do we raise a matrix to the n th power?  Ergodicity in Markov Chains.  When does a chain have equilibrium probabilities?
CS433 Modeling and Simulation Lecture 06 – Part 03 Discrete Markov Chains Dr. Anis Koubâa 12 Apr 2009 Al-Imam Mohammad Ibn Saud University.
Flows and Networks Plan for today (lecture 2): Questions? Continuous time Markov chain Birth-death process Example: pure birth process Example: pure death.
Markov Chains.
Discrete Time Markov Chains
Topics Review of DTMC Classification of states Economic analysis
TCOM 501: Networking Theory & Fundamentals
G12: Management Science Markov Chains.
Chapter 17 Markov Chains.
IEG5300 Tutorial 5 Continuous-time Markov Chain Peter Chen Peng Adapted from Qiwen Wang’s Tutorial Materials.
Copyright © 2005 Department of Computer Science CPSC 641 Winter Markov Chains Plan: –Introduce basics of Markov models –Define terminology for Markov.
Flows and Networks (158052) Richard Boucherie Stochastische Operations Research -- TW wwwhome.math.utwente.nl/~boucherierj/onderwijs/158052/ html.
1 Part III Markov Chains & Queueing Systems 10.Discrete-Time Markov Chains 11.Stationary Distributions & Limiting Probabilities 12.State Classification.
Андрей Андреевич Марков. Markov Chains Graduate Seminar in Applied Statistics Presented by Matthias Theubert Never look behind you…
048866: Packet Switch Architectures Dr. Isaac Keslassy Electrical Engineering, Technion Review.
TCOM 501: Networking Theory & Fundamentals
Homework 2 Question 2: For a formal proof, use Chapman-Kolmogorov Question 4: Need to argue why a chain is persistent, periodic, etc. To calculate mean.
If time is continuous we cannot write down the simultaneous distribution of X(t) for all t. Rather, we pick n, t 1,...,t n and write down probabilities.
2003 Fall Queuing Theory Midterm Exam(Time limit:2 hours)
Markov Chains Chapter 16.
Stochastic Process1 Indexed collection of random variables {X t } t   for each t  T  X t is a random variable T = Index Set State Space = range.
1 Markov Chains H Plan: –Introduce basics of Markov models –Define terminology for Markov chains –Discuss properties of Markov chains –Show examples of.
CS6800 Advanced Theory of Computation Fall 2012 Vinay B Gavirangaswamy
CDA6530: Performance Models of Computers and Networks Examples of Stochastic Process, Markov Chain, M/M/* Queue TexPoint fonts used in EMF. Read the TexPoint.
6. Markov Chain. State Space The state space is the set of values a random variable X can take. E.g.: integer 1 to 6 in a dice experiment, or the locations.
Flows and Networks Plan for today (lecture 5): Last time / Questions? Waiting time simple queue Little Sojourn time tandem network Jackson network: mean.
Flows and Networks Plan for today (lecture 5): Last time / Questions? Blocking of transitions Kelly / Whittle network Optimal design of a Kelly / Whittle.
Intro. to Stochastic Processes
1 Networks of queues Networks of queues reversibility, output theorem, tandem networks, partial balance, product-form distribution, blocking, insensitivity,
Flows and Networks Plan for today (lecture 6): Last time / Questions? Kelly / Whittle network Optimal design of a Kelly / Whittle network: optimisation.
Networks of Queues Plan for today (lecture 6): Last time / Questions? Product form preserving blocking Interpretation traffic equations Kelly / Whittle.
Flows and Networks Plan for today (lecture 4): Last time / Questions? Output simple queue Tandem network Jackson network: definition Jackson network: equilibrium.
1 Networks of queues Networks of queues reversibility, output theorem, tandem networks, partial balance, product-form distribution, blocking, insensitivity,
Networks Plan for today (lecture 8): Last time / Questions? Quasi reversibility Network of quasi reversible queues Symmetric queues, insensitivity Partial.
1 Networks of queues Networks of queues reversibility, output theorem, tandem networks, partial balance, product-form distribution, blocking, insensitivity,
Markov Chains and Random Walks. Def: A stochastic process X={X(t),t ∈ T} is a collection of random variables. If T is a countable set, say T={0,1,2, …
Markov Chains X(t) is a Markov Process if, for arbitrary times t1 < t2 < < tk < tk+1 If X(t) is discrete-valued If X(t) is continuous-valued i.e.
Chapter 61 Continuous Time Markov Chains Birth and Death Processes,Transition Probability Function, Kolmogorov Equations, Limiting Probabilities, Uniformization.
 { X n : n =0, 1, 2,...} is a discrete time stochastic process Markov Chains.
Chapter 3 : Problems 7, 11, 14 Chapter 4 : Problems 5, 6, 14 Due date : Monday, March 15, 2004 Assignment 3.
Model under consideration: Loss system Collection of resources to which calls with holding time  (c) and class c arrive at random instances. An arriving.
Flows and Networks Plan for today (lecture 2): Questions? Birth-death process Example: pure birth process Example: pure death process Simple queue General.
CS433 Modeling and Simulation Lecture 07 – Part 01 Continuous Markov Chains Dr. Anis Koubâa 14 Dec 2008 Al-Imam.
Flows and Networks Plan for today (lecture 6): Last time / Questions? Kelly / Whittle network Optimal design of a Kelly / Whittle network: optimisation.
CDA6530: Performance Models of Computers and Networks Chapter 3: Review of Practical Stochastic Processes.
Discrete Time Markov Chains
COMS Network Theory Week 5: October 6, 2010 Dragomir R. Radev Wednesdays, 6:10-8 PM 325 Pupin Terrace Fall 2010.
CS433 Modeling and Simulation Lecture 11 Continuous Markov Chains Dr. Anis Koubâa 01 May 2009 Al-Imam Mohammad Ibn Saud University.
11. Markov Chains (MCs) 2 Courtesy of J. Bard, L. Page, and J. Heyl.
Markov Processes What is a Markov Process?
8/14/04J. Bard and J. W. Barnes Operations Research Models and Methods Copyright All rights reserved Lecture 11 – Stochastic Processes Topics Definitions.
Flows and Networks Plan for today (lecture 3): Last time / Questions? Output simple queue Tandem network Jackson network: definition Jackson network: equilibrium.
Flows and Networks Plan for today (lecture 6): Last time / Questions? Kelly / Whittle network Optimal design of a Kelly / Whittle network: optimisation.
Flows and Networks (158052) Richard Boucherie Stochastische Operations Research -- TW wwwhome.math.utwente.nl/~boucherierj/onderwijs/158052/ html.
Networks (3TU) Plan for today (lecture 5): Last time / Questions? Tandem network Jackson network: definition Jackson network: equilibrium distribution.
Networks of queues Networks of queues reversibility, output theorem, tandem networks, partial balance, product-form distribution, blocking, insensitivity,
Flows and Networks Plan for today (lecture 3):
Discrete-time markov chain (continuation)
Much More About Markov Chains
Flows and Networks Plan for today (lecture 4):
IENG 362 Markov Chains.
Markov Chains Carey Williamson Department of Computer Science
TexPoint fonts used in EMF.
Flows and Networks Plan for today (lecture 6):
Networks of queues Networks of queues reversibility, output theorem, tandem networks, partial balance, product-form distribution, blocking, insensitivity,
September 1, 2010 Dr. Itamar Arel College of Engineering
Carey Williamson Department of Computer Science University of Calgary
TexPoint fonts used in EMF.
CS723 - Probability and Stochastic Processes
Presentation transcript:

Flows and Networks (158052) Richard Boucherie Stochastische Operations Research -- TW wwwhome.math.utwente.nl/~boucherierj/onderwijs/158052/ html Introduction to theorie of flows in complex networks: both stochastic and deterministic apects Size 5 ECTS 16 lectures : 8 by R.J. Boucherie focusing on stochastic networks 8 by W. Kern focusing on deterministic networks Common problem How to optimize resource allocation so as to maximize flow of items through the nodes of a complex network Material: handouts / downloads Exam: exercises / (take home) exam References: see website

Motivation and main question Motivation Production / storage system C:\Flexsim Demo\tutorial\Tutorial 3.fsm C:\Flexsim Demo\tutorial\Tutorial 3.fsm Internet Thomas Bonalds's animation of TCP.htm ( www-sop.inria.fr/mistral/personnel/Thomas.Bonald/tcp_eng.html ) trailer Thomas Bonalds's animation of TCP.htm ( www-sop.inria.fr/mistral/personnel/Thomas.Bonald/tcp_eng.html ) trailer Main questions How to allocate servers / capacity to nodes or how to route jobs through the system to maximize system performance, such as throughput, sojourn time, utilization QUESTIONS

Aim: Optimal design of Jackson network Consider an open Jackson network with transition rates Assume that the service rates and arrival rates are given Let the costs per time unit for a job residing at queue j be Let the costs for routing a job from station i to station j be (i) Formulate the design problem (allocation of routing probabilities) as an optimisation problem. (ii) Provide the solution to this problem

Flows and network: stochastic networks Contents 1.Introduction; Markov chains 2.Birth-death processes; Poisson process, simple queue; reversibility; detailed balance 3.Output of simple queue; Tandem network; equilibrium distribution 4.Jackson networks; Partial balance 5.Sojourn time simple queue and tandem network 6.Performance measures for Jackson networks: throughput, mean sojourn time, blocking 7.Application: service rate allocation for throughput optimisation Application: optimal routing

Today: Introduction / motivation course Discrete-time Markov chain Continuous-time Markov chain Next Exercises NOTE: Continuous time Markov chains not covered in first lecture. Continuous time Markov chains will be addressed in second lecture.

Today: Introduction / motivation course Discrete-time Markov chain Continuous-time Markov chain Next Exercises

AEX Continuous, per minute, per day Random process: reason increase / decrease ? Probability level 300 of 400 dec 2004 ? Given level 350 : buy or sell ? Markov chain : random walk

Gambler’s ruin Gambling game: on any turn –Win €1 w.p. p=0.4 –Lose €1 w.p. 1-p=0.6 –Continue to play until €N –If fortune reaches €0 you must stop –X n = amount after n plays –For –X n has the Markov property: conditional probability that given the entire history depends only on –X n is a discrete time Markov chain

Markov chain X n is time-homogeneous Transition probability State space : all possible states For gambler’s ruin For N=5: transition matrix Property

Markov chain : equilibrium distribution n-step transition probability Evaluate: Chapman-Kolmogorov equation n-step transition matrix Initial distribution Distribution at time n Matrix form

Markov chain: classification of states j reachable from i if there exists a path from i to j i and j communicate when j reachable from i and i reachable from j State i absorbing if p(i,i)=1 State i transient if there exists j such that j reachable from i and i not reachable from j Recurrent state i process returns to i infinitely often = non transient state State i periodic with period k>1 if k is smallest number such that all paths from i to i have length that is multiple of k Aperiodic state: recurrent state that is not periodic Ergodic Markov chain: alle states communicate, are recurrent and aperiodic (irreducible, aperiodic)

Markov chain : equilibrium distribution Assume: Markov chain ergodic Equilibrium distribution independent initial state stationary distribution normalising interpretation probability flux

Discrete time Markov chain: summary stochastic process X(t) countable or finite state space S Markov property time homogeneous independent t irreducible: each state in S reachable from any other state in S transition probabilities Assume ergodic (irreducible, aperiodic) global balance equations (equilibrium eqns) solution that can be normalised is equilibrium distribution if equilibrium distribution exists, then it is unique and is limiting distribution

Random walk Gambling game over infinite time horizon: on any turn –Win €1 w.p. p –Lose €1 w.p. 1-p –Continue to play –X n = amount after n plays –State space S = {…,-2,-1,0,1,2,…} –Time homogeneous Markov chain –For each finite time n : –But equilibrium?

Today: Introduction / motivation course Discrete-time Markov chain Continuous-time Markov chain Next Exercises

Today: Introduction / motivation course Discrete-time Markov chain Continuous-time Markov chain Next Exercises

Next time: [R+SN] section 1.1 – 1.3 Continuous – time Markov chains: Birth-death processes; Poisson process, simple queue; reversibility; detailed balance;

Today: Introduction / motivation course Discrete-time Markov chain Continuous-time Markov chain Next Exercises

Exercises: [R+SN] 1.1.2, 1.1.4, Give proof of Chapman-Kolmogorov equation For random walk, let Determine the possible states for N=10, and compute for all feasible j Consider the random walk with reflecting boundary, that has transition probabilities similar to random walk, except in state 0. When the process attempts to jump to the left in state 0, it stays at 0. The transition probabilities are Show that a solution of the global balance equations is For which values of p is this an equilibrium distribution?