Markov Games TCM Conference 2016 Chris Gann

Slides:



Advertisements
Similar presentations
. Markov Chains. 2 Dependencies along the genome In previous classes we assumed every letter in a sequence is sampled randomly from some distribution.
Advertisements

Applied Informatics Štefan BEREŽNÝ
Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc
Operations Research: Applications and Algorithms
. Markov Chains as a Learning Tool. 2 Weather: raining today40% rain tomorrow 60% no rain tomorrow not raining today20% rain tomorrow 80% no rain tomorrow.
. Computational Genomics Lecture 7c Hidden Markov Models (HMMs) © Ydo Wexler & Dan Geiger (Technion) and by Nir Friedman (HU) Modified by Benny Chor (TAU)
Hidden Markov Models Tunghai University Fall 2005.
. Hidden Markov Models - HMM Tutorial #5 © Ydo Wexler & Dan Geiger.
Topics Review of DTMC Classification of states Economic analysis
Eigenvalues and Eigenvectors
Our Group Members Ben Rahn Janel Krenz Lori Naiberg Chad Seichter Kyle Colden Ivan Lau.
Section 10.1 Basic Properties of Markov Chains
Chapter 17 Markov Chains.
Chapter 5 Probability Models Introduction –Modeling situations that involve an element of chance –Either independent or state variables is probability.
Matrices, Digraphs, Markov Chains & Their Use. Introduction to Matrices  A matrix is a rectangular array of numbers  Matrices are used to solve systems.
Operations Research: Applications and Algorithms
MAT 4830 Mathematical Modeling 4.4 Matrix Models of Base Substitutions II
1. Markov Process 2. States 3. Transition Matrix 4. Stochastic Matrix 5. Distribution Matrix 6. Distribution Matrix for n 7. Interpretation of the Entries.
What is the probability that the great-grandchild of middle class parents will be middle class? Markov chains can be used to answer these types of problems.
Matrix Multiplication To Multiply matrix A by matrix B: Multiply corresponding entries and then add the resulting products (1)(-1)+ (2)(3) Multiply each.
Maths for Computer Graphics
1 Markov Chains Tom Finke. 2 Overview Outline of presentation The Markov chain model –Description and solution of simplest chain –Study of steady state.
If time is continuous we cannot write down the simultaneous distribution of X(t) for all t. Rather, we pick n, t 1,...,t n and write down probabilities.
Finite Mathematics & Its Applications, 10/e by Goldstein/Schneider/SiegelCopyright © 2010 Pearson Education, Inc. 1 of 60 Chapter 8 Markov Processes.
Class 25: Question 1 Which of the following vectors is orthogonal to the row space of A?
Linear Equations in Linear Algebra
1 Matrix Addition, C = A + B Add corresponding elements of each matrix to form elements of result matrix. Given elements of A as a i,j and elements of.
. Markov Chains Tutorial #5 © Ilan Gronau. Based on original slides of Ydo Wexler & Dan Geiger.
Linear regression models in matrix terms. The regression function in matrix terms.
Graphs, relations and matrices
Arithmetic Operations on Matrices. 1. Definition of Matrix 2. Column, Row and Square Matrix 3. Addition and Subtraction of Matrices 4. Multiplying Row.
Table of Contents Matrices - Multiplication Assume that matrix A is of order m  n and matrix B is of order p  q. To determine whether or not A can be.
Final Exam Review II Chapters 5-7, 9 Objectives and Examples.
1 Chapter 6 – Determinant Outline 6.1 Introduction to Determinants 6.2 Properties of the Determinant 6.3 Geometrical Interpretations of the Determinant;
Eigenvalues and Eigenvectors
A rectangular array of numbers (we will concentrate on real numbers). A nxm matrix has ‘n’ rows and ‘m’ columns What is a matrix? First column First row.
CH – 11 Markov analysis Learning objectives:
A discrete-time Markov Chain consists of random variables X n for n = 0, 1, 2, 3, …, where the possible values for each X n are the integers 0, 1, 2, …,
When data from a table (or tables) needs to be manipulated, easier to deal with info in form of a matrix. Matrices FreshSophJunSen A0342 B0447 C2106 D1322.
Courtesy of J. Akinpelu, Anis Koubâa, Y. Wexler, & D. Geiger
Copyright (C) 2002 Houghton Mifflin Company. All rights reserved. 1 Understandable Statistics Seventh Edition By Brase and Brase Prepared by: Mistah Flynn.
1 Markov chains and processes: motivations Random walk One-dimensional walk You can only move one step right or left every time unit Two-dimensional walk.
Meeting 18 Matrix Operations. Matrix If A is an m x n matrix - that is, a matrix with m rows and n columns – then the scalar entry in the i th row and.
Class 26: Question 1 1.An orthogonal basis for A 2.An orthogonal basis for the column space of A 3.An orthogonal basis for the row space of A 4.An orthogonal.
How to Multiply Two Matrices. Steps for Matrix Multiplication 1.Determine whether the matrices are compatible. 2.Determine the dimensions of the product.
Matrices: Simplifying Algebraic Expressions Combining Like Terms & Distributive Property.
Matrices Digital Lesson. Copyright © by Houghton Mifflin Company, Inc. All rights reserved. 2 A matrix is a rectangular array of real numbers. Each entry.
Copyright © Cengage Learning. All rights reserved. 2 SYSTEMS OF LINEAR EQUATIONS AND MATRICES.
10.1 Properties of Markov Chains In this section, we will study a concept that utilizes a mathematical model that combines probability and matrices to.
Stochastic Processes and Transition Probabilities D Nagesh Kumar, IISc Water Resources Planning and Management: M6L5 Stochastic Optimization.
11. Markov Chains (MCs) 2 Courtesy of J. Bard, L. Page, and J. Heyl.
Goldstein/Schnieder/Lay: Finite Math & Its Applications, 9e 1 of 60 Chapter 8 Markov Processes.
Business Modeling Lecturer: Ing. Martina Hanová, PhD.
A rectangular array of numeric or algebraic quantities subject to mathematical operations. The regular formation of elements into columns and rows.
From DeGroot & Schervish. Example Occupied Telephone Lines Suppose that a certain business office has five telephone lines and that any number of these.
2.1 Matrix Operations 2. Matrix Algebra. j -th column i -th row Diagonal entries Diagonal matrix : a square matrix whose nondiagonal entries are zero.
Properties and Applications of Matrices
Linear Equations in Linear Algebra
V5 Stochastic Processes
Matrix Multiplication
Markov Chains Tutorial #5
PageRank and Markov Chains
Linear Equations in Linear Algebra
2. Matrix Algebra 2.1 Matrix Operations.
Courtesy of J. Akinpelu, Anis Koubâa, Y. Wexler, & D. Geiger
Objectives Multiply two matrices.
Markov Chains Tutorial #5
Discrete-time markov chain (continuation)
Random Processes / Markov Processes
Eigenvalues and Eigenvectors
Presentation transcript:

Markov Games TCM Conference 2016 Chris Gann

A taxi company has divided the city into three regions – Northside, Downtown, and Southside. By keeping track of pickups and deliveries, the company has found that of the fares picked up in Northside, 50% stay in that region, 20% are taken Downtown, and 30% go to Southside. Of the fares picked up Downtown, only 10% go to Northside, 40% stay Downtown, and 50% go to Southside. Of the fares picked up in Southside, 30% go to each of Northside and Downtown, while 40% stay in Southside. We would like to know what the distribution of taxis will be over time as they pick up and drop off successive fares. Suppose we want to know the probability that a taxi starting off Downtown, will be Downtown after letting off its seventh fare? The Taxi Problem

State Diagram This information can be represented in a state diagram which includes the three states D, N, and S corresponding to the three regions of the city the probabilities of a taxi transitioning from one region/state to another

Markov Chains These probabilities are constant and independent of previous behavior – this memorylessness of the system is called the Markov property. We assume that a transition – picking up and dropping off a fare – occurs each time the system is observed, and that observations occur at regular intervals. Systems with these characteristics are called Markov chains or Markov processes.

What is the probability that a taxi that starts off Downtown ends up in Northside after two fares? One possibility is that the taxi stays Downtown for the first fare, and then transitions to Northside for the second. The probability of this occurring is: But we could also have the taxi going to either Northside of Southside first, then transitioning to Northside: Computing the Probabilities

Since the taxi could follow the first, second or third path, the probability of starting and ending Downtown after two fares is But we could also have the taxi going to either Northside of Southside first, then transitioning to Northside: Computing the Probabilities

If we want to know the probability of a taxi transitioning from one region to another after just three fares, the computation would have more possible paths. Suppose we were interested in the probability of a taxi both starting and ending up Downtown. We can use a tree diagram to represent this calculation. If we multiply along all of the paths and sum the results we find that this probability is Tree Diagram

However, we originally asked what would be the probability that a taxi starting Downtown is back Downtown after seven fares. Considering the tree diagram for only three transitions, trying to compute this probability using a tree diagram would be unwieldy. Let’s look again at our first calculation – the probability of transitioning from D to N in two transitions. This has the structure of an inner product. That is it can be represented as the dot product of two vectors, or equivalently, the product of a row and a column of two matrices. Matrix Multiplication

Labeling these with matrices with D, S, and N indicates which probability is found in each entry. That is, each entry of these two simple matrices is the probability of going from the row label to the column label in one transition. Developing the Transition Matrix

We can create a square matrix, T, called the transition matrix, by constructing rows for the probabilities going from Southside and Northside as well. An entry T ij of this matrix is the probability of a transition from region i to region j. For example, T 32, is the probability of a fare that originates in Northside going to Southside. (Note the sum of entries across rows of T.) The Transition Matrix

What results when we multiply the transition matrix by itself? The highlighted entry results from the same computation that we already considered for of a taxi going from D to N in two fares. What are the other entries of ? What are the entries of ? ? The Transition Matrix

Some general features of a transition matrix for a Markov chain can be observed from this problem: 1.A transition matrix is square. Clearly the number of rows and the number of columns are both the same as the number of states. 2.All entries are between 0 and 1. This follows from the fact that the entries correspond to transition probabilities from one state to another. 3.The sum of the entries in any row must be 1. The sum of the entries in an entire row is the sum of the transition probabilities from one state to all other states. Since a transition is sure to take place, this sum must be 1. 4.The ij entry in the matrix gives the probability of being in state j after n transitions, with state i as the initial state. 5.The entries in the transition matrix are constant. A Markov chain model depends upon the assumption that the transition matrix does not change throughout the process.

Hi Ho Cherry O – A Markov Game A player starts with 0 cherries. A spinner determines how many cherries you gain or lose each turn. If the spinner lands on 1, 2, 3, or 4 cherries then you gain that many cherries. If the spinner lands on the bird or dog then you lose 2 cherries (or lose all your cherries if you have 2 or less). If the spinner lands on the spilled basket then you lose all your cherries. You must collect 10 or more cherries to win the game.