Entropy Rate of a Markov Chain

Slides:



Advertisements
Similar presentations
Generating Random Spanning Trees Sourav Chatterji Sumit Gulwani EECS Department University of California, Berkeley.
Advertisements

Markov Models.
Gibbs sampler - simple properties It’s not hard to show that this MC chain is aperiodic. Often is reversible distribution. If in addition the chain is.
Continuous-Time Markov Chains Nur Aini Masruroh. LOGO Introduction  A continuous-time Markov chain is a stochastic process having the Markovian property.
CS433 Modeling and Simulation Lecture 06 – Part 03 Discrete Markov Chains Dr. Anis Koubâa 12 Apr 2009 Al-Imam Mohammad Ibn Saud University.
Flows and Networks Plan for today (lecture 2): Questions? Continuous time Markov chain Birth-death process Example: pure birth process Example: pure death.
Markov Chains.
Link Analysis: PageRank
Discrete Time Markov Chains
Markov Chains 1.
. Computational Genomics Lecture 7c Hidden Markov Models (HMMs) © Ydo Wexler & Dan Geiger (Technion) and by Nir Friedman (HU) Modified by Benny Chor (TAU)
11 - Markov Chains Jim Vallandingham.
TCOM 501: Networking Theory & Fundamentals
The Rate of Concentration of the stationary distribution of a Markov Chain on the Homogenous Populations. Boris Mitavskiy and Jonathan Rowe School of Computer.
Random Walks Ben Hescott CS591a1 November 18, 2002.
Андрей Андреевич Марков. Markov Chains Graduate Seminar in Applied Statistics Presented by Matthias Theubert Never look behind you…
Entropy Rates of a Stochastic Process
Continuous Time Markov Chains and Basic Queueing Theory
6.896: Probability and Computation Spring 2011 Constantinos (Costis) Daskalakis lecture 2.
Overview of Markov chains David Gleich Purdue University Network & Matrix Computations Computer Science 15 Sept 2011.
More on Rankings. Query-independent LAR Have an a-priori ordering of the web pages Q: Set of pages that contain the keywords in the query q Present the.
What if time ran backwards? If X n, 0 ≤ n ≤ N is a Markov chain, what about Y n = X N-n ? If X n follows the stationary distribution, Y n has stationary.
048866: Packet Switch Architectures Dr. Isaac Keslassy Electrical Engineering, Technion Review.
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
Hidden Markov Models I Biology 162 Computational Genetics Todd Vision 14 Sep 2004.
Kurtis Cahill James Badal.  Introduction  Model a Maze as a Markov Chain  Assumptions  First Approach and Example  Second Approach and Example 
TCOM 501: Networking Theory & Fundamentals
Chapter 4: Stochastic Processes Poisson Processes and Markov Chains
1 TCOM 501: Networking Theory & Fundamentals Lecture 7 February 25, 2003 Prof. Yannis A. Korilis.
1 Markov Chains Algorithms in Computational Biology Spring 2006 Slides were edited by Itai Sharon from Dan Geiger and Ydo Wexler.
Expanders Eliyahu Kiperwasser. What is it? Expanders are graphs with no small cuts. The later gives several unique traits to such graph, such as: – High.
Sampling and Approximate Counting for Weighted Matchings Roy Cagan.
Problems, cont. 3. where k=0?. When are there stationary distributions? Theorem: An irreducible chain has a stationary distribution  iff the states are.
Basic Definitions Positive Matrix: 5.Non-negative Matrix:
Graphs, relations and matrices
CS6800 Advanced Theory of Computation Fall 2012 Vinay B Gavirangaswamy
6. Markov Chain. State Space The state space is the set of values a random variable X can take. E.g.: integer 1 to 6 in a dice experiment, or the locations.
Random Walks and Markov Chains Nimantha Thushan Baranasuriya Girisha Durrel De Silva Rahul Singhal Karthik Yadati Ziling Zhou.
Random Walks and Semi-Supervised Learning Longin Jan Latecki Based on : Xiaojin Zhu. Semi-Supervised Learning with Graphs. PhD thesis. CMU-LTI ,
Markov Decision Processes1 Definitions; Stationary policies; Value improvement algorithm, Policy improvement algorithm, and linear programming for discounted.
1 CS104 : Discrete Structures Chapter V Graph Theory.
6.896: Probability and Computation Spring 2011 Constantinos (Costis) Daskalakis lecture 3.
PageRank. s1s1 p 12 p 21 s2s2 s3s3 p 31 s4s4 p 41 p 34 p 42 p 13 x 1 = p 21 p 34 p 41 + p 34 p 42 p 21 + p 21 p 31 p 41 + p 31 p 42 p 21 / Σ x 2 = p 31.
Markov Chains and Random Walks. Def: A stochastic process X={X(t),t ∈ T} is a collection of random variables. If T is a countable set, say T={0,1,2, …
Markov Chains X(t) is a Markov Process if, for arbitrary times t1 < t2 < < tk < tk+1 If X(t) is discrete-valued If X(t) is continuous-valued i.e.
Random walks on undirected graphs and a little bit about Markov Chains Guy.
Relevant Subgraph Extraction Longin Jan Latecki Based on : P. Dupont, J. Callut, G. Dooms, J.-N. Monette and Y. Deville. Relevant subgraph extraction from.
CS433 Modeling and Simulation Lecture 07 – Part 01 Continuous Markov Chains Dr. Anis Koubâa 14 Dec 2008 Al-Imam.
Seminar on random walks on graphs Lecture No. 2 Mille Gandelsman,
Discrete Time Markov Chains
8.4 Closures of Relations Definition: The closure of a relation R with respect to property P is the relation obtained by adding the minimum number of.
COMS Network Theory Week 5: October 6, 2010 Dragomir R. Radev Wednesdays, 6:10-8 PM 325 Pupin Terrace Fall 2010.
Stochastic Processes and Transition Probabilities D Nagesh Kumar, IISc Water Resources Planning and Management: M6L5 Stochastic Optimization.
ST3236: Stochastic Process Tutorial 6
PERTEMUAN 26. Markov Chains and Random Walks Fundamental Theorem of Markov Chains If M g is an irreducible, aperiodic Markov Chain: 1. All states are.
Great Theoretical Ideas in Computer Science for Some.
Flows and Networks Plan for today (lecture 3): Last time / Questions? Output simple queue Tandem network Jackson network: definition Jackson network: equilibrium.
Theory of Computational Complexity Probability and Computing Lee Minseon Iwama and Ito lab M1 1.
Reliability Engineering
The Monte Carlo Method/ Markov Chains/ Metropolitan Algorithm from sec in “Adaptive Cooperative Systems” -summarized by Jinsan Yang.
Markov Chains and Random Walks
Markov Chains and Mixing Times
Random walks on undirected graphs and a little bit about Markov Chains
Markov Chains Mixing Times Lecture 5
Flows and Networks Plan for today (lecture 4):
Network analysis.
Haim Kaplan and Uri Zwick
Discrete-time markov chain (continuation)
Randomized Algorithms Markov Chains and Random Walks
Connectivity Section 10.4.
Presentation transcript:

Entropy Rate of a Markov Chain For a stationary Markov chain the entropy rate is given by: Where the conditional entropy is computed using the given stationary distribution. Recall that the stationary distribution μ is the solution of the equations: We explicitly express the conditional entropy in the following slide. for all j.

Conditional Entropy Rate for a SMC Theorem (Conditional Entropy rate of a MC): Let {Xi} be a SMC with stationary distribution μ and transition matrix P. Let X1 ~μ. Then the entropy rate is: Proof: Example (Two state MC): The entropy rate of the two state Markov chain in the previous example is: If the Markov chain is irreducible and aperiodic, it has unique stationary distribution on the states, and any initial distribution tends to the stationary distribution as n grows.

Example: ER of Random Walk As an example of stochastic process lets take the example of a random walk on a connected graph. Consider a graph with m nodes with weight Wij≥0 on the edge joining node i with node j. A particle walk randomly from node to node in this graph. The random walk is Xm is a sequence of vertices of the graph. Given Xn=i, the next vertex j is choosen from among the nodes connected to node i with a probability proportional to the weight of the edge connecting i to j. Thus,

ER of a Random Walk In this case the stationary distribution has a surprisingly simple form, which we will guess and verify. The stationary distribution for this MC assigns probability to node i proportional to the total weight of the edges emanating from node i. Let: Be the total weight of edges emanating from node i and let Be the sum of weights of all the edges. Then . We now guess that the stationary distribution is:

ER of Random Walk We check that μP=μ: Thus, the stationary probability of state i is proportional to the weight of edges emanating from node i. This stationary distribution has an interesting property of locality: It depends only on the total weight and the weight of edges connected to the node and therefore it does not change if the weights on some other parts of the graph are changed while keeping the total weight constant. The entropy rate can be computed as follows:

ER of Random Walk If all the edges have equal weight, , the stationary distribution puts weight Ei/2E on node i, where Ei is the number of edges emanating from node i and E is the total number of edges in the graph. In this case the entropy rate of the random walk is: Apparently the entropy rate, which is the average transition entropy, depends only on the entropy of the stationary distribution and the total number of edges

Example Random walk on a chessboard. Let’s king move at random on a 8x8 chessboard. The king has eight moves in the interior, five moves at the edges and three moves at the corners. Using this and the preceding results, the stationary probabilities are, respectively, 8/420, 5/420 and 3/420, and the entropy rate is 0.92log8. The factor of 0.92 is due to edge effects; we would have an entropy rate of log8 on an infinite chessboard. Find the entropy of the other pieces for exercize! It is easy to see that a stationary random walk on a graph is time reversible; that is, the probability of any sequence of states is the same forward or backward: The converse is also true, that is any time reversible Markov chain can be represented as a random walk on an undirected weighted graph.