Random Walks Ben Hescott CS591a1 November 18, 2002.

Slides:



Advertisements
Similar presentations
The Cover Time of Random Walks Uriel Feige Weizmann Institute.
Advertisements

. Markov Chains. 2 Dependencies along the genome In previous classes we assumed every letter in a sequence is sampled randomly from some distribution.
Lecture 6  Calculating P n – how do we raise a matrix to the n th power?  Ergodicity in Markov Chains.  When does a chain have equilibrium probabilities?
CS433 Modeling and Simulation Lecture 06 – Part 03 Discrete Markov Chains Dr. Anis Koubâa 12 Apr 2009 Al-Imam Mohammad Ibn Saud University.
1 The Monte Carlo method. 2 (0,0) (1,1) (-1,-1) (-1,1) (1,-1) 1 Z= 1 If  X 2 +Y 2  1 0 o/w (X,Y) is a point chosen uniformly at random in a 2  2 square.
Markov Chains 1.
11 - Markov Chains Jim Vallandingham.
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
Андрей Андреевич Марков. Markov Chains Graduate Seminar in Applied Statistics Presented by Matthias Theubert Never look behind you…
Entropy Rates of a Stochastic Process
Tutorial 8 Markov Chains. 2  Consider a sequence of random variables X 0, X 1, …, and the set of possible values of these random variables is {0, 1,
6.896: Probability and Computation Spring 2011 Constantinos (Costis) Daskalakis lecture 2.
Overview of Markov chains David Gleich Purdue University Network & Matrix Computations Computer Science 15 Sept 2011.
Markov Chains Lecture #5
Analysis of Network Diffusion and Distributed Network Algorithms Rajmohan Rajaraman Northeastern University, Boston May 2012 Chennai Network Optimization.
CS774. Markov Random Field : Theory and Application Lecture 16 Kyomin Jung KAIST Nov
1 Discrete Structures & Algorithms Graphs and Trees: II EECE 320.
1 Representing Graphs. 2 Adjacency Matrix Suppose we have a graph G with n nodes. The adjacency matrix is the n x n matrix A=[a ij ] with: a ij = 1 if.
Computational problems, algorithms, runtime, hardness
Algorithmic and Economic Aspects of Networks Nicole Immorlica.
1 Mazes In The Theory of Computer Science Dana Moshkovitz.
If time is continuous we cannot write down the simultaneous distribution of X(t) for all t. Rather, we pick n, t 1,...,t n and write down probabilities.
Advanced Topics in Data Mining Special focus: Social Networks.
Expanders Eliyahu Kiperwasser. What is it? Expanders are graphs with no small cuts. The later gives several unique traits to such graph, such as: – High.
Complexity 1 Mazes And Random Walks. Complexity 2 Can You Solve This Maze?
1 On the Computation of the Permanent Dana Moshkovitz.
Sampling and Approximate Counting for Weighted Matchings Roy Cagan.
Ramanujan Graphs of Every Degree Adam Marcus (Crisply, Yale) Daniel Spielman (Yale) Nikhil Srivastava (MSR India)
Problems, cont. 3. where k=0?. When are there stationary distributions? Theorem: An irreducible chain has a stationary distribution  iff the states are.
Accelerating Simulated Annealing for the Permanent and Combinatorial Counting Problems.
Undirected ST-Connectivity In Log Space Omer Reingold Slides by Sharon Bruckner.
Monte Carlo Methods in Partial Differential Equations.
CS6800 Advanced Theory of Computation Fall 2012 Vinay B Gavirangaswamy
6. Markov Chain. State Space The state space is the set of values a random variable X can take. E.g.: integer 1 to 6 in a dice experiment, or the locations.
Entropy Rate of a Markov Chain
Nattee Niparnan. Easy & Hard Problem What is “difficulty” of problem? Difficult for computer scientist to derive algorithm for the problem? Difficult.
1 Random Walks on Graphs: An Overview Purnamrita Sarkar, CMU Shortened and modified by Longin Jan Latecki.
DATA MINING LECTURE 13 Pagerank, Absorbing Random Walks Coverage Problems.
CSE 326: Data Structures NP Completeness Ben Lerner Summer 2007.
6.896: Probability and Computation Spring 2011 Constantinos (Costis) Daskalakis lecture 3.
Many random walks are faster than one Noga AlonTel Aviv University Chen AvinBen Gurion University Michal KouckyCzech Academy of Sciences Gady KozmaWeizmann.
Markov Chains and Random Walks. Def: A stochastic process X={X(t),t ∈ T} is a collection of random variables. If T is a countable set, say T={0,1,2, …
Data Structures & Algorithms Graphs
Graph Colouring L09: Oct 10. This Lecture Graph coloring is another important problem in graph theory. It also has many applications, including the famous.
Random walks on undirected graphs and a little bit about Markov Chains Guy.
Relevant Subgraph Extraction Longin Jan Latecki Based on : P. Dupont, J. Callut, G. Dooms, J.-N. Monette and Y. Deville. Relevant subgraph extraction from.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Seminar on random walks on graphs Lecture No. 2 Mille Gandelsman,
Artur Czumaj DIMAP DIMAP (Centre for Discrete Maths and it Applications) Computer Science & Department of Computer Science University of Warwick Testing.
Discrete Time Markov Chains
Graph theory and networks. Basic definitions  A graph consists of points called vertices (or nodes) and lines called edges (or arcs). Each edge joins.
Strings Basic data type in computational biology A string is an ordered succession of characters or symbols from a finite set called an alphabet Sequence.
Complexity and Efficient Algorithms Group / Department of Computer Science Testing the Cluster Structure of Graphs Christian Sohler joint work with Artur.
Date: 2005/4/25 Advisor: Sy-Yen Kuo Speaker: Szu-Chi Wang.
COMS Network Theory Week 5: October 6, 2010 Dragomir R. Radev Wednesdays, 6:10-8 PM 325 Pupin Terrace Fall 2010.
PERTEMUAN 26. Markov Chains and Random Walks Fundamental Theorem of Markov Chains If M g is an irreducible, aperiodic Markov Chain: 1. All states are.
11. Markov Chains (MCs) 2 Courtesy of J. Bard, L. Page, and J. Heyl.
The Poincaré Constant of a Random Walk in High- Dimensional Convex Bodies Ivona Bezáková Thesis Advisor: Prof. Eric Vigoda.
Theory of Computational Complexity Probability and Computing Lee Minseon Iwama and Ito lab M1 1.
Random Sampling Algorithms with Applications Kyomin Jung KAIST Aug ERC Workshop.
Complexity and Efficient Algorithms Group / Department of Computer Science Testing the Cluster Structure of Graphs Christian Sohler joint work with Artur.
Randomized Algorithms Hung Dang, Zheyuan Gao, Irvan Jahja, Loi Luu, Divya Sivasankaran.
Theory of Computational Complexity Probability and Computing Chapter Hikaru Inada Iwama and Ito lab M1.
Markov Chains and Random Walks
Random walks on undirected graphs and a little bit about Markov Chains
Path Coupling And Approximate Counting
Prof. Paolo Ferragina, Algoritmi per "Information Retrieval"
Randomized Algorithms Markov Chains and Random Walks
Prof. Paolo Ferragina, Algoritmi per "Information Retrieval"
Ilan Ben-Bassat Omri Weinstein
Presentation transcript:

Random Walks Ben Hescott CS591a1 November 18, 2002

Random Walk Definition Given an undirected, connected graph G(V,E) with |V| = n, |E| = m a random “step” in G is a move from some node u to a randomly selected neighbor v. A random walk is a sequence of these random steps starting from some initial node.

Points to note Processes is discrete G is not necessarily planar G is not necessarily fully connected A walk can back in on itself Can consider staying in same place as a move

Questions How many steps to get from u to v How many steps to get back to initial node How many steps to visit every node Easy questions to answer if we consider a simple example

Regular Graphs The expected number of steps to get from vertex u to v in a regular graph is n-1, The expected number of steps to get back to starting point is n for a regular graph. Expected number of steps to visit every node in a regular graph is

Triangle Example Consider probabilities of being at a particular vertex at each step in walk. Each of these can be consider a vector,

Transition Matrix We can use a matrix to represent transition probabilities, consider adjacency matrix A and diagonal matrix, D, with entries 1/d(i) where d(i) is degree of node i. Then we can define matrix M = DA For triangle d(i) = 2 so M = Note for triangle Pr[a to b] = Pr[b to a]

Markov Chains - Generalized Random Walks A Markov Chain is a stochastic process defined on a set of states with matrix of transition probabilities. The process is discrete, namely it is only in one state at given time step (0, 1, 2, …) Next move does not depend on previous moves, formally

Markov Chain Definitions Define vector where the i-th entry is the probability that the chain is in state t Note: Notice that we can then calculate everything given q0 and P.

More Definitions Consider question where am I after t steps, define t-step probability Question is this my first time at node j? Consider probability visit state j at some time t>0 when started at state i. Consider how many steps to get from state i to j. given,otherwise

Even More Definitions Consider f ii State i is called transient f ii < 1 State i is called persistent if f ii = 1 If state i is persistent and h ii is infinite then i is null-persistent If state i is persistent and h ii is not infinite then i is non-null-persistent Turns out every Markov Chain is either transient or non-null-persistent

Almost there A strong component of a directed graph G is a subgraph C of G where for each edge e ij there is a directed path from i to j and from j to i. A Markov Chain is irreducible if underlying graph G consists of a single strong component. A stationary distribution for a Markov Chain with transition matrix P is distribution  s.t. P  =  The periodicity of state i is max int T for which there is a q 0 and a>0 s.t. for all t, if then t is in arithmetic progression A state is periodic if T>1 and aperiodic otherwise An ergodic Markov Chain is one where all states are aperiodic and non-null persistent

Fundamental Theorem of Markov Chains Given any irreducible, finite, aperoidic Markov Chain then all of the following hold –The chain is ergodic –There is a unique stationary distribution  where for –for –Given N(i,t) number of time chain visits state i in t steps then

Random Walk is a Markov Chain Consider G a connected, non-bipartite, undirected graph with |V| = n, |E| = m. There is a corresponding Markov Chain Consider states of chain to be set of vertices Define transitional matrix to be

Interesting facts M G is irreducible since G is connected and undirected Notice the periodicity is the gcd of all cycles in G - closed walk –Smallest walk is 2 go one step and come back –Since G is non-bipartite then there is odd length cycle –GCD is then 1 so M G is aperiodic

Fundamental Theorem Holds So we have a stationary distribution, P  =  But Good news, Also get

Hitting Time Generally h ij, the expected number of steps needed before reaching node j when starting from node i is the hitting time. The commute time is the expected number of steps to reach node j when starting from i and then returning back to i. The commute is bounded by 2m We can express hitting time in terms of commute time as

Lollipop Hitting time from i to j not necessarily same a time from j to i. Consider the kite or lollipop graph. Here

Cover Time How long until we visit every node? The expected number of steps to reach every node in a graph G starting from node i is called the cover time, C i (G) Consider the maximum cover time over all nodes [Matthews]The maximum cover time of any graph with n nodes is (1+1/2+…+1/n) times the maximum hitting time between any two nodes - 2m(n-1)

Coupon Collector Want to collect n different coupons and every day Stop and Shop sends a different coupon at random - how long do have to wait before you can by food? Consider cover time on on a complete graph, here cover time is O(nlgn)

Mixing Rate Going back to probability, could ask how quickly do we converge to the stationary (limiting) distribution? We call this rate the mixing rate of the random walk. We saw How fast

Mixing with the Eigenvalues How do we calculate - yes eigenvalues! Since we are considering probability transitions as a matrix why not use spectral techniques Need graph G to be non-bipartite First need to make sure P is symmetric, not true unless G is regular

More decomposition Need to make P symmetric. Recall that P = DA, where D is diagonal matrix with entries of d(u), degree of node u Consider Claim this let us work with spectral since Now mixing rate is

Graph Connectivity Recall the graph connectivity problem, Given undirected graph G(V,E) want to find out if node s is connected to node t. Can do this in deterministic polynomial time. But what about space, would like to do this in a small amount of space. Recall hitting time of a graph is at most n 3 Try a walk of 2n 3 steps to get there, needs only lg(n) space to count number of steps.

Sampling So what’s the fuss? We can use random walks to sample, since we have the powerful notion of stationary distribution and on a regular graph this is uniform distribution, we can get at random elements. More importantly we can get a random sample from an exponentially large set.

Covering Problems Jerrun, Valiant, and Vazirani - Babai product estimator, or enumeration - self reducibilty Given V a set and V 1, V 2, V 3, …,V m subsets –For all i, |Vi| is polynomial time computable –Can sample from V uniformly at random –For all v in V, can determine efficiently that v is in V i Can we get size of union of subsets, or can we enumerate V in polynomial time

Permanent Want to count the number of perfect matchings in a bipartite graph. This is the permanent of a matrix Given a n x n matrix A the permanent is This is #P-complete

Commercial Break #P is the coolest class, defined as counting class. Consider the number of solutions to the problem Hard - [Toda] PH is contained in #P Want to show hyp-PH is also contained in #P

How to use random walk for permanent approximation Given graph G with d(u) > n/2 want to generate a random perfect matching First notice that input graph is bipartite. Instead consider graph of perfect matchings Let each node be a perfect matching in G - problem is how to connect Need to consider near-perfect matchings a matching with n/2-1 edges

Permanent Cont. Connect near perfect matchings with an edge if they have n/2-2 edges in common and connect a perfect matching to all of the near perfect matchings contained in it to create a graph H. Notice degree of H is bounded by 3n Walk (randomly) a polynomial number of steps in H - if node is a perfect matching - good - otherwise try again.

Volume Want to be able to calculate the volume of a convex body in n dimensions Computing convex polytope is #P-Hard No fear randomization is here Want to be able to fit this problem into enumeration problem

Volume Cont. Given C convex body in n dimensions, assume that C contains the origin Further assume that C contains the unit ball and itself is contained in a ball of radius r<n 3/2. Define C i = intersection of C and ball around origin of radius

Volume Cont. Then we get And we know Vol(C i ) Now only need to be able to get a element uniformly at random - use a walk Difficult to do walk, create a grid and walk from intersections of Ci’s Stationary distribution is not uniform but a distribution with density function proportional to local conductance

Partial Reference List L. Lovasz. Random Walks on Graphs: A survey. Combinatorics Paul Erdos is Eighty Volume 2, p1-46 R. Motwani, P. Raghavan Randomized Algorithms. Cambridge University 1995