Random walks on undirected graphs and a little bit about Markov Chains

Slides:



Advertisements
Similar presentations
The Cover Time of Random Walks Uriel Feige Weizmann Institute.
Advertisements

Great Theoretical Ideas in Computer Science
Theory of Computing Lecture 18 MAS 714 Hartmut Klauck.
Approximation, Chance and Networks Lecture Notes BISS 2005, Bertinoro March Alessandro Panconesi University La Sapienza of Rome.
Gibbs sampler - simple properties It’s not hard to show that this MC chain is aperiodic. Often is reversible distribution. If in addition the chain is.
Random Walks Ben Hescott CS591a1 November 18, 2002.
Entropy Rates of a Stochastic Process
Great Theoretical Ideas in Computer Science.
1 Discrete Structures & Algorithms Graphs and Trees: II EECE 320.
Algorithmic and Economic Aspects of Networks Nicole Immorlica.
1 Optimization problems such as MAXSAT, MIN NODE COVER, MAX INDEPENDENT SET, MAX CLIQUE, MIN SET COVER, TSP, KNAPSACK, BINPACKING do not have a polynomial.
Complexity 19-1 Complexity Andrei Bulatov More Probabilistic Algorithms.
Computability and Complexity 24-1 Computability and Complexity Andrei Bulatov Approximation.
Expanders Eliyahu Kiperwasser. What is it? Expanders are graphs with no small cuts. The later gives several unique traits to such graph, such as: – High.
1 On the Computation of the Permanent Dana Moshkovitz.
Entropy Rate of a Markov Chain
Random Walks and Markov Chains Nimantha Thushan Baranasuriya Girisha Durrel De Silva Rahul Singhal Karthik Yadati Ziling Zhou.
The Theory of NP-Completeness 1. Nondeterministic algorithms A nondeterminstic algorithm consists of phase 1: guessing phase 2: checking If the checking.
The Theory of NP-Completeness 1. What is NP-completeness? Consider the circuit satisfiability problem Difficult to answer the decision problem in polynomial.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
Expanders via Random Spanning Trees R 許榮財 R 黃佳婷 R 黃怡嘉.
Markov Chains and Random Walks. Def: A stochastic process X={X(t),t ∈ T} is a collection of random variables. If T is a countable set, say T={0,1,2, …
NP-Complete problems.
Random walks on undirected graphs and a little bit about Markov Chains Guy.
Seminar on random walks on graphs Lecture No. 2 Mille Gandelsman,
Presented by Dajiang Zhu 11/1/2011.  Introduction of Markov chains Definition One example  Two problems as examples 2-SAT Algorithm (simply introduce.
NPC.
PERTEMUAN 26. Markov Chains and Random Walks Fundamental Theorem of Markov Chains If M g is an irreducible, aperiodic Markov Chain: 1. All states are.
The Theory of NP-Completeness 1. Nondeterministic algorithms A nondeterminstic algorithm consists of phase 1: guessing phase 2: checking If the checking.
Given this 3-SAT problem: (x1 or x2 or x3) AND (¬x1 or ¬x2 or ¬x2) AND (¬x3 or ¬x1 or x2) 1. Draw the graph that you would use if you want to solve this.
COSC 3101A - Design and Analysis of Algorithms 14 NP-Completeness.
Theory of Computational Complexity Probability and Computing Lee Minseon Iwama and Ito lab M1 1.
1 Lecture 5 (part 2) Graphs II (a) Circuits; (b) Representation Reading: Epp Chp 11.2, 11.3
Random Sampling Algorithms with Applications Kyomin Jung KAIST Aug ERC Workshop.
CHAPTER SIX T HE P ROBABILISTIC M ETHOD M1 Zhang Cong 2011/Nov/28.
Krishnendu ChatterjeeFormal Methods Class1 MARKOV CHAINS.
Theory of Computational Complexity Probability and Computing Chapter Hikaru Inada Iwama and Ito lab M1.
TU/e Algorithms (2IL15) – Lecture 11 1 Approximation Algorithms.
ICS 353: Design and Analysis of Algorithms NP-Complete Problems King Fahd University of Petroleum & Minerals Information & Computer Science Department.
An Introduction to Graph Theory
The Theory of NP-Completeness
The NP class. NP-completeness
NP-Completeness (2) NP-Completeness Graphs 4/13/2018 5:22 AM x x x x x
More NP-complete problems
P & NP.
Algorithms for Big Data: Streaming and Sublinear Time Algorithms
Lap Chi Lau we will only use slides 4 to 19
Markov Chains and Random Walks
Topics in Algorithms Lap Chi Lau.
More NP-complete Problems
Markov Chains Mixing Times Lecture 5
NP-Completeness (2) NP-Completeness Graphs 7/23/ :02 PM x x x x
NP-Completeness Proofs
Graph theory Definitions Trees, cycles, directed graphs.
Great Theoretical Ideas in Computer Science
Greedy Algorithms / Minimum Spanning Tree Yin Tat Lee
NP-Completeness Yin Tat Lee
Discrete Mathematics for Computer Science
ICS 353: Design and Analysis of Algorithms
NP-Completeness (2) NP-Completeness Graphs 11/23/2018 2:12 PM x x x x
Haim Kaplan and Uri Zwick
Randomized Algorithms Markov Chains and Random Walks
3.5 Minimum Cuts in Undirected Graphs
NP-Complete Problems.
Graphs and Algorithms (2MMD30)
NP-Completeness Yin Tat Lee
The Theory of NP-Completeness
Instructor: Aaron Roth
NP-Completeness (2) NP-Completeness Graphs 7/9/2019 6:12 AM x x x x x
Presentation transcript:

Random walks on undirected graphs and a little bit about Markov Chains Guy

One dimensional Random walk A random walk on a line. A line is n 1 2 3 4 5 If the walk is on 0 it goes into 1. Else it goes to i+1 or to i-1 with probability 1/2 What is the expected number of steps to go to n?

The expected time function T(n)=0 T(i)=1+(T(i+1)+T(i-1))/2, i0 T(0)=1+T(1) Add all equation gives T(n-1)=2n-1. From that we get: T(n-2)=4n-4 T(i)=2(n-i)n-(n-i)2 T(0)=n2

2-SAT: definition A 2-CNF formula: (x+ ¬y)& (¬ x+ ¬ z) &(x+y) &(z+¬ x) (x+y) and the others are called CLAUSES CNF means that we want to satisfy all of them (x+ ¬y) is satisfied if X=T or y=F. The question: is there a satisfying assignment? 2n is not a number the computer can run even for n=70

Remark: the case of two literals very special If three literals per clause: NPC Not only that, but can not be approximated better than 7/8 Obviously for (x+¬y+¬z) if we draw a random value the probability that the clause is satisfied is 1-1/8=7/8. Best approximation possible Also hard 3-SAT(5)

The random algorithm for 2-SAT 2-SAT : Start with an arbitrary assignment Let C be a non satisfied clause. Choose at random one of the two literals of C and flip its value. We know that if the variables are x1 and x2 the optimum disagrees with us on x1 or on x2. Distance to OPT: with probability ½ smaller by 1 and with probability ½ larger by 1 (worst case). Thus E(RW)≤n2

RP algorithm (can make a mistake) If you do these changes 2n2 times the probability that we do not get a truth assigment is ½ if one exists You can do n4 and the probability that if there is a truth assignment we don’t find it is 1/n2 What we use here is called the Markov inequality that states there are at most 1/3 of the numbers in a collection of numbers that are at least 3 times the average There are several deterministic algorithms for 2-SAT

Shuffling cards Take the top card and place it with a random place (including first place). One of n. A simple claim: if a card is in a random place it will be in a random place after next move. We know Pr(card is in place i)=1/n For the card to be in i after next step three possibilities

Probability of the card to be in place i One possibility: it is not the first card and it is in place i. Chooses one of i-1 first places. Second possibility (disjoint events) its at place 1 and exchanged with i. Third possibility: it is in place i+1 and the first card is placed in one of the places after i+1 1/n*(i-1)/n+1/n*1/n+1/n*1/(n-i)=1/n

Stopping time If all the cards have been upstairs its random Check the lowes card. To go up by 1 G(1/n) thus expectation n. To go from second last to third last G(2/n) and expectation n/2. This gives n+n/2+n/3+….= n(ln n+Ө(1)) FAST

Random Walks on undirected graphs Given a graph choose a neigbor at random with probability 1/d(v)

Random Walks Given a graph choose a vertex at random.

Random Walks Given a graph choose a vertex at random.

Random Walks Given a graph choose a vertex at random.

Random Walks Given a graph choose a vertex at random.

Random Walks Given a graph choose a vertex at random.

Random Walks Given a graph choose a vertex at random.

Markov chains Generalization of random walks on undirected graph. The graph is directed The sum over the values of outgoing edges is 1 but it does not need to be uniform. We have a matrix P=(pij) with pij is the probability that on state i it will go to j. Say that for Π0 =(x1,x2,….,xn)

Changing from state to state Probability(state=i)=j xj * pji This is the inner product of the current state and column j of the matrix. Therefore if we are in distribution Π0 after one step the distribution is at state: Π1= Π0 *P And after i stages its in Πi= Π0 *Pi What is a steady state?

Steady state Steady state is Π so that: Π*P= Π. For any i and any round the probability that it is in i is the same always. Conditions for convergence: The graph has to be strongly connected. Otherwise there may be many components that have no edges out. No steady state. hii the time to get back to i from i is finite Non periodic. Slightly complex for Markov chains. For random walks on undirected graphs: not a bipartite graph.

The bipartite graph example If the graph is bipartite and V1 and V2 are its sets then if we start with V1 we can not be on V1 vertex in after odd number of transitions. Therefore a steady state is not possible. So we need for random walks on graphs that the graph is connected and not bipartite. The other property will follow.

Fundamental theorem of Markov chains Theorem: Given an aperiodic MC so that hii is not infinite for any i, and non-reducible Markov chain than: 1) There is a unique steady state. Thus to find the steady state just find Π*P= Π 2) hii=1/Πi . Geometric distribution argument. Remark: the mixing time is how fast the chain gets to (very close to) the steady state.

Because its unique you just have to find the correct Π For random walks in an undirected graph we claim that the steady state is: (2d1/m,2d2/m,……,2dm/m) It is trivial to show that this is the steady state. Multiply this vector and the i column. The only important ones are neighbors of i. So sum(j,i)E 1/dj*(2dj/m)=2di/m

The expected time to visit all the vertices of a graph A matrix is doubly stochastic if and only if all columns also sum to 1. Exercise (very simple): doubly stochastic then {1/n} or uniform is the steady state. Define a new Markov chain of edges with directions which means 2m states. The walk is defined naturally. Exercise: show that the Matrix is doubly stochastic

The time we spend on every edge By the above, we spend the same time on every edge in the two directions over all edges (of course you have to curve the noise. We are talking on a limit here). Now, say that I want to bound hij the expected time we get from i to j.

Showing hij+hji≤ 2m Assume that we start at i--->j By the Markov chain of the edges it will take 2m steps until we do this move again Forget about how you got to j (no memory). Since we are doing now i---->j again we know that: a) As it was in j, it returned to i. This is half the inequality hji b) Now it goes i----> j. This takes at most hij c) Since this takes at most 2m the claim follows.

Consider a spanning tree and a walk on the spanning tree Choose any paths that traverses the tree so that the walk goes from a parent to a child once and back once. Per parent child we have hij+hji≤2m Thus over the n-1 edges of the graph the cover time is at most 2m(n-1)<n3

Tight example: the clique n/2 vertices u1 un/2 u2

Bridges For bridges hij+hji=2m (follows from proof). Say you are the intersection vertex u1 of the clique and of the path. Its harder to go right than left. Thus it takes about n2 time to go from u1 to u2 and the same time to go from u2 to u3 etc. This gives Ω(n3) Funny name: Lollipop graph.

Exponential cover time for directed graphs

Universal sequences Say that the graph is d-regular. And the edges of every vertex are denoted by 1,2,….,d. A walks series for d=20 would be 7,14,12,19,3,4,8,6…….. The question, will it cover all graphs? We show that if the series is large enough it covers all graphs.

The length of the sequence Let L=4dn2(dn+2)(log n+1) What is the probability that a graph is covered? Let N=4dn2, and think of the above as (dn+2)(ln n+1) different parts of sequences of length N. By previous proof a graph will be cover in expectation in 2n2d steps. With probability at most ½ it is not covered in N steps.

The number of d regular graphs From n2 possible edges choose n*d Thus we have to choose n*d elements and this is bounded by n2n*d graphs. The probability that a graph is not covered is exp(-nd*lg n) The union bound implies that with large probability all graphs are covered.

One advise for every n And with this one advice we can cover all graphs with high probability. Thus we can find out for example if s and t are in the same connected components. Later it was proved that with O(log n) space you can check if s and t are in the same connected component!! Even though you can remember O(1) items only.