Shuffling by semi-random transpositions Elchanan Mossel, U.C. Berkeley Joint work with Yuval Peres and Alistair Sinclair.

Slides:



Advertisements
Similar presentations
Optimal Space Lower Bounds for All Frequency Moments David Woodruff MIT
Advertisements

Card shuffling and Diophantine approximation Omer Angel, Yuval Peres, David Wilson Annals of Applied Probability, to appear.
Circuit and Communication Complexity. Karchmer – Wigderson Games Given The communication game G f : Alice getss.t. f(x)=1 Bob getss.t. f(y)=0 Goal: Find.
Chapter Three: Closure Properties for Regular Languages
Gibbs sampler - simple properties It’s not hard to show that this MC chain is aperiodic. Often is reversible distribution. If in addition the chain is.
Learning Juntas Elchanan Mossel UC Berkeley Ryan O’Donnell MIT Rocco Servedio Harvard.
Section 7.4: Closures of Relations Let R be a relation on a set A. We have talked about 6 properties that a relation on a set may or may not possess: reflexive,
Spring 2015 Lecture 5: QuickSort & Selection
On the Spread of Viruses on the Internet Noam Berger Joint work with C. Borgs, J.T. Chayes and A. Saberi.
6.896: Probability and Computation Spring 2011 Constantinos (Costis) Daskalakis lecture 2.
CPSC 668Set 7: Mutual Exclusion with Read/Write Variables1 CPSC 668 Distributed Algorithms and Systems Fall 2009 Prof. Jennifer Welch.
1 List Coloring and Euclidean Ramsey Theory TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A Noga Alon, Tel Aviv.
Data Structures, Spring 2006 © L. Joskowicz 1 Data Structures – LECTURE 4 Comparison-based sorting Why sorting? Formal analysis of Quick-Sort Comparison.
From PET to SPLIT Yuri Kifer. PET: Polynomial Ergodic Theorem (Bergelson) preserving and weakly mixing is bounded measurable functions polynomials,integer.
Complexity 19-1 Complexity Andrei Bulatov More Probabilistic Algorithms.
Dariusz Kowalski University of Connecticut & Warsaw University Andrzej Pelc University of Quebec en Outaouais Broadcasting in Undirected Ad hoc Radio Networks.
The k-server Problem Study Group: Randomized Algorithm Presented by Ray Lam August 16, 2003.
Preference Analysis Joachim Giesen and Eva Schuberth May 24, 2006.
Expanders Eliyahu Kiperwasser. What is it? Expanders are graphs with no small cuts. The later gives several unique traits to such graph, such as: – High.
DAST 2005 Week 4 – Some Helpful Material Randomized Quick Sort & Lower bound & General remarks…
1 Biased card shuffling and the asymmetric exclusion process Elchanan Mossel, Microsoft Research Joint work with Itai Benjamini, Microsoft Research Noam.
Mixing Times of Markov Chains for Self-Organizing Lists and Biased Permutations Prateek Bhakta, Sarah Miracle, Dana Randall and Amanda Streib.
Mixing Times of Self-Organizing Lists and Biased Permutations Sarah Miracle Georgia Institute of Technology.
cover times, blanket times, and majorizing measures Jian Ding U. C. Berkeley James R. Lee University of Washington Yuval Peres Microsoft Research TexPoint.
Ch. 8 & 9 – Linear Sorting and Order Statistics What do you trade for speed?
Asaf Cohen (joint work with Rami Atar) Department of Mathematics University of Michigan Financial Mathematics Seminar University of Michigan March 11,
GROUPS & THEIR REPRESENTATIONS: a card shuffling approach Wayne Lawton Department of Mathematics National University of Singapore S ,
Random Walks and Markov Chains Nimantha Thushan Baranasuriya Girisha Durrel De Silva Rahul Singhal Karthik Yadati Ziling Zhou.
Minimal Spanning Trees What is a minimal spanning tree (MST) and how to find one.
The Polynomial Time Algorithm for Testing Primality George T. Gilbert.
Small subgraphs in the Achlioptas process Reto Spöhel, ETH Zürich Joint work with Torsten Mütze and Henning Thomas TexPoint fonts used in EMF. Read the.
Quantum Computing MAS 725 Hartmut Klauck NTU TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A.
Edge-disjoint induced subgraphs with given minimum degree Raphael Yuster 2012.
1 New Coins from old: Computing with unknown bias Elchanan Mossel, U.C. Berkeley
By J. Burns and J. Pachl Based on a presentation by Irina Shapira and Julia Mosin Uniform Self-Stabilization 1 P0P0 P1P1 P2P2 P3P3 P4P4 P5P5.
Time to Equilibrium for Finite State Markov Chain 許元春(交通大學應用數學系)
Diagonalization and Similar Matrices In Section 4.2 we showed how to compute eigenpairs (,p) of a matrix A by determining the roots of the characteristic.
Cover times, blanket times, and the GFF Jian Ding Berkeley-Stanford-Chicago James R. Lee University of Washington Yuval Peres Microsoft Research.
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
Algorithms 1.Notion of an algorithm 2.Properties of an algorithm 3.The GCD algorithm 4.Correctness of the GCD algorithm 5.Termination of the GCD algorithm.
UNIT 5.  The related activities of sorting, searching and merging are central to many computer applications.  Sorting and merging provide us with a.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Unique Games Approximation Amit Weinstein Complexity Seminar, Fall 2006 Based on: “Near Optimal Algorithms for Unique Games" by M. Charikar, K. Makarychev,
To be presented by Maral Hudaybergenova IENG 513 FALL 2015.
Kevin Stevenson AST 4762/5765. What is MCMC?  Random sampling algorithm  Estimates model parameters and their uncertainty  Only samples regions of.
Advanced Algorithms Analysis and Design
Sorting by placement and Shift Sergi Elizalde Peter Winkler By 資工四 B 周于荃.
Theory of Computational Complexity M1 Takao Inoshita Iwama & Ito Lab Graduate School of Informatics, Kyoto University.
Krishnendu ChatterjeeFormal Methods Class1 MARKOV CHAINS.
Advanced Algorithms Analysis and Design By Dr. Nazir Ahmad Zafar Dr Nazir A. Zafar Advanced Algorithms Analysis and Design.
Theory of Computational Complexity Probability and Computing Chapter Hikaru Inada Iwama and Ito lab M1.
CSCE 668 DISTRIBUTED ALGORITHMS AND SYSTEMS
Markov Chains and Mixing Times
Random walks on undirected graphs and a little bit about Markov Chains
Markov Chains Mixing Times Lecture 5
Section 4.1 Eigenvalues and Eigenvectors
Spectral Clustering.
Reconstruction on trees and Phylogeny 1
Phase Transitions In Reconstruction Yuval Peres, U.C. Berkeley
Markov chain monte carlo
Reconstruction on trees and Phylogeny 2
GROUPS & THEIR REPRESENTATIONS: a card shuffling approach
Hidden Markov Models Part 2: Algorithms
Haim Kaplan and Uri Zwick
Numerical Analysis Lecture 16.
CSCE 668 DISTRIBUTED ALGORITHMS AND SYSTEMS
Compact routing schemes with improved stretch
Clustering.
Locality In Distributed Graph Algorithms
Presentation transcript:

Shuffling by semi-random transpositions Elchanan Mossel, U.C. Berkeley Joint work with Yuval Peres and Alistair Sinclair

Shuffling by random transpositions At each step choose two independent uniformly chosen cards and exchange them. A A

Shuffling by random transpositions Thm[Diaconis-Shahshahani-81]: The mixing time of the random transpositions shuffle is (½ + o(1)) n log n. One can prove an O(n log n) upper bound can using “marking” (more later). Proof of an  (n log n) lower bound: At each step “touch” 2 random cards. Until time (n log n)/4 there are  (n 1/2 ) untouched cards ) permutation is not random.

The cyclic to random shuffle At step i exchange card at location (i mod n) with a uniformly chosen card. A

History of the cyclic to random shuffle Shuffle introduced by Thorp (65). Aldous and Diaconis (86) asked what is the mixing time? Mironov posed again and proved O(n log n) upper bound using marking.

Why do we care? General question: Is systematic scan faster than random update? (other examples: Diaconis-Ram ; Benjamini-Berger- Hoffman-M for asymmetric exclusion; Gaussian fields etc.). Would be nice to find a “natural problem” where the mixing time is strictly between  (n) and  (n log n) Mironov: Cyclic to random may tell us a lot about a widely used crypto algorithm RC4.

The RC4 algorithm Mironov: Let’s study algorithm assuming j is random. Slow mixing corresponds to weak crypto. More than 10 6 hits in google

Upper Bounds - Broder’s Marking Broder’s Marking argument: Call the two pointers L t and R t. Start by marking the first card that is pointed by L 1. At time t, mark card pointed by L t if either: The card at R t is marked or R t = L t.

Broder’s Marking A A LR A R=L

Broder’s marking By induction: Given the time and set of marked cards and their positions, the permutation on the marked cards is uniform. ) The time when all cards are marked is a strong uniform time (permutation is random given the time). In order to prove upper bound, need to bound the “marking time”. For random transpositions easy: By coupon collector estimate this time is O(n log n). Mironov: delicate analysis for cyclic to random.

A general n log n upper bound Thm: [M-Peres-Sinclair] An O(n log n) upper bound on the mixing time holds for any shuffle where: At step t we exchange cards L t and R t where R t are i.i.d. uniform in {0,…,n-1}. The sequence L t is independent of R t. L t can be random, deterministic etc. Cyclic to random is given by L t = t mod n. Top to random is given by L t = 0 for all n. Random transpositions is given by L t i.i.d uniform. Pf: Careful analysis of the marking process.

A general n log n upper bound Proof In more detail: May assume that L t is deterministic. Partition time into intervals of length 2n. In such an interval look at pairs of times s < t such that L s = L t (there are at least n such pairs). We can mark card x if: at time s, x is chosen by R s. R r  L t for s < r < t. R t is one of the marked cards. Letting m i (u i ) be the (un)-marked card at interval i, gives E[u i+1 | F i ] · u i (1 – c m i ) for c > 0. Will skip the rest of the proof. LsLs RsRs x LtLt x RtRt

Cyclic to random shuffle – lower bound? Mironov proved c n lower bound for some c > 1 using parity as a test function: Each shuffle changes the parity with probability (1 – 1/n). After t steps, resulting parity = original parity with probability: Q: Is next to random faster than random transpositions? Note: All cards are touched by time n.

n log n lower bound for cyclic to random shuffle Thm[M-Peres-Sinclair]: The cyclic to random shuffle has a mixing time  (n log n). More precisely: And here is how the proof goes:

Step 1: Homogenizing the chain Problem: The chain is not time homogenous. Can be easily fixed: Consider a chain where at time t:  ( 0) swaped with  (U), where U is uniform. Rotate all cards to the left:  ’(k) =  (k+1 mod n). Clearly chain is equivalent + It is homogenous. From now on study homogenized chain.

One card chain Markov chain for a single card: Eigenvalues satisfy = (1 – 1/n)  where (n-1)  n – n  n = 0. Want to show slow mixing ) want close to 1.

Asymptotics of eigen values and functions = (1 – 1/n)  where (n-1)  n – n  n = 0. Let  -1 = 1 + z/n and get (1+z/n) n – n (1+z/n) + (n-1) = 0 ! e z – z – 1 = 0. Lemma 1: e z – z – 1 has non-zero complex roots. Lemma 2: If  is a root, then M has an eigenvalue such that 1-| | = (1+ <  )/n + O(1/n 2 ). Lemma 3: The eigenvector f corresponding to is “smooth”: |f| 1 · C |f| 2. Will write |f| for either. Pfs: Complex analysis … Remark: Numerically, the smallest non-zero root is  = 2.088… … i

The test function Take f to be an eigenfunction of M corresponding to the eigenvalue closest to 1. Define the test function F Easy: E[f] = 0 ) E[F] = 0. Easy: E[F(id t )] = t |f| 2. A Longer calculation gives: E(F) = 0 E(F(id t )) = t |f| 2 E(F 2 ) = |f| 4 /n

The main Lemma Remains to bound E[|F(id t )| 2 ]. Main Lemma: ) as long as | | 2t ~ ¸ (4t + n)/n 2 the id t and  (where  is uniform) have large total variation distance (2 nd moment method). Since 1 - | | = O(1/n): )  1 ¸  (n log n) E(F) = 0 E(F(id t )) = t |f| 2 E(F 2 ) = |f| 4 /n

Proof of main Lemma The main lemma can be proved using Wilson’s method and the properties of and f. Or it can be done more directly using coupling: Lemma:

Proof of main Lemma Pf idea: “Couple” the following two processes: Process 1: cards i and j move independently. Process 2: The location of cards i and j in the real process. In process 1: Remains to bound the difference between the processes using coupling. Will skip the details …

We’ve seen that the mixing time of the pseudo-random next to random shuffle has the same mixing time as the random transposition shuffle. Proof is not that hard. Problem: How general is the phenomenon? In particular: Open problem: Are there any sequences (deterministic/random) I t, such that the I t to random shuffle mixes in less than n log n time? Conclusion and Open problems