An Improved Liar Game Strategy From a Deterministic Random Walk

Slides:



Advertisements
Similar presentations
Another question consider a message (sequence of characters) from {a, b, c, d} encoded using the code shown what is the probability that a randomly chosen.
Advertisements

Ulams Game and Universal Communications Using Feedback Ofer Shayevitz June 2006.
Noise, Information Theory, and Entropy (cont.) CS414 – Spring 2007 By Karrie Karahalios, Roger Cheng, Brian Bailey.
Error Control Code.
Gillat Kol (IAS) joint work with Ran Raz (Weizmann + IAS) Interactive Channel Capacity.
“Ulam‘s” Liar Game with Lies in an Interval Benjamin Doerr (MPI Saarbrücken, Germany) joint work with Johannes Lengler and David Steurer (Universität des.
Games of Prediction or Things get simpler as Yoav Freund Banter Inc.
Information Theoretical Security and Secure Network Coding NCIS11 Ning Cai May 14, 2011 Xidian University.
Information Theory Introduction to Channel Coding Jalal Al Roumy.
Day 2 Information theory ( 信息論 ) Civil engineering ( 土木工程 ) Cultural exchange.
Asymptotic Enumerators of Protograph LDPCC Ensembles Jeremy Thorpe Joint work with Bob McEliece, Sarah Fogal.
Coding Theory: Packing, Covering, and 2-Player Games Robert Ellis Menger Day 2008: Recent Applied Mathematics Research Advances April 14, 2008.
The Goldreich-Levin Theorem: List-decoding the Hadamard code
2/28/03 1 The Virtues of Redundancy An Introduction to Error-Correcting Codes Paul H. Siegel Director, CMRR University of California, San Diego The Virtues.
DANSS Colloquium By Prof. Danny Dolev Presented by Rica Gonen
Adaptive Coding from a Diffusion Process on the Integer Line Robert Ellis October 26, 2009 Joint work with Joshua Cooper, University of South Carolina.
A 2-player game for adaptive covering codes Robert B. Ellis Texas A&M coauthors: Vadim Ponomarenko, Trinity University Catherine Yan, Texas A&M.
15-853Page :Algorithms in the Real World Error Correcting Codes I – Overview – Hamming Codes – Linear Codes.
Mario Vodisek 1 HEINZ NIXDORF INSTITUTE University of Paderborn Algorithms and Complexity Erasure Codes for Reading and Writing Mario Vodisek ( joint work.
Rényi-Ulam liar games with a fixed number of lies Robert B. Ellis Illinois Institute of Technology University of Illinois at Chicago, October 26, 2005.
Hamming Codes 11/17/04. History In the late 1940’s Richard Hamming recognized that the further evolution of computers required greater reliability, in.
Games, Hats, and Codes Mira Bernstein Wellesley College SUMS 2005.
©2003/04 Alessandro Bogliolo Background Information theory Probability theory Algorithms.
Identification of strategies for liar-type games via discrepancy from their linear approximations Robert Ellis October 14 th, 2011 AMS Sectional Meeting,
The Hat Game 11/19/04 James Fiedler. References Hendrik W. Lenstra, Jr. and Gadiel Seroussi, On Hats and Other Covers, preprint, 2002,
The Multiplicative Weights Update Method Based on Arora, Hazan & Kale (2005) Mashor Housh Oded Cats Advanced simulation methods Prof. Rubinstein.
An Improved Liar Game Strategy From a Deterministic Random Walk Robert Ellis February 22 nd, 2010 Peled Workshop, UIC Joint work with Joshua Cooper, University.
Information and Coding Theory Linear Block Codes. Basic definitions and some examples. Juris Viksna, 2015.
Information Coding in noisy channel error protection:-- improve tolerance of errors error detection: --- indicate occurrence of errors. Source.
Threshold Phenomena and Fountain Codes Amin Shokrollahi EPFL Joint work with M. Luby, R. Karp, O. Etesami.
Error Control Code. Widely used in many areas, like communications, DVD, data storage… In communications, because of noise, you can never be sure that.
1 Private codes or Succinct random codes that are (almost) perfect Michael Langberg California Institute of Technology.
Information Theory Linear Block Codes Jalal Al Roumy.
Channel Coding Binit Mohanty Ketan Rajawat. Recap…  Information is transmitted through channels (eg. Wires, optical fibres and even air)  Channels are.
10.1 Chapter 10 Error Detection and Correction Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
Error Detection. Data can be corrupted during transmission. Some applications require that errors be detected and corrected. An error-detecting code can.
Some Computation Problems in Coding Theory
Raptor Codes Amin Shokrollahi EPFL. BEC(p 1 ) BEC(p 2 ) BEC(p 3 ) BEC(p 4 ) BEC(p 5 ) BEC(p 6 ) Communication on Multiple Unknown Channels.
Error Detecting and Error Correcting Codes
An identity for dual versions of a chip-moving game Robert B. Ellis April 8 th, 2011 ISMAA 2011, North Central College Joint work with Ruoran Wang.
Online Learning Model. Motivation Many situations involve online repeated decision making in an uncertain environment. Deciding how to invest your money.
An identity for dual versions of a chip-moving game Robert B. Ellis April 8 th, 2011 ISMAA 2011, North Central College Joint work with Ruoran Wang.
8 Coding Theory Discrete Mathematics: A Concept-based Approach.
An Improved Liar Game Strategy From a Deterministic Random Walk
Hamming codes. Golay codes.
The Viterbi Decoding Algorithm
Game Theory Just last week:
Error Detection and Correction
: An Introduction to Computer Networks
Advanced Computer Networks
15-853:Algorithms in the Real World
Chapter 10 Error Detection And Correction
Scalar Quantization – Mathematical Model
Chapter 6.
COT 5611 Operating Systems Design Principles Spring 2012
COT 5611 Operating Systems Design Principles Spring 2014
RS – Reed Solomon List Decoding.
EEC-484/584 Computer Networks
Dr. Clincy Professor of CS
Information Redundancy Fault Tolerant Computing
RAID Redundant Array of Inexpensive (Independent) Disks
Chapter Nine: Data Transmission
Information-Theoretic Security
Error Detection and Correction
Parameterized Complexity of Even Set (and others)
The Selection Problem.
Types of Errors Data transmission suffers unpredictable changes because of interference The interference can change the shape of the signal Single-bit.
Error Correction Coding
Watermarking with Side Information
An Improved Liar Game Strategy From a Deterministic Random Walk
Presentation transcript:

An Improved Liar Game Strategy From a Deterministic Random Walk Robert Ellis April 9th, 2010 San Diego State University Joint work with Joshua Cooper, University of South Carolina

Outline of Talk Coding theory overview Diffusion processes on Z 2 Outline of Talk Coding theory overview Packing (error-correcting) & covering codes Coding as a 2-player game Liar game and pathological liar game Diffusion processes on Z Simple random walk (linear machine) Liar machine Pathological liar game, alternating question strategy Improved pathological liar game bound Reduction to liar machine Discrepancy analysis of liar machine versus linear machine Concluding remarks

Coding Theory Overview 3 Coding Theory Overview Codewords: fixed-length strings from a finite alphabet Primary uses: Error-correction for transmission in the presence of noise Compression of data with or without loss Viewpoints: Packings and coverings of Hamming balls in the hypercube 2-player perfect information games Applications: Cell phones, compact disks, deep-space communication

Coding Theory: (n,e)-Codes 4 Coding Theory: (n,e)-Codes Transmit blocks of length n Noise changes ≤ e bits per block (||||1 ≤ e) Repetition code 111, 000 length: n = 3 e = 1 information rate: 1/3  x1…xn (x1+1)…(xn+ n) Richard Hamming 110 010 000 101 111 Received: Decoded: blockwise majority vote (After second bullet) The Hamming assumption is that noise is “adversarial”; that is, an adversary may choose up to e bits to invert in the binary symmetric channel. A code can be successful with high probability under the Shannon assumption (usually there is a unique decoding), but can fail under the Hamming assumption (there exists a situation with non-unique decoding).

Coding Theory – A Hamming (7,1)-Code 5 Coding Theory – A Hamming (7,1)-Code Length n=7, corrects e=1 error 1 received 1 1 1 1 1 3 errors: incorrect decoding 1 1 error: correct decoding Encoding and decoding is automated by using a codeword generator matrix and a decoding parity-check matrix. The 3 right-most bits are parity-check bits. They are defined as the sum of the first 4 bits with no bits left out, the 2nd, 3rd, and 4th bits left out, respectively. decoded

A Repetition Code as a Packing 6 A Repetition Code as a Packing (3,1)-code: 111, 000 Pairwise distance = 3  1 error can be corrected The M codewords of an (n,e)-code correspond to a packing of Hamming balls of radius e in the n-cube 110 011 101 111 111 110 101 011 000 010 001 100 100 010 001 000 ((3,1)-repetition code) A packing of 2 radius-1 Hamming balls in the 3-cube

A (5,1)-Packing Code as a 2-Player Game 7 A (5,1)-Packing Code as a 2-Player Game (5,1)-code: 11111, 10100, 01010, 00001 Paul Carole 11111 What is the 1st bit? 10100 What is the 2nd bit? 01010 00001 What is the 3rd bit? What is the 4th bit? 1 1 >1 What is the 5th bit? # errors Paul represents the receiver, waiting to get the bits of the message. Carole is a combination of the sender and the noise. If we assume that Carole sends at most one error, we can track the possibility of each codeword being the intended message. Because the (5,1)-code is a packing in the 5-cube, Paul is always able to eliminate all but at most one codeword from Carole’s answers. (If he eliminates all codewords, Carole has defaulted (violated the bound on errors in the game).) In this game, we say that Carole has “lied” once in the 2nd question about 01010 being the correct codeword. 11111 11111 10100 10100 01010 01010 01010 00001 00001 00001 00001 00001 01111 00100 00010 00011 00100 00010 00010 00010

(3,1)-packing code and (3,1)-covering code “perfect code” 8 Covering Codes Covering is the companion problem to packing Packing: (n,e)-code Covering: (n,R)-code 110 011 101 111 000 010 001 100 packing radius length covering radius (3,1)-packing code and (3,1)-covering code “perfect code” By convention, coding theorists use e for the packing radius and R for the covering radius. For the rest of this talk, we assume e=R. Hamming (n,1)-codes are perfect codes. They exist when the size of the Hamming ball divides the size of the n-cube. There is no perfect (5,1)-code. This is because the 5-cube volume, 32, is not divisible by size of the radius 1 Hamming ball, 6. 11111 11111 00100 10100 01111 00010 01010 10111 00001 00001 11000 (5,1)-packing code (5,1)-covering code

Optimal Length 5 Packing & Covering Codes 9 Optimal Length 5 Packing & Covering Codes (5,1)-packing code 11111 11110 11101 11011 10111 01111 11100 11010 11001 10110 10101 10011 01110 01101 01011 00111 11000 10100 10010 10001 01100 01010 01001 00110 00101 00011 10000 01000 00100 00010 00001 (5,1)-covering code 11111 00000 11110 11101 11011 10111 01111 11100 11010 11001 10110 10101 10011 01110 01101 01011 00111 These two codes can be proved optimal by exhaustive combinatorial analysis, though there more powerful techniques. This packing code cannot be converted into an optimal covering code by the addition of Hamming balls. (It would take 4 more Hamming balls, as no three uncovered words are in the same ball.) The optimal packing and covering code sizes, times (n+1), bracket 2^n. Sphere bound: 11000 10100 10010 10001 01100 01010 01001 00110 00101 00011 10000 01000 00100 00010 00001 00000

A (5,1)-Covering Code as a Football Pool 10 A (5,1)-Covering Code as a Football Pool Round 1 Round 2 Round 3 Round 4 Round 5 Bet 1 W W W W W 11111 Bet 2 L W W W W 01111 Bet 3 W L W W W 10111 Bet 4 W W L L L 11000 Bet 5 L L W L L 00100 Bet 6 L L L W L 00010 Football means soccer! 5 rounds of football are played, each with outcome (W)in or (L)ose, with respect to the home team. A bet is a set of predictions on each of the 5 rounds. A payoff is made for a bet with at most 1 incorrect prediction. The minimum number of bets needed to guarantee a payoff is 7, because the problem is equivalent to finding an optimal (5,1)-covering code. A Finnish magazine for soccer pools published one such code a couple of years before it was discovered by Swiss scientist Marcel J.E. Golay. We can also view as a 2-player game, moving codeword/bets to the right as their predictions conflict with game results in rounds 1-5. Bet 7 L L L L W 00001 Payoff: a bet with ≤ 1 bad prediction Question. Min # bets to guarantee a payoff? Ans.=7

Codes with Feedback (Adaptive Codes) 11 Codes with Feedback (Adaptive Codes) Feedback Noiseless, delay-less report of actual received bits Improves the number of decodable messages E.g., from 20 to 28 messages for an (8,1)-code 1, 0, 1, 1, 0 1, 1, 1, 1, 0 sender receiver With feedback, the sender can adjust the transmission according to a pre-agreed strategy. Codes employing feedback are called “adaptive codes” Feedback does not help for n<=7 for packing codes. Noise 1, 1, 1, 1, 0 Elwyn Berlekamp Noiseless Feedback

A (5,1)-Adaptive Packing Code as a 2-Player Liar Game 12 A (5,1)-Adaptive Packing Code as a 2-Player Liar Game Paul Carole A Is the message C or D? Y B Is the message A or C? N C Is the message B? N D Is the message D? N 1 >1 Is the message C? Y # lies Message A B C D The question “Is the message C or D” corresponds to the first bit of the codewords for C and D being 1, and for those of A and B being 0. Carole’s first response of “Y” corresponds to transmitting the first bit as 1. Now Paul knows the original encoding of A and B will not be completely successful, so he may adapt to a new encoding. His second question “Is the message A or C?” assigns a 1 to the second bit of A and C, and a 0 to the second bit of B and D. For A and B, this means in the second bit of the adapted encoding. Original encoding 00101 01110 11000 01010 10011 1**** 1**** 10*** 1000* Adapted encoding 11*** 10*** 101** 100** 1000* 10000 10001 Y $ 1, N $ 0

A (5,1)-Adaptive Covering Code as a Football Pool 13 A (5,1)-Adaptive Covering Code as a Football Pool Round 1 Round 2 Round 3 Round 4 Round 5 Bet 1 W W L L W Bet 1 Bet 2 Bet 2 W L W W L Bet 3 Bet 3 W Bet 4 Bet 5 Bet 4 L Bet 6 Bet 5 L 1 >1 Bet 6 L # bad predictions (# lies) Now we consider companion problem in which Paul wants to guarantee that at least one message survives. This is in the same way that the covering code is the companion problem to the packing code. We start with the football pool problem, but now allow the predictions of the bet to be filled out one-by-one, after the previous round has been played. Carole adversarially supplies the outcome of the round, and tries to disqualify all of Paul’s bets. In the second round, a bet with no bad predictions is worth more than a bet with a bad prediction. So the 1-2 split in the bets with 0 bad predictions is balanced by the 3-0 split in the bets with 1 bad prediction. By allowing adaptivity in choosing predictions, the number of betting cards needed reduces from 7 to 6! Carole W L L W L Payoff: a bet with ≤ 1 bad prediction Question. Min # bets to guarantee a payoff? Ans.=6

Optimal (5,1)-Codes Code type Optimal size (5,1)-code 4 14 Optimal (5,1)-Codes Code type Optimal size (5,1)-code 4 (5,1)-adaptive code Sphere bound 5 1/3 (= 25/(5+1) ) (5,1)-adaptive covering code 6 (5,1)-covering code 7

Adaptive Codes: Results and Questions 15 Adaptive Codes: Results and Questions Sizes of optimal adaptive packing codes Binary, fixed e ≥ sphere bound - ce (Spencer `92) Binary, e=1,2,3 =sphere bound - O(1), exact solutions (Pelc; Guzicki; Deppe) Q-ary, e=1 =sphere bound - c(q,e), exact solution (Aigner `96) Q-ary, e linear unknown if rate meets Hamming bound for all e. (Ahlswede, C. Deppe, and V. Lebedev) Sizes of optimal adaptive covering codes Binary, fixed e · sphere bound + Ce Binary, e=1,2 =sphere bound + O(1), exact solution (Ellis, Ponomarenko, Yan `05) Near-perfect adaptive codes Q-ary, symmetric or “balanced”, e=1 exact solution (Ellis `04+) General channel, fixed e asymptotic first term (Ellis, Nyman `09)

Outline of Talk Coding theory overview Diffusion processes on Z 16 Outline of Talk Coding theory overview Packing (error-correcting) & covering codes Coding as a 2-player game Liar game and pathological liar game Diffusion processes on Z Simple random walk (linear machine) Liar machine Pathological liar game, alternating question strategy Improved pathological liar game bound Reduction to liar machine Discrepancy analysis of liar machine versus linear machine Concluding remarks

Linear Machine on Z 11 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 17 1 2 3 4 5 6 7 8 17

Linear Machine on Z 5.5 5.5 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 1 2 3 4 5 6 7 8 18

Time-evolution: 11 £ binomial distribution of {-1,+1} coin flips Linear Machine on Z 2.75 5.5 2.75 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 Time-evolution: 11 £ binomial distribution of {-1,+1} coin flips 19

Liar Machine on Z Liar machine time-step Number chips left-to-right 1,2,3,… Move odd chips right, even chips left (Reassign numbers every time-step) t=0 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 11 chips Approximates linear machine Preserves indivisibility of chips 20

Liar Machine on Z Liar machine time-step Number chips left-to-right 1,2,3,… Move odd chips right, even chips left (Reassign numbers every time-step) t=1 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 21

Liar Machine on Z Liar machine time-step Number chips left-to-right 1,2,3,… Move odd chips right, even chips left (Reassign numbers every time-step) t=2 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 22

Liar Machine on Z Liar machine time-step Number chips left-to-right 1,2,3,… Move odd chips right, even chips left (Reassign numbers every time-step) t=3 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 23

Liar Machine on Z Liar machine time-step Number chips left-to-right 1,2,3,… Move odd chips right, even chips left (Reassign numbers every time-step) t=4 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 24

Liar Machine on Z Liar machine time-step Number chips left-to-right 1,2,3,… Move odd chips right, even chips left (Reassign numbers every time-step) t=5 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 25

Liar Machine on Z Liar machine time-step Number chips left-to-right 1,2,3,… Move odd chips right, even chips left (Reassign numbers every time-step) t=6 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 26

Liar Machine on Z Liar machine time-step Number chips left-to-right 1,2,3,… Move odd chips right, even chips left (Reassign numbers every time-step) t=7 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 Height of linear machine at t=7 l1-distance: 5.80 l∞-distance: 0.98 27

Discrepancy for Two Discretizations Liar machine: round-offs spatially balanced Rotor-router model/Propp machine: round-offs temporally balanced The liar machine has poorer discrepancy… but provides bounds to the pathological liar game. Interval length=L. Liar machine: O(sqrt(t)) if L >sqrt(t)/2. O(L ln(t/L^2)) if L <= sqrt(t)/2. 28

Proof of Liar Machine Pointwise Discrepancy f_{t+1}(z): For f_{t+1} at position z, Delta brings along an extra half-chip from any right-arrow at position z-1, time t; Delta^{-1} brings along an extra half-chip from any left-arrow at position z+1, time t. Recursion gives f_t in terms of f_0 and L^s and Delta’s applied to arrows, for all intermediate time steps 0,1,…,t-1 L^tf_0 is just the linear machine’s g_t, and chi is expressed as the alternating sum of discrete Kronecker delta functions 2nd to last line: (Delta-Delta^{-1})…) is bimodal on its support, and inner sum over i is a spatially alternating sum of shifts of the same bimodal function, so at most 4 times maximum, which is 3/s Summing over s gives the last line. 29

The Liar Game as a Diffusion Process A priori: M=#chips, n=#rounds, e=max #lies Initial configuration: f0 = M ¢ 0 Each round, obtain ft+1 from ft by: (1) Paul 2-colors the chips (2) Carole moves one color class left, the other right Final configuration: fn Winning conditions Original variant (Berlekamp, Rényi, Ulam) Pathological variant (Ellis, Yan) Interval length=L. Liar machine: O(sqrt(t)) if L >sqrt(t)/2. O(L ln(t/L^2)) if L <= sqrt(t)/2. 30

Pathological Liar Game Bounds Fix n, e. Define M*(n,e) = minimum M such that Paul can win the pathological liar game with parameters M,n,e. Sphere Bound (E,P,Y `05) For fixed e, M*(n,e) · sphere bound + Ce (Delsarte,Piret `86) For e/n 2 (0,1/2), M*(n,e) · sphere bound ¢ n ln 2 . (C,E `09+) For e/n 2 (0,1/2), using the liar machine, M*(n,e) = sphere bound ¢ . The sphere bound is obtained by showing that Paul can do no better than if he were allowed to subdivide chips, and looking at the cumulative distribution of M times the binomial distribution of {-1,+1} coin flips at g_n(-n)+…+g_n(-n+2e). 31

Liar Machine vs. (6,1)-Pathological Liar Game 32 Liar Machine vs. (6,1)-Pathological Liar Game 9 chips 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 t=0 We reduce the pathological liar game to the liar machine as follows: Paul mimics the liar machine by numbering the chips left-to-right 1,2,3,… Coloring the odd chips green and the even chips purple. Paul thereby restricts his playing strategy. (This strategy is sometimes called the “alternating question strategy” in the literature.) We prove that Carole can do no better than moving the green chips right and purple chips left. 9 chips 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 disqualified

Liar Machine vs. (6,1)-Pathological Liar Game 33 Liar Machine vs. (6,1)-Pathological Liar Game 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 t=1 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 disqualified

Liar Machine vs. (6,1)-Pathological Liar Game 34 Liar Machine vs. (6,1)-Pathological Liar Game 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 t=2 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 disqualified

Liar Machine vs. (6,1)-Pathological Liar Game 35 Liar Machine vs. (6,1)-Pathological Liar Game 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 t=3 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 disqualified

Liar Machine vs. (6,1)-Pathological Liar Game 36 Liar Machine vs. (6,1)-Pathological Liar Game 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 t=4 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 disqualified

Liar Machine vs. (6,1)-Pathological Liar Game 37 Liar Machine vs. (6,1)-Pathological Liar Game 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 t=5 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 disqualified

Liar Machine vs. (6,1)-Pathological Liar Game 38 Liar Machine vs. (6,1)-Pathological Liar Game 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 t=6 No chips survive: Paul loses 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 disqualified

Comparison of Processes 39 Comparison of Processes Process Optimal #chips Linear machine 9 1/7 (6,1)-Pathological liar game 10 (6,1)-Liar machine 12 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 disqualified (6,1)-Liar machine started with 12 chips after 6 rounds

Loss from Liar Machine Reduction 40 Loss from Liar Machine Reduction 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 t=3 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 disqualified Why is there typically loss in the reduction to the Liar Machine? Re-start the 6-round 1-lie game with 10 chips. After 3 rounds we reach this configuration. Paul’s question determined by the liar machine is shown in the middle. But his optimal question is shown on the bottom. The chip at position -3 is “worth” more than a chip at position -1. So we color more chips purple at position -1 to compensate. Paul’s optimal 2-coloring: 9 -9 -8 -7 -6 -5 -4 -3 -2 -1 1 2 3 4 5 6 7 8 disqualified

Reduction to Liar Machine \documentclass{slides}\pagestyle{empty} \usepackage{amsmath,multirow,amssymb} \usepackage[usenames]{color} \setlength{\textwidth}{8.75in} \newcommand{\E}{\mathrm{E}} \newcommand{\R}{\mathcal{R}} \renewcommand{\P}{\mathrm{P}} \newcommand{\ho}{H} \newcommand{\h}{\widetilde{H}} \newcommand{\V}{\mathrm{Var}} \newcommand{\Gnp}{\mathcal{G}(n,p)} \newcommand{\SigM}{\hat{\Sigma}_{\mathrm{R}}(H)} \newcommand{\HCoreSG}[3]{\mathcal{H}_{#2,#3}(#1)} \newcommand{\SRep}[2]{\mathcal{S}_{#2}(#1)} \begin{document} {\bf Proof sketch.} {\bf (1)} $M$-chip liar game position vectors: $$\mathcal{U} = \{(u(1),\ldots,u(M)\}\in \mathbb{N}^M:u(1)\leq \cdots\leq u(M)\}$$ with weak majorization partial order $u\leq v$ provided for all $1\leq k\leq M$, \vspace{-.15in} $$\sum_{j=1}^k u(j)\leq \sum_{j=1}^k v(j).$$ {\bf (2)} Paul's questions partition odd- versus even-indexed chips.\\ Carole moves odd chips ($\textsc{odd}(u)$) or even chips ($\textsc{even}(u)$) \vspace{-.25in} {\bf (3)} $u\leq v\Rightarrow \textsc{odd}(u)\leq \textsc{odd}(v)$.\\ By induction, $\textsc{even}^t(u)\leq \{\textsc{odd},\textsc{even}\}^t(u)\leq \textsc{odd}^t(u)$. {\bf (4)} Set $u=(u(1),\ldots,u(M))=(0,\ldots,0)$.\\ If $\textsc{odd}^n(u)\leq e$, Paul wins the $(n,e)$-pathological liar game. {\bf (5)} Rescale to liar machine: $\textsc{odd}^t(u)(i)=j \Leftrightarrow f_t(-t+2j){\small+\!\!=}1.$ \end{document} 41

Saving One Chip in the Liar Machine 42 Saving One Chip in the Liar Machine

Summary: Pathological Liar Game Theorem

44 Further Exploration Tighten the discrepancy analysis for the special case of initial chip configuration f0=M 0. Generalize from binary questions to q-ary questions, q ¸ 2. Improve analysis of the original liar game from Spencer and Winkler `92; solve the optimal rate of q-ary adaptive block codes for all fractional error rates. Prove general pointwise and interval discrepancy theorems for various discretizations of random walks.

45 Reading List This paper: Linearly bounded liars, adaptive covering codes, and deterministic random walks, preprint (see homepage). The liar machine Joel Spencer and Peter Winkler. Three thresholds for a liar. Combin. Probab. Comput.1(1):81-93, 1992. The pathological liar game Robert Ellis, Vadim Ponomarenko, and Catherine Yan. The Renyi-Ulam pathological liar game with a fixed number of lies. J. Combin. Theory Ser. A 112(2):328-336, 2005. Discrepancy of deterministic random walks Joshua Cooper and Joel Spencer, Simulating a Random Walk with Constant Error, Combin. Probab. Comput. 15 (2006), no. 06, 815-822. Joshua Cooper, Benjamin Doerr, Joel Spencer, and Gabor Tardos. Deterministic random walks on the integers. European J. Combin., 28(8):2072-2090, 2007.