Graph Guessing Games & non-Shannon Information Inequalities First workshop on Entropy and Information Inequalities 17 April 2013 Søren Riis Queen Mary.

Slides:



Advertisements
Similar presentations
6.896: Topics in Algorithmic Game Theory Lecture 21 Yang Cai.
Advertisements

Bayesian Networks, Winter Yoav Haimovitch & Ariel Raviv 1.
Introduction to Algorithms
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Complexity ©D Moshkovitz 1 Approximation Algorithms Is Close Enough Good Enough?
Approximation Algorithms Chapter 5: k-center. Overview n Main issue: Parametric pruning –Technique for approximation algorithms n 2-approx. algorithm.
Parallel Scheduling of Complex DAGs under Uncertainty Grzegorz Malewicz.
GOLOMB RULERS AND GRACEFUL GRAPHS
II. Linear Block Codes. © Tallal Elshabrawy 2 Last Lecture H Matrix and Calculation of d min Error Detection Capability Error Correction Capability Error.
Isaac Newton Institute 9th of May 2012 Søren Riis Queen Mary University of London Valiant’s Shift problem: A reduction to a problem about graph guessing.
April 2, 2015Applied Discrete Mathematics Week 8: Advanced Counting 1 Random Variables In some experiments, we would like to assign a numerical value to.
Optimization of Pearl’s Method of Conditioning and Greedy-Like Approximation Algorithm for the Vertex Feedback Set Problem Authors: Ann Becker and Dan.
The number of edge-disjoint transitive triples in a tournament.
Fast FAST By Noga Alon, Daniel Lokshtanov And Saket Saurabh Presentation by Gil Einziger.
Entropy Rates of a Stochastic Process
Computational Game Theory
Hardness Results for Problems P: Class of “easy to solve” problems Absolute hardness results Relative hardness results –Reduction technique.
Network Coding Theory: Consolidation and Extensions Raymond Yeung Joint work with Bob Li, Ning Cai and Zhen Zhan.
CSE115/ENGR160 Discrete Mathematics 02/07/12
Computational Geometry Seminar Lecture 1
Perfect Graphs Lecture 23: Apr 17. Hard Optimization Problems Independent set Clique Colouring Clique cover Hard to approximate within a factor of coding.
NP-Complete Problems Reading Material: Chapter 10 Sections 1, 2, 3, and 4 only.
Approximation Algorithms
Theta Function Lecture 24: Apr 18. Error Detection Code Given a noisy channel, and a finite alphabet V, and certain pairs that can be confounded, the.
Distributed Combinatorial Optimization
Graph Theory Ch.5. Coloring of Graphs 1 Chapter 5 Coloring of Graphs.
Maximal Independent Set Distributed Algorithms for Multi-Agent Networks Instructor: K. Sinan YILDIRIM.
Copyright © Cengage Learning. All rights reserved. CHAPTER 11 ANALYSIS OF ALGORITHM EFFICIENCY ANALYSIS OF ALGORITHM EFFICIENCY.
Some Surprises in the Theory of Generalized Belief Propagation Jonathan Yedidia Mitsubishi Electric Research Labs (MERL) Collaborators: Bill Freeman (MIT)
The Hat Game 11/19/04 James Fiedler. References Hendrik W. Lenstra, Jr. and Gadiel Seroussi, On Hats and Other Covers, preprint, 2002,
Zvi Kohavi and Niraj K. Jha 1 Capabilities, Minimization, and Transformation of Sequential Machines.
CS774. Markov Random Field : Theory and Application Lecture 08 Kyomin Jung KAIST Sep
Computing and Communicating Functions over Sensor Networks A.Giridhar and P. R. Kumar Presented by Srikanth Hariharan.
Introduction to Proofs
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Independence and Bernoulli.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 1 – Slide 1 of 34 Chapter 11 Section 1 Random Variables.
Section 1.8. Section Summary Proof by Cases Existence Proofs Constructive Nonconstructive Disproof by Counterexample Nonexistence Proofs Uniqueness Proofs.
1 2. Independence and Bernoulli Trials Independence: Events A and B are independent if It is easy to show that A, B independent implies are all independent.
1 Combinatorial Algorithms Parametric Pruning. 2 Metric k-center Given a complete undirected graph G = (V, E) with nonnegative edge costs satisfying the.
A logicians approach to Network Coding Søren Riis The Isaac Newton Institute, Cambridge 7 February, 2012 Søren Riis The Isaac Newton Institute, Cambridge.
Independence and Bernoulli Trials. Sharif University of Technology 2 Independence  A, B independent implies: are also independent. Proof for independence.
Maximum density of copies of a graph in the n-cube John Goldwasser Ryan Hansen West Virginia University.
CSE 589 Part VI. Reading Skiena, Sections 5.5 and 6.8 CLR, chapter 37.
Chapter 10 Graph Theory Eulerian Cycle and the property of graph theory 10.3 The important property of graph theory and its representation 10.4.
ICS 253: Discrete Structures I Induction and Recursion King Fahd University of Petroleum & Minerals Information & Computer Science Department.
Linear Program Set Cover. Given a universe U of n elements, a collection of subsets of U, S = {S 1,…, S k }, and a cost function c: S → Q +. Find a minimum.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Conditional Probability Mass Function. Introduction P[A|B] is the probability of an event A, giving that we know that some other event B has occurred.
Implicit Hitting Set Problems Richard M. Karp Erick Moreno Centeno DIMACS 20 th Anniversary.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
1 What happens to the location estimator if we minimize with a power other that 2? Robert J. Blodgett Statistic Seminar - March 13, 2008.
Secret Sharing Non-Shannon Information Inequalities Presented in: Theory of Cryptography Conference (TCC) 2009 Published in: IEEE Transactions on Information.
1 On the Channel Capacity of Wireless Fading Channels C. D. Charalambous and S. Z. Denic School of Information Technology and Engineering, University of.
Approximation Algorithms Duality My T. UF.
Network Formation Games. NFGs model distinct ways in which selfish agents might create and evaluate networks We’ll see two models: Global Connection Game.
Approximation Algorithms based on linear programming.
Theory of Computational Complexity Probability and Computing Ryosuke Sasanuma Iwama and Ito lab M1.
Network Topology Single-level Diversity Coding System (DCS) An information source is encoded by a number of encoders. There are a number of decoders, each.
ICS 353: Design and Analysis of Algorithms NP-Complete Problems King Fahd University of Petroleum & Minerals Information & Computer Science Department.
CSE15 Discrete Mathematics 02/08/17
The minimum cost flow problem
Sum of Squares, Planted Clique, and Pseudo-Calibration
Chapter 5. Optimal Matchings
Optimal Query Processing Meets Information Theory
1.3 Modeling with exponentially many constr.
3.3 Applications of Maximum Flow and Minimum Cut
ICS 353: Design and Analysis of Algorithms
3.5 Minimum Cuts in Undirected Graphs
1.3 Modeling with exponentially many constr.
Locality In Distributed Graph Algorithms
Presentation transcript:

Graph Guessing Games & non-Shannon Information Inequalities First workshop on Entropy and Information Inequalities 17 April 2013 Søren Riis Queen Mary University of London

Overview 1.Digraph* Guessing Games 2.Graph Entropy** (causal networks) 3.Network Coding ↔ Guessing Numbers ↔ Graph Entropy 4.Examples and some basic results 5.Application of Non-Shannon Information Inequalities 6.Graph where G and G d have different Shannon bounds 7.The superman conjecture and other false propositions 8.Final Remarks and Open Questions * From now on by “Graph” we mean “Digraph” ** In a network Coding sense and not in the sense of Körner

3 1. Graph Guessing Games We will consider a class of problems that looks “recreational” However, the problems go straight to the heart of core issues in Information Theory and Network Coding Guessing Game (Riis 1997) Requirements: n players are each given a die with s-sides Rules: Each player rolls their die. A player is NOT allowed to see the value of his/her own die, but is allowed to look at the dice values of the other n-1 players The players have to make their guess simultaneously (no communication allowed). Outcome: The players win if each player guesses correctly the value of their own die. Task: The players have (in advance) to choose a strategy that maximizes the probability that they win.

1. Graph Guessing Games (continued) Question: Assume n=100 and s=6. What is the probability that each player guesses correctly the value of their own die? Naïve, but wrong answer based on the following (correct) premises: Premise: Each player has no information about their own die value. Premise: Each player has probability 1/6 of guessing their own dice value. Premise: The 100 dice values are independent Conclusion (false): The probability that all players guess correctly is (1/6) 100

5 Graph Guessing Games (continued) Let us check the argument: 1)Premise: Each player has no relevant information about their own dice. 2) Premise: Each player has probability 1/6 of guessing their own die value. 3) Premise: The 100 dice values are independent 4)Conclusion: The probability that all players guess correctly is (1/6) 100 Mistake based on a serious fallacy.

Graph Guessing Games (continued) 4)The probability all players guess correctly is (1/6) 100 Mistake based on a serious fallacy. Let Z j =1 if player j guesses correctly his/her own dice value. Let Z j =0 otherwise. P(z j =1)=1/6 for j=1,2,3,….,100 P(players wins)= P(z 1 =1 ∧ z 2 =1 ∧ z 3 =1 ∧... ∧ z 100 =1 )= P(z 1 =1 ∧ z 2 =1 ∧.. ∧ z 99 =1|z 100 =1) P(z 100 =1) ≤ p(z 100 =1)=1/6 The players can arrange it so the conditional probability holds with certainty How?

7 Graph Guessing Games (continued) The probability all players are correct depend on their guessing strategy. If the players arrange it such that one player is correct if and only if all players are correct they win with probability 1/6 If each player “assume” that the sum of all dice values is 0 modulo 6, and guess accordingly, then all players are correct with probability 1/6

Graph Guessing Games (continued) The uncoordinated guessing strategy succeeds with probability (1/s) n The optimal guessing strategy succeeds with probability 1/s Thus by cooperating the players can achieve a probability that is s (n-1) times more likely of succeeding than a uncoordinated guessing strategy. The power n-1 in the factor s (n-1) plays an important role in our theory.

9 Graph Guessing Games (continued) Graph Guessing Game (Riis 2005) Requirements: n players are each given a die with s-sides Rules: Each player rolls their die. Each player is sitting in a node in a given graph. A player has only access to the dice values of their direct predecessors. The players have to make their guess simultaneously Outcome: The players win if each player guesses correctly the value of their own die. Task: The players have (in advance) to choose a strategy that maximizes the probability that they win.

10 Graph Guessing Games (continued) Definition: The guessing number of a graph G (or strategy) is α if the players can succeed with a probability that is s α times higher than the probability of success when using uncoordinated random guessing. The complete graph K n with bi-directed edges, corresponds to the guessing game we already considered. Proposition: The guessing number guess(K n ) of K n is n-1 because we can do s (n-1) times better than uncoordinated random guessing. Notice that the guessing number of K n does not depend on the size s of the “alphabet”

11 Graph Guessing Games (continued) Guess(directed cycle)= 1Guess(C 4 )=2

12

Graph Entropy (causal networks) Consider stochastic variables x 1,x 2,…,x n ε A that are assigned to the nodes in a (di)graph.* Assume that the value of a stochastic variable in a node is determined by its direct predecessor nodes. A stochastic variable in a node with NO predecessors is determined deterministically. ** *Its possible to be more general and consider digraphs with weighted nodes. ** Different from dependence in Bayesian nets as we allows loops. Example: Stochastic variables x,y,z,u and v are causally related as specified in the graph.

Graph Entropy (causal networks) Stochastic variables x,y,z,u and v are causally related as specified in the graph. We can consider the rate region of a causal network. We will take a different approach: We are interested in determining the maximal Entropy of H(x 1,x 2,….,x n ) subject to the causal requirements. We normalize the entropy by taking logarithms in base |A|.

Graph Entropy (causal networks) We will take a different approach: We are interested In determining the maximal Entropy of H(x 1,x 2,….,x n ) subject to the causal requirements. We normalize the entropy by taking logarithms in base |A|. Consider a clique (bi-directed edges) of n nodes. Consider it as a causal network of stochastic variables over a finite alphabet A. What is the maximal Entropy H(x 1,x 2,…,x n )? Answer: n-1 =Guess(K n )

16 Graph Entropy (causal networks) What is the maximal Entropy of a directed cycle? Answer: 1 = Guess(directed cycle) What is the maximal Entropy Of C 4 ? Answer: 2 = Guess(C 4 )

Graph Entropy (causal networks) Definition: The Entropy of a graph G=(V,E) (causal network) is the maximal Entropy of H(V) subject to the constrains: 1)H(j| U)=0 whenever U contains the in- neighborhood of j 2)H(vertex) ≦ 1

Graph Entropy (causal networks) Theorem: (Riis 2007, Gadouleau,Riis 2010) For each graph G and for each alphabet size s=2,3,4,… Guess(G,s)=Entropy(G,s) Let Guess(G)=lim s->∞ Guess(G,s) (Limit can be shown to exist) Then Guess(G)=Entropy(G)

Graph Entropy (causal networks) Example: Consider the causal network C 5 (or guessing game C 5 ). What is the maximal Entropy of H(1,2,3,4,5) when H(1),H(2),….,H(5) ≤ 1 and the stochastic variables are subject to the causal constrains? Answer: 2.5

Graph Entropy (causal networks) What is the maximal Entropy of H(1,2,3,4,5) when H(1),H(2),….,H(5) ≤ 1 and the stochastic variables are subject to the causal constrains? Answer: 2.5 Two different methods to obtain the lower bound: Method 1: Provide an optimal guessing strategy of the guessing game C 5 Method 2: Construct an Entropic vector H: P(1,2,3,4,5) -> R subject to the causal constraints

Graph Entropy (causal networks) Consider C 5 with bi-directed edges. Let A be an alphabet with s letters. Assume that s is a square number i.e. that s=t 2 for some t=2,3,4,…

Graph Entropy (causal networks) W.l.o.g. we can assume that each of the 5 players is assigned two dice (each of t sides). One labeled “ l ” for left, and one labeled “ r ” for right.

Graph Entropy (causal networks) Optimal Guessing strategy for C 5 : The players guess according to the following rule: Each “ left ” die has the same value as its corresponding “ right ” die. This strategy succeeds with probability (1/t) 5 = (1/s) 2.5

Graph Entropy (causal networks) This shows that pentagon has guessing number ≥ 2.5 if s=4,9,16,25,… is square number. This kind of argument can be used to show that Guess( C 5 ) ≥ 2.5 Proposition : Guess( C 5 ) = Entropy( C 5 ) = 2.5

Graph Entropy (causal networks) Proof of proposition: Let H be the Entropy function that satisfies the causal constraints given by C 5. In general H satisfies H(X,Y,Z) + H(Z ) ≤ H(X,Z) + H(Y,Z). If we let X={1}, Y={3} and Z={4,5} we get: H(1,2,3,4,5)+H(4,5) = H(1,3,4,5) + H(4,5) ≤ H(1,4,5) + H(3,4,5) = H(1,4)+H(3,5) ≤ H(1) +H(3) +H(4) +H(5) and thus H(1,2,3,4,5) ≤ H(1)+H(3)+H(4)+H(5)-H(4,5) Next notice that H(1,2,3,4,5)-H(2,5) =H(2,3,4,5)-H(2,5)= H(3,4 | 2,5) = H(4 | 2,5) ≤ H(4|5) = H(4,5)-H(5) which shows that H(1,2,3,4,5) ≤ H(2,5) + H(4,5)-H(5) ≤ H(4,5)+H(2) Adding up: 2H(1,2,3,4,5) ≤ H(1)+H(2)+H(3)+H(4)+H(5) ≤ 5 Thus: H has Entropy H(1,2,3,4,5) ≤ 2.5

26 Network Coding ↔ Guessing Numbers ↔ Graph Entropy Let N be a network with k input/output pairs. If we identify each input node with its corresponding output node, we get a graph G N. Theorem:[Riis 2005] N has a Network coding solution over alphabet A if and only if G N has guessing number  k (equivalent = k) over A. 26

27 Network Coding ↔ Guessing Numbers ↔ Graph Entropy (continued) An even a stronger result is valid: Theorem: [Riis 2005] Let N be a network with k input/output pairs. The number of distinct network coding solutions to N is identical to the number of distinct guessing strategies for G N that achieves guessing number k (the solutions are counted with respect to the same alphabet A). 27

28 Network Coding ↔ Guessing Numbers ↔ Graph Entropy (continued)

29 Network Coding ↔ Guessing Numbers ↔ Graph Entropy (continued)

30 Network Coding ↔ Guessing Numbers ↔ Graph Entropy (continued) The networks in b, c and d all appear by “ splitting ” the graph G in a. “ Splitting ” is the opposite operation as identifying input/output pairs. The networks in b, c and d have exactly the same number of network coding solutions (over a fixed alphabet A). This number is identical to the number of guessing strategies (over A) that achieve guessing number 3 for G. 30

Examples and some basic results

Application of Non-Shannon Information Inequalities Definition: The pseudo entropy of a graph is the minimal upper bound on the guessing number that can be derived from H(X,Y,Z) + H(Z) ≤ H(X,Z) + H(Y,Z) and H(Ø)=0 together with the constraints H(vertex) ≦ 1, and H(j| V)=0 whenever all incoming nodes to j belong to V. Example: C 5 has pseudo entropy = 2.5 (= Entropy(C 5 ))

Application of Non-Shannon Information Inequalities The entropy calculation in the previous section used Shannon ’ s information inequalities to derive an upper bound on the guessing number. Definition: For the Zhang and Yeung (=ZY) non-Shanon Information Inequalities we can define the ZY- entropy of a graph is the minimal upper bound on the guessing number that can be derived from ZY information inequalities. I(A; B) ≤ 2I(A; B | C) + I(A; C | B) + I(B; C | A) + I(A; B | D) + I(C; D) ( H(X,Y,Z) + H(Z) ≤ H(X,Z) + H(Y,Z) not needed) and H(Ø)=0 together with the constraints H(vertex) ≦ 1, and H(j| V)=0 whenever all incoming nodes to j belong to V. Example: pseudo entropy ≽ zy-entropy ≽ entropy

Application of Non-Shannon Information Inequalities Building block for many constructions

Application of Non-Shannon Information Inequalities Theorem: (Baber, Chistofides, Dang, Riis, Vaughan) Let R^ denote the graph R with edge (9,10) removed. For R^; The Shannon bound is 114/17= 6.705….. The Zhang-Yeung bound is 1212/181 = 6.696…. The Dougherty-Freilng-Zeger bound is 59767/8929 = 6.693… The Ingleton bound is 20/3 = 6.666…. Strong evidence this is the only such example on the app 12 million undirected graphs on ≤ nodes

Application of Non-Shannon Information Inequalities 5 : {12, 21, 23, 32, 34, 43, 45, 54, 51, 15} -3/2 : 0 = H(EmptySet) 1/2 : 1 >= H(1) 1/2 : 1 >= H(2) 1/2 : 1 >= H(3) 1/2 : 1 >= H(4) 1/2 : 1 >= H(5) 1/2 : 0 >= H(1,3,5) - H(1,2,3,5) 1/2 : 0 >= H(2,4,5) - H(1,2,4,5) 1/2 : 0 = H(1,2,5) - H(2,5) 1/2 : 0 = H(1,2,3,4,5) - H(2,3,4,5) 1/2 : 0 = H(1,2,3) - H(1,3) 1/2 : 0 = H(1,2,3,4,5) - H(1,3,4,5) 1/2 : 0 = H(2,3,4,5) - H(2,4,5) 1/2 : 0 = H(1,3,4,5) - H(1,3,5) 1/2 : 0 = H(1,2,4,5) - H(1,2,4) 1/2 : 0 >= H(EmptySet) + H(1,3) - H(1) - H(3) 1/2 : 0 >= H(EmptySet) + H(1,4) - H(1) - H(4) 1/2 : 0 >= H(1) + H(1,2,4) - H(1,2) - H(1,4) 1/2 : 0 >= H(EmptySet) + H(2,5) - H(2) - H(5) 1/2 : 0 >= H(1,2) + H(1,2,3,5) - H(1,2,3) - H(1,2,5) #Total : # 5/2 >= H(1,2,3,4,5) Glimpse of the data files. Derivation of Upper bound on Shannon bound for C 5

Application of Non-Shannon Information Inequalities Glimpse of the data files. Derivation of lower bound for the DFZ case Derivation of upper bound for DFZ case App pages (for all 214 DFZ inequalities)

Graph where G and G d have different Shannon bounds Let G be R where node 1 is turned into a Superman; I.e. there is an arc from every node to node 1. Let G d denote the dual graph. Node 1 is a Luthor node in G d Theorem: (Baber, Chistofides, Dang, Riis, Vaughan) For G; The Shannon bound 27/4=6.75 is optimal and equal the ingelton bound for G For G d ; The Shannon bound is 34/5=6.8 The ZY-bound is 61/9=6.777… The DFZ-bound is 359/53 = 6.773… The Ingelton Bound is 27/4=6.75

39 The superman conjecture and other false propositions Theorem: D. Christofides and K. Markström (2011) Each perfect graph* G on n nodes has Guessing/Entropy number given by n - minimal number of cliques that covers G. * A perfect graph is a graph in which the chromatic number of every induced subgraph equals the size of the largest clique of that subgraph. It can be shown that a graph is perfect if and only if the graph has no odd holes or odd anti-holes. Conjecture: D. Christofides and K. Markström (2011) The Guessing number of an undirected Graph is given by n - fractional clique cover number i.e. there is an optimal strategy that occurs by splitting each die into multiple die and dividing players into groups where each play the clique game for their shared die. The Graph R violates this conjecture.

The superman conjecture and other false propositions Superman conjecture/question: Does there exist an undirected graph whose asymptotic guessing number increases when a single directed edge is added? We were able to clone a part of R to achieve a graph R c on 13 nodes that provides a counter example to the superman conjecture.

Final remarks and open questions It is possible to use the Vanos graph to construct a graph with the Shannon bound 6 and ZY-bound 35/6=5.833… Sun (PhD thesis 2011) The graph R is simpler and has proved to be more powerful in our research.

Final remarks and Open Questions Is it possible to amplify the above result and construct graph(s) where there is a substantial gap between the pseudo Entropy, the zy-Entropy and the genuine Entropy? Improve the upper bounds and the lower bounds for the Entropy of R^ Can generalize the causal structures (Graphs) to lattices. Do the graphs G and G d have different graph entropies?

43 Final Remarks and Open Problems Spin off: Network Coding, Guessing Game and Graph entropy approach have lead to new areas of research Combinatorial Representations (2011) (With Peter Cameron and Max Gadouleau) A generalisation of Matroid theory Memoryless computation (2011) (With Max Gadouleau) Dynamic communication networks (2011) (With Max Gadouleau) Graph entropy and non-Shannon information inequalities (Yun Sun PhD Thesis 2011) Construction of new classes of communication networks (with Max Gadouleau) (2010) New max-flow min-cut theorem for multiuser communication (with Max Gadouleau) (2011) Thank you 33