1 Coinduction Principle for Games Michel de Rougemont Université Paris II & LRI.

Slides:



Advertisements
Similar presentations
Efficient classification for metric data Lee-Ad GottliebWeizmann Institute Aryeh KontorovichBen Gurion U. Robert KrauthgamerWeizmann Institute TexPoint.
Advertisements

1 Property testing and learning on strings and trees Michel de Rougemont University Paris II & LRI Joint work with E. Fischer, Technion, F. Magniez, LRI.
Clustering II.
Property testing of Tree Regular Languages Frédéric Magniez, LRI, CNRS Michel de Rougemont, LRI, University Paris II.
Euclidean Algorithm Applied Symbolic Computation CS 567 Jeremy Johnson.
Cook’s Theorem The Foundation of NP-Completeness.
Large Vocabulary Unconstrained Handwriting Recognition J Subrahmonia Pen Technologies IBM T J Watson Research Center.
MIT and James Orlin © Game Theory 2-person 0-sum (or constant sum) game theory 2-person game theory (e.g., prisoner’s dilemma)
1 Markov Decision Processes: Approximate Equivalence Michel de Rougemont Université Paris II & LRI
1 The Monte Carlo method. 2 (0,0) (1,1) (-1,-1) (-1,1) (1,-1) 1 Z= 1 If  X 2 +Y 2  1 0 o/w (X,Y) is a point chosen uniformly at random in a 2  2 square.
Randomized Algorithms Kyomin Jung KAIST Applied Algorithm Lab Jan 12, WSAC
Markov Chains 1.
CS774. Markov Random Field : Theory and Application Lecture 04 Kyomin Jung KAIST Sep
Discounting the Future in Systems Theory Chess Review May 11, 2005 Berkeley, CA Luca de Alfaro, UC Santa Cruz Tom Henzinger, UC Berkeley Rupak Majumdar,
Clustering II.
Foundations of Data-Flow Analysis. Basic Questions Under what circumstances is the iterative algorithm used in the data-flow analysis correct? How precise.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
A New Evolutionary Algorithm for Multi-objective Optimization Problems Multi-objective Optimization Problems (MOP) –Definition –NP hard By Zhi Wei.
Metrics for real time probabilistic processes Radha Jagadeesan, DePaul University Vineet Gupta, Google Inc Prakash Panangaden, McGill University Josee.
Programming Language Semantics Denotational Semantics Chapter 5 Part II.
L16: Micro-array analysis Dimension reduction Unsupervised clustering.
NORM BASED APPROACHES FOR AUTOMATIC TUNING OF MODEL BASED PREDICTIVE CONTROL Pastora Vega, Mario Francisco, Eladio Sanz University of Salamanca – Spain.
2.3 General Conditional Expectations 報告人:李振綱. Review Def (P.51) Let be a nonempty set. Let T be a fixed positive number, and assume that for each.
If time is continuous we cannot write down the simultaneous distribution of X(t) for all t. Rather, we pick n, t 1,...,t n and write down probabilities.
1 Hybrid Agent-Based Modeling: Architectures,Analyses and Applications (Stage One) Li, Hailin.
Data Flow Analysis Compiler Design Nov. 8, 2005.
Tutorial 10 Iterative Methods and Matrix Norms. 2 In an iterative process, the k+1 step is defined via: Iterative processes Eigenvector decomposition.
Maximum Entropy Model LING 572 Fei Xia 02/07-02/09/06.
Reinforcement Learning Game playing: So far, we have told the agent the value of a given board position. How can agent learn which positions are important?
1 Biased card shuffling and the asymmetric exclusion process Elchanan Mossel, Microsoft Research Joint work with Itai Benjamini, Microsoft Research Noam.
Problems, cont. 3. where k=0?. When are there stationary distributions? Theorem: An irreducible chain has a stationary distribution  iff the states are.
An efficient distributed protocol for collective decision- making in combinatorial domains CMSS Feb , 2012 Minyi Li Intelligent Agent Technology.
1 MCMC Style Sampling / Counting for SAT Can we extend SAT/CSP techniques to solve harder counting/sampling problems? Such an extension would lead us to.
1 Approximate Satisfiability and Equivalence Michel de Rougemont University Paris II & LRI Joint work with E. Fischer, Technion, F. Magniez, LRI, LICS.
Markov Decision Processes1 Definitions; Stationary policies; Value improvement algorithm, Policy improvement algorithm, and linear programming for discounted.
The Selection Problem. 2 Median and Order Statistics In this section, we will study algorithms for finding the i th smallest element in a set of n elements.
Approximate schemas Michel de Rougemont, LRI, University Paris II.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Markov Models and Simulations Yu Meng Department of Computer Science and Engineering Southern Methodist University.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
1 Parrondo's Paradox. 2 Two losing games can be combined to make a winning game. Game A: repeatedly flip a biased coin (coin a) that comes up head with.
COMPSCI 102 Introduction to Discrete Mathematics.
Projective Geometry Hu Zhan Yi. Entities At Infinity The ordinary space in which we lie is Euclidean space. The parallel lines usually do not intersect.
5. Maximum Likelihood –II Prof. Yuille. Stat 231. Fall 2004.
Information Bottleneck versus Maximum Likelihood Felix Polyakov.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
x Examples of Fixed Points Infinite Fixed Points.
Solving Scalar Linear Systems A Little Theory For Jacobi Iteration
Example 31: Solve the following game by equal gains method: Y I II I II I X II II
Gaoal of Chapter 2 To develop direct or iterative methods to solve linear systems Useful Words upper/lower triangular; back/forward substitution; coefficient;
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
Dynamic Programming & Hidden Markov Models. Alan Yuille Dept. Statistics UCLA.
Program Analysis Mooly Sagiv Tel Aviv University Sunday Scrieber 8 Monday Schrieber.
Theory of Computational Complexity Probability and Computing Lee Minseon Iwama and Ito lab M1 1.
Xiaowei Ying, Kai Pan, Xintao Wu, Ling Guo Univ. of North Carolina at Charlotte SNA-KDD June 28, 2009, Paris, France Comparisons of Randomization and K-degree.
PROBABILITY AND COMPUTING RANDOMIZED ALGORITHMS AND PROBABILISTIC ANALYSIS CHAPTER 1 IWAMA and ITO Lab. M1 Sakaidani Hikaru 1.
SIMILARITY SEARCH The Metric Space Approach
Theorem of Banach stainhaus and of Closed Graph
Markov Chains and Mixing Times
Path Coupling And Approximate Counting
Probabilistic Models for Linear Regression
Totally Disjoint Multipath Routing in Multihop Wireless Networks Sonia Waharte and Raoef Boutaba Presented by: Anthony Calce.
On Statistical Model Checking of Stochastic Systems
SOLUTION OF NONLINEAR EQUATIONS
Metrics for real time probabilistic processes
The Selection Problem.
Totally Asynchronous Iterative Algorithms
Presentation transcript:

1 Coinduction Principle for Games Michel de Rougemont Université Paris II & LRI

2 Coinduction (inspired by D. Kozen) Naive notion Induction: F monotone, F has a unique least-fixed point. Coinduction: F is antimonotone, unique smallest greatest fixed-point. Example: Even numbers over N General Principle Function F defined by an equation, admits a fixed-point F*. We want to show that F* satisfies a property: it suffices to show that the property is preserved by the equation. Foundation: Let B a Banach Space, R is a linear Operator, of spectral radius <1. It admits a fixed point. Applications to Computer Science: 1.Evolutionary Games 2. Combinatorics 3.Stochastic processes

3 1. Evolutionary Dynamics Example: Rock-Scissor-Paper: Mixed strategy= density of agents playing pure strategies Replicator Strategy:

4 2. Enumerative Combinatorics Coinductive Counting (J. Rutten), Stream Calculus Male Bees=Drones D (Q,D)  Q Female Bees=Queens Q Q  D How many Q ancestors at level k? Q D Q Q D D Q Q D Q D Q k q0q0 q1q1

5 3. Probabilistic Processes Equivalence of Markov Chains ? Metric Analogue of Bisimulation (Desharnais and al.) D-bar measure in Statistics Approximate Equivalence Property of Markov Chains: …. p 1-p

6 Coinduction to compare 2 processes Property of Markov Chains: Examples from D. Kozen, Lics 2006 Generation of a biased coin (q) given a biased coin (p). Consider 3 different processes.

7 Biased coin simulation Algorithm : qflip(q): If q >p( if pflip=head) return head else return qflip(q-p/1-p) ) else ( if pflip=head) return qflip(q/p) else return tail ) p q 0 1 q’ 0 1 Given: pflip a biased coin (head, tail) with probability (p,1-p). Task(q) : Generate a biased coin with probability (q,1-q). q p 0 1 q’ 0

8 Strategy : qflip(q) Convergence: output with probability Min(p,1-p) at each step. Halts with probability 1. H(q)=Probability qflip(q)=head p q 0 1 q’ 0 1

9 Time estimation of qflip(q) Estimated Time: p q 0 1 Algorithm : qflip(q): If q >p( if pflip=head) return head else return qflip(q-p/1-p) ) else ( if pflip=head) return qflip(q/p) else return tail ) q’ 0 1

10 Strategy 1: qflip(q) Estimated Time: p q 01 Algorithm 1 : qflip(q): If ( q>1-p) ( if pflip=head) return qflip((q-1+p)/p) else return head Else if q >p( if pflip=head) return head else return qflip(q-p/1-p) ) else ( if pflip=head) return qflip(q/p) else return tail ) 1-p

11 Comparison between Strategies 0 and 1

12 Bounded Linear Operators B, Banach Space, R is a linear Operator, of spectral radius <1. Affine Operator τ(e) =a+Re Φ closed non empty region preserved by τ. Conclusion: there exists a fixed point e* (τ(e*) =a+Re*) s.t e* is in Φ. Example: E(q) is bounded. B Φ E*

13 Coinduction principle Co induction principle:

14 Application Co induction principle

15 Case 1 Co induction setting

16 Strategy 2: qflip(q) Estimated Time: p q 01 Algorithm 2 : qflip(q): If ( q>1-p) ( if pflip=head) return qflip((q-1+p)/p) else return head Else If ( q>0.5) ( if pflip=head) return tail else qflip(q/1-p) Else if q >p( if pflip=head) return head else return qflip(q-p/1-p) ) else ( if pflip=head) return qflip(q/p) else return tail ) 1-p 1/2

17 New Application Co induction principle on pairs (E,E’)

18 Probabilistic Processes Equivalence of Markov Chains ? Are M1,M2 ε-close ? Metric Analogue of Bisimulation (Desharnais and al.) Distance d between distributions obtained by iterations, also the maximum fixed-point of a Functional F. Property of Markov Chains: …. p-ε 1-p+ ε …. p 1-p M1M1 M2M2

19 Conclusion General Principle: Given τ linear bounded operator of spectral radius <1, Φ closed non empty region preserved by τ, we conclude that there exists a fixed point e* in Φ. Applications: 1.Stochastic processes. Compare Expected time between two fractal processes. 2.Evolutionary Games. Compare convergence time. 3.Analysis of Streams.