Chernoff Bounds, and etc.

Slides:



Advertisements
Similar presentations
Part VI NP-Hardness. Lecture 23 Whats NP? Hard Problems.
Advertisements

“Devo verificare un’equivalenza polinomiale…Che fò? Fò dù conti” (Prof. G. Di Battista)
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
1 The Monte Carlo method. 2 (0,0) (1,1) (-1,-1) (-1,1) (1,-1) 1 Z= 1 If  X 2 +Y 2  1 0 o/w (X,Y) is a point chosen uniformly at random in a 2  2 square.
Probabilistic algorithms Section 10.2 Giorgi Japaridze Theory of Computability.
Randomized Algorithms Kyomin Jung KAIST Applied Algorithm Lab Jan 12, WSAC
Probabilistic Algorithms Michael Sipser Presented by: Brian Lawnichak.
Randomized Algorithms Randomized Algorithms CS648 Lecture 8 Tools for bounding deviation of a random variable Markov’s Inequality Chernoff Bound Lecture.
Complexity 18-1 Complexity Andrei Bulatov Probabilistic Algorithms.
Machine Learning Week 2 Lecture 2.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Perfect and Statistical Secrecy, probabilistic algorithms, Definitions of Easy and Hard, 1-Way FN -- formal definition.
Randomized Computation Roni Parshani Orly Margalit Eran Mantzur Avi Mintz
Probabilistic Complexity. Probabilistic Algorithms Def: A probabilistic Turing Machine M is a type of non- deterministic TM, where each non-deterministic.
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
The Goldreich-Levin Theorem: List-decoding the Hadamard code
Evaluating Hypotheses
Randomized Computation
Quantum Automata Formalism. These are general questions related to complexity of quantum algorithms, combinational and sequential.
13. The Weak Law and the Strong Law of Large Numbers
–Def: A language L is in BPP c,s ( 0  s(n)  c(n)  1,  n  N) if there exists a probabilistic poly-time TM M s.t. : 1.  w  L, Pr[M accepts w]  c(|w|),
Complexity ©D. Moshkovitz 1 And Randomized Computations The Polynomial Hierarchy.
Statistical Hypothesis Testing. Suppose you have a random variable X ( number of vehicle accidents in a year, stock market returns, time between el nino.
1 Dr. Jerrell T. Stracener EMIS 7370 STAT 5340 Probability and Statistics for Scientists and Engineers Department of Engineering Management, Information.
Estimation and Hypothesis Testing. The Investment Decision What would you like to know? What will be the return on my investment? Not possible PDF for.
Hypothesis Testing.
Statistical Review We will be working with two types of probability distributions: Discrete distributions –If the random variable of interest can take.
Topics: Statistics & Experimental Design The Human Visual System Color Science Light Sources: Radiometry/Photometry Geometric Optics Tone-transfer Function.
Approximation Algorithms Pages ADVANCED TOPICS IN COMPLEXITY THEORY.
Trust-Aware Optimal Crowdsourcing With Budget Constraint Xiangyang Liu 1, He He 2, and John S. Baras 1 1 Institute for Systems Research and Department.
Analysis of Algorithms
Statistics for Managers Using Microsoft Excel, 5e © 2008 Pearson Prentice-Hall, Inc.Chap 8-1 Statistics for Managers Using Microsoft® Excel 5th Edition.
1 CS546: Machine Learning and Natural Language Discriminative vs Generative Classifiers This lecture is based on (Ng & Jordan, 02) paper and some slides.
1 Law of large numbers. Central limit theorem. Confidence interval. Hypothesis testing. Types of errors. Lecture 9.
1 2 Probabilistic Computations  Extend the notion of “efficient computation” beyond polynomial-time- Turing machines.  We will still consider only.
. CLASSES RP AND ZPP By: SARIKA PAMMI. CONTENTS:  INTRODUCTION  RP  FACTS ABOUT RP  MONTE CARLO ALGORITHM  CO-RP  ZPP  FACTS ABOUT ZPP  RELATION.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
ENGG 2040C: Probability Models and Applications Andrej Bogdanov Spring Limit theorems.
Chapter 4-5 DeGroot & Schervish. Conditional Expectation/Mean Let X and Y be random variables such that the mean of Y exists and is finite. The conditional.
 2005 SDU Lecture13 Reducibility — A methodology for proving un- decidability.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Pearson Prentice-Hall, Inc.Chap 8-1 Statistics for Managers Using Microsoft® Excel 5th Edition.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Unique Games Approximation Amit Weinstein Complexity Seminar, Fall 2006 Based on: “Near Optimal Algorithms for Unique Games" by M. Charikar, K. Makarychev,
Chapter 5. Continuous Random Variables. Continuous Random Variables Discrete random variables –Random variables whose set of possible values is either.
Section 10.5 Let X be any random variable with (finite) mean  and (finite) variance  2. We shall assume X is a continuous type random variable with p.d.f.
1 Section 13.1 Turing Machines A Turing machine (TM) is a simple computer that has an infinite amount of storage in the form of cells on an infinite tape.
1 Turing Machines and Equivalent Models Section 13.1 Turing Machines.
Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Hypothesis Testing. Suppose we believe the average systolic blood pressure of healthy adults is normally distributed with mean μ = 120 and variance σ.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
1 8.4 Extensions to the Basic TM Extended TM’s to be studied: Multitape Turing machine Nondeterministic Turing machine The above extensions make no increase.
Theory of Computational Complexity Probability and Computing Ryosuke Sasanuma Iwama and Ito lab M1.
Linear Models & Clustering Presented by Kwak, Nam-ju 1.
©The McGraw-Hill Companies, Inc. 2008McGraw-Hill/Irwin Estimation and Confidence Intervals Chapter 9.
PROBABILITY AND COMPUTING RANDOMIZED ALGORITHMS AND PROBABILISTIC ANALYSIS CHAPTER 1 IWAMA and ITO Lab. M1 Sakaidani Hikaru 1.
Theory of Computational Complexity Probability and Computing Chapter Hikaru Inada Iwama and Ito lab M1.
Probabilistic Algorithms
Random Testing: Theoretical Results and Practical Implications IEEE TRANSACTIONS ON SOFTWARE ENGINEERING 2012 Andrea Arcuri, Member, IEEE, Muhammad.
Pushdown Automata.
CONCEPTS OF ESTIMATION
Ch13 Empirical Methods.
Topic 3: Prob. Analysis Randomized Alg.
13. The Weak Law and the Strong Law of Large Numbers
Tomado del libro Cormen
The Polynomial Hierarchy
9. Limit Theorems.
Theorem 9.3 A problem Π is in P if and only if there exist a polynomial p( ) and a constant c such that ICp (x|Π) ≤ c holds for all instances x of Π.
CSE 105 theory of computation
13. The Weak Law and the Strong Law of Large Numbers
Presentation transcript:

Chernoff Bounds, and etc. Presented by Kwak, Nam-ju

Topics A General Form of Chernoff Bounds Brief Idea of Proof for General Form of Chernoff Bounds More Tight form of Chernoff Bounds Application of Chernoff Bounds: Amplification Lemma of Randomized Algorithm Studies Chebyshev’s Inequality Application of Chebyshev’s Inequality Other Considerations

A General Form of Chernoff Bounds Assumption Xi’s: random variables where Xi∈{0, 1} and 1≤i≤n. P(Xi=1)=pi and therefore E[Xi]=pi. X: a sum of n independent random variables, that is, μ: the mean of X

A General Form of Chernoff Bounds When δ >0

Brief Idea of Proof for General Form of Chernoff Bounds Necessary Backgrounds Marcov’s Inequality For any random variable X≥0, When f is a non-decreasing function,

Brief Idea of Proof for General Form of Chernoff Bounds Necessary Backgrounds Upper Bound of M.G.F.

Brief Idea of Proof for General Form of Chernoff Bounds Proof of One General Case (proof)

Brief Idea of Proof for General Form of Chernoff Bounds Proof of One General Case Here, put a value of t which minimize the above expression as follows:

Brief Idea of Proof for General Form of Chernoff Bounds Proof of One General Case As a result,

More Tight form of Chernoff Bounds The form just introduced has no limitation in choosing the value of δ other than that it should be positive. When we restrict the range of the value δ can have, we can have tight versions of Chernoff Bounds.

More Tight form of Chernoff Bounds When 0<δ<1 Compare these results with the upper bo und of the general case:

Application of Chernoff Bounds: Amplification Lemma of Randomized Algorithm Studies A probabilistic Turing machine is a nondeterministic Turing machine in which each nondeterministic step has two choices. (coin-flip step) Error probability: The probability that a certain probabilistic TM produces a wrong answer for each trial. Class BPP: a set of languages which can be recognized by polynomial time probabilistic Turing Machines with an error probability of 1/3.

Application of Chernoff Bounds: Amplification Lemma of Randomized Algorithm Studies However, even though the error probability is over 1/3, if it is between 0 and 1/2 (exclusively), it belongs to BPP. By the amplification lemma, we can construct an alternative probabilistic Turing machine recognizing the same language with an error probability 2-a where a is any desired value. By adjusting the value of a, the error probability would be less than or equal to 1/3.

How to construct the alternative TM? (For a given input x) Application of Chernoff Bounds: Amplification Lemma of Randomized Algorithm Studies How to construct the alternative TM? (For a given input x) Select the value of k. Simulate the original TM 2k times. If more than k simulations result in accept, accept; otherwise, reject. Now, prove how it works.

Xi’s: 1 if the i-th simulation produces a wrong answer; otherwise, 0. Application of Chernoff Bounds: Amplification Lemma of Randomized Algorithm Studies Xi’s: 1 if the i-th simulation produces a wrong answer; otherwise, 0. X: the summation of 2k Xi’s, which means the number of wrongly answered simulations among 2k ones. ε: the error probability X~B(2k, ε) μ=E[X]=2k ε

Application of Chernoff Bounds: Amplification Lemma of Randomized Algorithm Studies P(X>k): the probability that more than hal f of the 2k simulations get a wrong answe r. We will show that P(X>k) can be less tha n 2-a for any a, by choosing k appropriatel y.

Here we set δ as follows: Application of Chernoff Bounds: Amplification Lemma of Randomized Algorithm Studies Here we set δ as follows: Therefore, by the Chernoff Bounds,

To make the upper bound less than or eq ual to 2-a, Application of Chernoff Bounds: Amplification Lemma of Randomized Algorithm Studies To make the upper bound less than or eq ual to 2-a,

Here, we can guarantee the right term is po sitive when 0<ε<1/2. Application of Chernoff Bounds: Amplification Lemma of Randomized Algorithm Studies Here, we can guarantee the right term is po sitive when 0<ε<1/2.

Chebyshev’s Inequality For a random variable X of any probabilistic distribution with mean μ and standard deviation σ, To derive the inequality, utilize Marcov’s in equality.

Application of Chebyshev’s Inequality Use of the Chebyshev Inequality To Calculate 95% Upper Confidence Limits for DDT Contaminated Soil Concentrations at a Using Chebyshev’s Inequality to Determine Sample Size in Biometric Evaluation of Fingerprint Data Superfund Site.

Application of Chebyshev’s Inequality For illustration, assume we have a large body of text, for example articles from a publication. Assume we know that the articles are on average 1000 characters long with a standard deviation of 200 characters. From Chebyshev's inequality we can then deduce that at least 75% of the articles have a length between 600 and 1400 characters (k = 2).

Other Considerations The only restriction Markov’s Inequality impose is that X should be non-negative. It even doesn’t matter whether the standard deviation is infinite or not. e.g. a random variable X with P.D.F. it has a finite mean but a infinite standard deviation.

Other Considerations P.D.F. E[X] Var(x)

Conclusion Chernoff’s Bounds provide relatively nice upper bounds without too much restrictions. With known mean and standard deviation, Chebyshev’s Inequality gives tight upper bounds for the probability that a certain random variable is within a fixed distance from the mean of it.

Conclusion Any question?