Download presentation
Presentation is loading. Please wait.
1
Testing with Alternative Distances
Gautam βGβ Kamath FOCS 2017 Workshop: Frontiers in Distribution Testing October 14, 2017 Based on joint works with Jayadev Acharya Cornell Constantinos Daskalakis MIT John Wright MIT
2
The story so farβ¦ Test whether π=π versus β 1 π,π β₯π
β 1 π,π >π 0< β 1 π,π β€π π=π Test whether π=π versus β 1 π,π β₯π Domain of [π] Success probability β₯2/3 Goal: Strongly sublinear sample complexity π( π 1βπΎ ) for some πΎ>0 Identity testing (samples from π, known π) Ξ( π / π 2 ) samples [BFFKRβ01, Pβ08, VVβ14] Closeness testing (samples from π, π) Ξ( max { π 2/3 / π 4/3 , π / π 2 } ) samples [BFRSWβ00, Vβ11, CDVVβ14]
3
Generalize: Different Distances
π=π or β 1 π,π β₯π? π 1 π,π β€ π 1 or π 2 π,π β₯ π 2 ?
4
Generalize: Different Distances
π 1 π,π β€ π 1 or π 2 π,π β₯ π 2 ? Are π and π π 1 -close in π 1 (.,.), or π 2 -far in π 2 (.,.)? Distances of interest: β 1 , β 2 , π 2 , KL, Hellinger Classic identity testing: π 1 =0, π 2 = β 1 Can we characterize sample complexity for each pair of distances? Which distribution distances are sublinearly testable? [DKWβ18] Wait, butβ¦ why?
5
Wait, but⦠why? Tolerance for model misspecification
Useful as a proxy in classical testing problems π 1 as π 2 distance is useful for composite hypothesis testing Monotonicity, independence, etc. [ADKβ15] Other distances are natural in certain testing settings π 2 as Hellinger distance is sometimes natural in multivariate settings Bayes networks, Markov chains [DPβ17,DDGβ17] Costisβ talk
6
Wait, but⦠why? Tolerance for model misspecification
Useful as a proxy in classical testing problems π 1 as π 2 distance is useful for composite hypothesis testing Monotonicity, independence, etc. [ADKβ15] Other distances are natural in certain testing settings π 2 as Hellinger distance is sometimes natural in multivariate settings Bayes networks, Markov chains [DPβ17,DDGβ17] Costisβ talk
7
Tolerance Is π equal to π, or are they far from each other?
ideal model Is π equal to π, or are they far from each other? But why do we know π exactly? Models are inexact Measurement errors Imprecisions in nature π, π may be βphilosophicallyβ equal, but not literally equal When can we test π 1 π,π β€ π 1 versus β 1 π,π β₯π? CLT approximationsβ¦ Read data point wrongβ¦ observed model
8
Tolerance π 1 π,π β€ π 1 vs. β 1 π,π β₯π? β 1 π,π β€π/2 vs. β 1 π,π β₯π?
β 1 π,π >π π 2 < β 1 π,π β€π β 1 π,π β€ π 2 β 1 π,π >π 0< β 1 π,π β€π π=π β 1 π,π >π π 2 π,π > π 2 /2 β 1 π,π β€π π 1 π,π β€ π 1 vs. β 1 π,π β₯π? What π 1 ? How about β 1 ? β 1 π,π β€π/2 vs. β 1 π,π β₯π? No! Ξ(π/ log π ) samples [VVβ10] Chill out, relaxβ¦ π 2 -distance: π 2 π,π = πβΞ£ π π β π π π π Cauchy-Schwarz: π 2 π,π β₯ β 1 2 (π,π) π 2 π,π β€ π 2 /4 vs. β 1 π,π β₯π? Yes! π( π / π 2 ) samples [ADKβ15] π 2 π,π β€ π 2 /2
9
Details for a π 2 -Tolerant Tester
Goal: Distinguish (i) π 2 π,π β€ π versus (ii) β 1 2 π,π β₯ π 2 Draw ππππ π ππ(π) samples from π (βPoissonizationβ) π π : number of appearances of symbol π π π ~ ππππ π ππ πβ
π π π π βs are now independent! Statistic: π= πβΞ£ π π βπβ
π π 2 β π π πβ
π π Acharya, Daskalakis, K. Optimal Testing for Properties of Distributions. NIPS 2015.
10
Details for a π 2 -Tolerant Tester
Goal: Distinguish (i) π 2 π,π β€ π versus (ii) β 1 2 π,π β₯ π 2 Statistic: π= πβ[π] π π βπβ
π π 2 β π π πβ
π π π π : # of appearances of π; π: # of samples πΈ π =πβ
π 2 (π, π) (i): πΈ π β€πβ
π 2 2 , (ii): πΈ π β₯πβ
π 2 Can bound variance of π with some work Need to avoid low prob. elements of π Either ignore π such that π π β€ π 2 10π ; or Mix lightly (π( π 2 )) with uniform distribution (also in [Gβ16]) Apply Chebyshevβs inequality Side-Note: Pearsonβs π 2 -test uses statistic π π π βπ β
π π 2 πβ
π π Subtracting π π in the numerator gives an unbiased estimator and importantly may hugely decrease variance [Zeltermanβ87] [VVβ14, CDVVβ14, DKNβ15] Acharya, Daskalakis, K. Optimal Testing for Properties of Distributions. NIPS 2015.
11
Tolerant Identity Testing
Harder (Implicit in [DKβ16]) Daskalakis, K., Wright. Which Distribution Distances are Sublinearly Testable? SODA 2018.
12
Tolerant Testing Takeaways
Can handle β 2 or π 2 tolerance at no additional cost Ξ( π / π 2 ) samples KL, Hellinger, or β 1 tolerance are expensive Ξ(π/ log π ) samples KL result based off hardness of entropy estimation Closeness testing (π unknown): Even π 2 tolerance is costly! Only β 2 tolerance is free Proven via hardness of β 1 -tolerant identity testing Since π is unknown, π 2 is no longer a polynomial
13
Application: Testing for Structure
Composite hypothesis testing Test against a class of distributions! πβπΆ versus β 1 π,πΆ >π Example: πΆ = all monotone distributions π π βs are monotone non-increasing Others: unimodality, log-concavity, monotone hazard rate, independence All can be tested in Ξ π / π 2 samples [ADKβ15] Same complexity as vanilla uniformity testing! min πβπΆ β 1 (π,π)
14
Testing by Learning Goal: Distinguish πβπΆ from β 1 π,πΆ >π
Learn-then-Test: Learn hypothesis πβπΆ such that πβπΆβ π 2 π,π β€ π 2 /2 (needs cheap βproper learnerβ in π 2 ) β 1 π,πΆ >πβ β 1 π,π >π (automatic since πβπΆ) Perform βtolerant testingβ Given sample access to π and description of π, distinguish π 2 π,π β€ π 2 /2 from β 1 π,π >π Tolerant testing (step 2) is π( π / π 2 ) NaΓ―ve approach (using β 1 instead of π 2 ) would require Ξ©(π/ log π ) Proper learners in π 2 (step 1)? Claim: This is cheap NaΓ―ve approach of using L1 tolerant testing would require n/log n samples. This is why different distances are important!
15
Hellinger Testing Change π 2 instead of π 1
Hellinger distance: π» 2 π,π = 1 2 πβ[π] π π β π π 2 Between linear and quadratic relationship with β 1 π» 2 π,π β€ β 1 π,π β€π»(π,π) Natural distance when considering a collection of iid samples Comes up in some multivariate testing problems 2:55) Testing π=π vs. π» π,π β₯π? Trivial results via β 1 testing Identity: π( π / π 4 ) samples Closeness: π(max { π 2/3 / π 8/3 , π / π 4 } ) samples
16
Hellinger Testing But you can do better!
Testing π=π vs. π» π,π β₯π? Trivial results via β 1 testing Identity: π( π / π 4 ) samples Closeness: π(max { π 2/3 / π 8/3 , π / π 4 } ) samples But you can do better! Identity: Ξ( π / π 2 ) samples No extra cost for β 2 or π 2 tolerance either! Closeness: Ξ(min { π 2/3 / π 8/3 , π 3/4 / π 2 } ) samples LB and previous UB in [DKβ16] Similar chi-squared statistics as [ADKβ15] and [CDVVβ14] Some tweaks and more careful analysis to handle Hellinger Daskalakis, K., Wright. Which Distribution Distances are Sublinearly Testable? SODA 2018.
17
Miscellanea π=π vs. πΎπΏ π,π β₯π?
Trivially impossible, due to ratio between π π and π π π π =πΏ, π π =0, πΏβ0 Upper bounds for Ξ©(π/ log π ) testing problems? i.e., πΎπΏ π,π β€ π 2 /4 vs. β 1 π,π β₯π? Use estimators mentioned in Jiantaoβs talk
18
Thanks!
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.