1 On approximating the number of relevant variables in a function Dana Ron & Gilad Tsur Tel-Aviv University.

Slides:



Advertisements
Similar presentations
Lower Bounds for Additive Spanners, Emulators, and More David P. Woodruff MIT and Tsinghua University To appear in FOCS, 2006.
Advertisements

Xiaoming Sun Tsinghua University David Woodruff MIT
Tight Lower Bounds for the Distinct Elements Problem David Woodruff MIT Joint work with Piotr Indyk.
On the degree of symmetric functions on the Boolean cube Joint work with Amir Shpilka.
A threshold of ln(n) for approximating set cover By Uriel Feige Lecturer: Ariel Procaccia.
Computational Privacy. Overview Goal: Allow n-private computation of arbitrary funcs. –Impossible in information-theoretic setting Computational setting:
The Communication Complexity of Approximate Set Packing and Covering
Gillat Kol joint work with Ran Raz Locally Testable Codes Analogues to the Unique Games Conjecture Do Not Exist.
Ioannis Caragiannis, Jason A. Covey, Michal Feldman, Christopher M. Homan, Christos Kaklamanis, Nikos Karanikolask, Ariel D. Procaccia, Je ff rey S. Rosenschein.
Distributional Property Estimation Past, Present, and Future Gregory Valiant (Joint work w. Paul Valiant)
An Ω(n 1/3 ) Lower Bound for Bilinear Group Based Private Information Retrieval Alexander Razborov Sergey Yekhanin.
Learning Juntas Elchanan Mossel UC Berkeley Ryan O’Donnell MIT Rocco Servedio Harvard.
1 Distribution-free testing algorithms for monomials with a sublinear number of queries Elya Dolev & Dana Ron Tel-Aviv University.
Distribution-free testing algorithms for monomials with a sublinear number of queries Elya Dolev & Dana Ron Tel-Aviv University.
Approximating Average Parameters of Graphs Oded Goldreich, Weizmann Institute Dana Ron, Tel Aviv University.
Infinite Horizon Problems
Proclaiming Dictators and Juntas or Testing Boolean Formulae Michal Parnas Dana Ron Alex Samorodnitsky.
CPSC 689: Discrete Algorithms for Mobile and Wireless Systems Spring 2009 Prof. Jennifer Welch.
Testing of Clustering Noga Alon, Seannie Dar Michal Parnas, Dana Ron.
Sublinear Algorithms for Approximating Graph Parameters Dana Ron Tel-Aviv University.
Exact Learning of Boolean Functions with Queries Lisa Hellerstein Polytechnic University Brooklyn, NY AMS Short Course on Statistical Learning Theory,
Michael Bender - SUNY Stony Brook Dana Ron - Tel Aviv University Testing Acyclicity of Directed Graphs in Sublinear Time.
Testing Metric Properties Michal Parnas and Dana Ron.
On Proximity Oblivious Testing Oded Goldreich - Weizmann Institute of Science Dana Ron – Tel Aviv University.
On Testing Convexity and Submodularity Michal Parnas Dana Ron Ronitt Rubinfeld.
1 On the Benefits of Adaptivity in Property Testing of Dense Graphs Joint work with Mira Gonen Dana Ron Tel-Aviv University.
1 Algorithmic Aspects in Property Testing of Dense Graphs Oded Goldreich – Weizmann Institute Dana Ron - Tel-Aviv University.
1 On the Benefits of Adaptivity in Property Testing of Dense Graphs Joint works with Mira Gonen and Oded Goldreich Dana Ron Tel-Aviv University.
Integer Programming Difference from linear programming –Variables x i must take on integral values, not real values Lots of interesting problems can be.
DAST 2005 Week 4 – Some Helpful Material Randomized Quick Sort & Lower bound & General remarks…
On Testing Computability by small Width OBDDs Oded Goldreich Weizmann Institute of Science.
Finding a maximum independent set in a sparse random graph Uriel Feige and Eran Ofek.
In a World of BPP=P Oded Goldreich Weizmann Institute of Science.
Dana Moshkovitz, MIT Joint work with Subhash Khot, NYU.
Correlation testing for affine invariant properties on Shachar Lovett Institute for Advanced Study Joint with Hamed Hatami (McGill)
A Property Testing Double-Feature of Short Talks Oded Goldreich Weizmann Institute of Science Talk at Technion, June 2013.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
Quantum Computing MAS 725 Hartmut Klauck NTU TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A.
Streaming Algorithms Piotr Indyk MIT. Data Streams A data stream is a sequence of data that is too large to be stored in available memory Examples: –Network.
Transitive-Closure Spanner of Directed Graphs Kyomin Jung KAIST 2009 Combinatorics Workshop Joint work with Arnab Bhattacharyya MIT Elena Grigorescu MIT.
Edge-disjoint induced subgraphs with given minimum degree Raphael Yuster 2012.
The Integers. The Division Algorithms A high-school question: Compute 58/17. We can write 58 as 58 = 3 (17) + 7 This forms illustrates the answer: “3.
Asymptotic Analysis-Ch. 3
Testing the independence number of hypergraphs
Analysis of Algorithms CS 477/677 Instructor: Monica Nicolescu Lecture 7.
Finding Correlations in Subquadratic Time Gregory Valiant.
Eric Allender Rutgers University Curiouser and Curiouser: The Link between Incompressibility and Complexity CiE Special Session, June 19, 2012.
Approximation Algorithms based on linear programming.
Sketching complexity of graph cuts Alexandr Andoni joint work with: Robi Krauthgamer, David Woodruff.
Theory of Computational Complexity M1 Takao Inoshita Iwama & Ito Lab Graduate School of Informatics, Kyoto University.
Randomness and Computation
New Characterizations in Turnstile Streams with Applications
Dana Ron Tel Aviv University
On Testing Dynamic Environments
Vitaly Feldman and Jan Vondrâk IBM Research - Almaden
On Approximating the Number of Relevant Variables in a Function
From dense to sparse and back again: On testing graph properties (and some properties of Oded)
Lecture 18: Uniformity Testing Monotonicity Testing
NP-Completeness Yin Tat Lee
Background: Lattices and the Learning-with-Errors problem
COMS E F15 Lecture 2: Median trick + Chernoff, Distinct Count, Impossibility Results Left to the title, a presenter can insert his/her own image.
Nikhil Bansal, Shashwat Garg, Jesper Nederlof, Nikhil Vyas
Pseudo-derandomizing learning and approximation
The Curve Merger (Dvir & Widgerson, 2008)
NP-Completeness Yin Tat Lee
The Subgraph Testing Model
Every set in P is strongly testable under a suitable encoding
Complexity Theory: Foundations
Presentation transcript:

1 On approximating the number of relevant variables in a function Dana Ron & Gilad Tsur Tel-Aviv University

2 Computing num of relevant variables Consider f : {0,1} n  {0,1} A variable x i is relevant w.r.t. f if exists assignment y 1 …y n  {0,1} n s.t. f(y 1 …y i …y n )  f(y 1 …  y i …y n ) Denote num of relevant variables by r(f). f y f(y) Given query access to f: Computing r(f) exactly may require exp(n) many queries (e.g. hard to distinguish between all-0 function (r(f)=0) and random function that gives value 1 to a single point (r(f)=n) ) 0 0 f(y)=1

3 Relaxed notions One natural relaxation: Compute value r’ s.t. r(f)/B ≤ r’ ≤ B r(f). As previous example shows, requires exp(n) queries. Another relaxation suggested by Property Testing (PT): Given integer k and 0 <  < 1, distinguish (w.h.p) between:  f has at most k relevant variables (is a k-junta)  f is  -far from every k-junta (i.e. must modify f on more than  -fraction of domain to get k-junta) Query complexity of testing k-juntas: O(k log k) [Blais] and  (k) [Chockler, Gutfreund] With high constant prob

4 Relaxed notions Question: Go below Õ(k) if combine two relaxations? Given integer k and 0 < ,  < 1, distinguish between:  f is a k-junta  f is  -far from every (1+  )k-junta Question raised in [Fischer,Kindler,Ron,Safra,Samorodnitsky] recently considered in [Blais,Brody,Matulef]: Using their (cool) communication complexity technique, give l.b.  ( min{(k/t) 2,k} ) for distinguishing between k-juntas and  - far from (k+t)-juntas (for constant  ). In our (multiplicative) formulation: cannot go below linear in k for  = O(1/k 1/2 ) What about larger , in particular constant (e.g. 1)?

5 Relaxed notions – our results (I) Thm 1. Distinguishing between: f is a k-junta and f is  -far from every (1+  )k-junta requires  (k/log k) queries for constant  and . Thm 2. Distinguishing between: f is a k-junta and f is  -far from every (1+  )k-junta can be done by performing O(k  -1  -2 log(1/  ) ) queries

6 Relaxed notions – our results (II) Thm 3. For f that is linear, distinguishing between: f has at most k relevant vars and f has more than (1+  )k vars can be done by performing O(log(1/  )/  2 ) queries. Given that relaxed PT (const  ) is not much easier than standard PT (  =0), consider another relaxation: Compute (approx) num of variables for restricted classes of function. In particular linear functions and, more generally, degree-d polynomials (over GF(2) n ). For linear functions: distinguishing between at most k relevant vars and more than k+1:  (min{k,n-k}) [BBM]

7 Relaxed notions – our results (II) Thm 5. For d < log(k) and const , distinguishing between: f is a degree-d poly with k relevant variables and f is  -far from every degree-d poly with 2k variables requires  (2 d /d) Is exp(d) necessary (for d < log(k) )? Thm 4. For f that is degree-d poly, distinguishing between: f has at most k relevant vars and f has more than (1+  )k vars can be done by performing O(2 d log(1/  )/  2 ) queries.

8 Upper Bounds Thm 3. For f that is linear, distinguishing between: f has at most k relevant vars and f has more than (1+  )k vars can be done by performing O(log(1/  )/  2 ) queries. Algorithm constructs set(s) S of vars by selecting each var (independently) w.p. 1/2k. x 1 x 2 x 3 x 4 x 5 x 6 x 7 x 8 x 9 x 10 Observation 1: If f has >(1+  )k relevant vars then prob. S contains some relevant var is significantly larger than if f has  k relevant vars (by additive term of  (  )) Observation 2: Can distinguish between “empty” S and “non-empty” S (for linear f) by performing const num of queries: Compare value of f on pairs of assignments that differ only on S (random assignments to S) f( ) vs f( ) (e.g., f(x)=x 5 +x 8 will show difference)

9 Upper Bounds Thm 4. For f that is degree-d poly, distinguishing between: f has at most k relevant vars and f has more than (1+  )k vars can be done by performing O(2 d log(1/  )/  2 ) queries. Same as for linear functions except that “test for relevant variables” uses O(2 d ) queries (influence of each variable is at least 2 -d ) Thm 2. Distinguishing between: f is a k-junta and f is  -far from every (1+  )k-junta can be done by performing O((k/  )( log(1/  )/  2 ) )queries. Same as for linear functions except that “test for relevant variables” uses O(k/  ) queries. Analysis more subtle since need to lower bound influence of sets of variables.

10 Lower Bounds Proof by reduction from Distinct Elements Problem: Given query access to string s of length t, approximate num of distinct elements in s (denoted d(s)) [Valiant&Valiant] prove l.b. of  (t/log t) on distinguishing btwn ≤ t/16 and >t/2 distinct elements. Will use to give l.b. of  (k/log k) on distinguishing btwn k-junta and  (1)-far from 2k-junta. Show for k=n/8 (extends to any k by padding), and use t=n. Thm 1. Distinguishing between: f is a k-junta and f is  -far from every (1+  )k-junta requires  (k/log k) queries for constant  and .

11 The Reduction Given string s=s[0],s[1],…,s[n-1] with d(s) ≤ n- ℓ distinct elements in { ℓ +1,…,n}. Define f s by f s (y) = f(j,y’) = y s[j] Example: s = f s ( ) = Claim: (1) f s is a (d(s)+log(n))-junta. (2) f s is  (1)-far from every (d(s)/2)-junta (d(s)=  (n)) (3) A query to f s can be answered by performing a single query to s. Let ℓ = log(n). Think of y  {0,1} n = {0,1} ℓ x {0,1} n- ℓ as y  [n] x {0,1} n- ℓ so y=(j,y’) (e.g = (2,01) ) {0,…,n-1}

12 Lower Bounds Claim: (1) f s is a (d(s)+log(n))-junta. (2) f s is  (1)-far from every (d(s)/2)-junta (d(s)=  (n)) (3) A query to f s can be answered by performing a single query to s. Combining claim with [VV]’s l.b., we get Thm. 1 (for k=  (n)). Thm 1. Distinguishing between: f is a k-junta and f is  -far from every (1+  )k-junta requires  (k/log k) queries for constant  and . Thm 5. For d < log(k) and const , distinguishing between: f is a degree-d poly with k relevant variables and f is  -far from every degree-d poly with 2k variables requires  (2 d /d) Proof of l.b. for degree-d polynomials: also applies reduction from DE but takes care that functions be degree-d polynomials.

13 Summary Consider relaxed notions of computing number of relevant variables For “multiplicative relaxation” of PT show that can do a little better than “standard” PT, but cannot do much better. For restricted function classes: linear functions and degree-d polynomials can do much better. (For monotone functions, cannot). What about other classes of functions?

14 Thanks