KITPC 2011.3 Osamu Watanabe Tokyo Inst. of Tech. Finding Most-Likely Solution of the Perturbed k -Linear-Equation Problem k -Linear-Equation = k LIN 渡辺.

Slides:



Advertisements
Similar presentations
Optimal Lower Bounds for 2-Query Locally Decodable Linear Codes Kenji Obata.
Advertisements

Randomness Conductors Expander Graphs Randomness Extractors Condensers Universal Hash Functions
Routing Complexity of Faulty Networks Omer Angel Itai Benjamini Eran Ofek Udi Wieder The Weizmann Institute of Science.
A threshold of ln(n) for approximating set cover By Uriel Feige Lecturer: Ariel Procaccia.
Inapproximability of MAX-CUT Khot,Kindler,Mossel and O ’ Donnell Moshe Ben Nehemia June 05.
Price Of Anarchy: Routing
Max Cut Problem Daniel Natapov.
Approximation Algorithms Chapter 14: Rounding Applied to Set Cover.
Heuristics for the Hidden Clique Problem Robert Krauthgamer (IBM Almaden) Joint work with Uri Feige (Weizmann)
Wavelength Assignment in Optical Network Design Team 6: Lisa Zhang (Mentor) Brendan Farrell, Yi Huang, Mark Iwen, Ting Wang, Jintong Zheng Progress Report.
1 NP-Complete Problems. 2 We discuss some hard problems:  how hard? (computational complexity)  what makes them hard?  any solutions? Definitions 
Approximating the Domatic Number Feige, Halldorsson, Kortsarz, Srinivasan ACM Symp. on Theory of Computing, pages , 2000.
A polylogarithmic approximation of the minimum bisection Robert Krauthgamer The Hebrew University Joint work with Uri Feige.
1/17 Optimal Long Test with One Free Bit Nikhil Bansal (IBM) Subhash Khot (NYU)
Complexity 16-1 Complexity Andrei Bulatov Non-Approximability.
Computability and Complexity 23-1 Computability and Complexity Andrei Bulatov Search and Optimization.
Why almost all k-colorable graphs are easy A. Coja-Oghlan, M. Krivelevich, D. Vilenchik.
Coloring the edges of a random graph without a monochromatic giant component Reto Spöhel (joint with Angelika Steger and Henning Thomas) TexPoint fonts.
It’s all about the support: a new perspective on the satisfiability problem Danny Vilenchik.
Analysis of Algorithms CS 477/677
Why almost all k-colorable graphs are easy ? A. Coja-Oghlan, M. Krivelevich, D. Vilenchik.
2-Layer Crossing Minimisation Johan van Rooij. Overview Problem definitions NP-Hardness proof Heuristics & Performance Practical Computation One layer:
Computability and Complexity 24-1 Computability and Complexity Andrei Bulatov Approximation.
Job Scheduling Lecture 19: March 19. Job Scheduling: Unrelated Multiple Machines There are n jobs, each job has: a processing time p(i,j) (the time to.
Complexity 1 Mazes And Random Walks. Complexity 2 Can You Solve This Maze?
Finding a maximum independent set in a sparse random graph Uriel Feige and Eran Ofek.
(work appeared in SODA 10’) Yuk Hei Chan (Tom)
Packing Element-Disjoint Steiner Trees Mohammad R. Salavatipour Department of Computing Science University of Alberta Joint with Joseph Cheriyan Department.
1 Joint work with Shmuel Safra. 2 Motivation 3 Motivation.
Eigenvectors of random graphs: nodal domains James R. Lee University of Washington Yael Dekel and Nati Linial Hebrew University TexPoint fonts used in.
Dana Moshkovitz, MIT Joint work with Subhash Khot, NYU.
The Quasi-Randomness of Hypergraph Cut Properties Asaf Shapira & Raphael Yuster.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
The Complexity of Optimization Problems. Summary -Complexity of algorithms and problems -Complexity classes: P and NP -Reducibility -Karp reducibility.
Week 10Complexity of Algorithms1 Hard Computational Problems Some computational problems are hard Despite a numerous attempts we do not know any efficient.
Satisfiable k -CNF formulas above the threshold Danny Vilenchik.
Testing the independence number of hypergraphs
Markov Chains and Random Walks. Def: A stochastic process X={X(t),t ∈ T} is a collection of random variables. If T is a countable set, say T={0,1,2, …
1/19 Minimizing weighted completion time with precedence constraints Nikhil Bansal (IBM) Subhash Khot (NYU)
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
New algorithms for Disjoint Paths and Routing Problems
Minimizing Delay in Shared Pipelines Ori Rottenstreich (Technion, Israel) Joint work with Isaac Keslassy (Technion, Israel) Yoram Revah, Aviran Kadosh.
CSC 413/513: Intro to Algorithms
Why almost all satisfiable k - CNF formulas are easy? Danny Vilenchik Joint work with A. Coja-Oghlan and M. Krivelevich.
Analysis of Boolean Functions and Complexity Theory Economics Combinatorics …
Hongyu Liang Institute for Theoretical Computer Science Tsinghua University, Beijing, China The Algorithmic Complexity.
CHAPTER SIX T HE P ROBABILISTIC M ETHOD M1 Zhang Cong 2011/Nov/28.
TU/e Algorithms (2IL15) – Lecture 11 1 Approximation Algorithms.
Stochastic Streams: Sample Complexity vs. Space Complexity
New Characterizations in Turnstile Streams with Applications
Richard Anderson Lecture 26 NP-Completeness
A Quantum Local Lemma Or Sattath TAU and HUJI Andris Ambainis
Polynomial integrality gaps for
Richard Anderson Lecture 26 NP-Completeness
Possibilities and Limitations in Computation
Computability and Complexity
Subhash Khot Dept of Computer Science NYU-Courant & Georgia Tech
Turnstile Streaming Algorithms Might as Well Be Linear Sketches
Nikhil Bansal, Shashwat Garg, Jesper Nederlof, Nikhil Vyas
ICS 353: Design and Analysis of Algorithms
CSCI B609: “Foundations of Data Science”
The Curve Merger (Dvir & Widgerson, 2008)
Uncertain Compression
Coverage Approximation Algorithms
On the effect of randomness on planted 3-coloring models
Topic 3: Prob. Analysis Randomized Alg.
The Byzantine Secretary Problem
CSCI B609: “Foundations of Data Science”
Contagious sets in random graphs
Presentation transcript:

KITPC Osamu Watanabe Tokyo Inst. of Tech. Finding Most-Likely Solution of the Perturbed k -Linear-Equation Problem k -Linear-Equation = k LIN 渡辺 治 東京工業大学

Preliminaries: Problem & Algorithm Most Likely 3LIN Instance: system of 3-linear equations F generated following 3LIN(n,p,q) Ans: an assignment a = (a1,...,an) that is most likely for F ⇔ minimizing # of broken eqn.s MAX-3LIN MAX-3XSAT

Computational Learning Data Mining Inverse Problem Observed Data reasoning solution randomness e.g., noise s3s3 s1s1 s2s2 s4s4 Problem of Finding Most Likely Solution

Input: x Task: Find a most likely solution s for x Inverse Problem for Computational Learning ← observed data Most Likely Solution Problem max arg Pr [ g ( s, r ) = x ] r s Complexity and Algorithms 1. Optimization problem often NP-hard in the worst case 2.Interesting algorithms are known that works well on average ! x = g ( s, r ) with r : D good scenario

Target Problems System of Linear Equations (mod 2) Instance: system of linear equations F generated randomly under a certain scenario Ans: an assignment for var.s of F that is most likely for the scenario Ex. ・ linear code decoding ・ bisection ・ solving perturbed eqn.s ↑ many applications

random generation model: 3LIN(n,p,q) Generate a eqn. system F for n, p, and q 1. randomly generate a target solution a* = (a1,...,an) ; 2. for each 3-tuple of variables, select it with prob. p and make an eqn. consistent to a* ; 3. flip the righthand side value b of the eqn. with prob. q and then add it to F ; eqn. density noise prob.

3LIN(n,p,q) n : # of variables p : eqn. density param. q : noise prob. some remarks ・ fix our planted solution a* = (a1,...,an) to (+1,...,+1) ・ average # of eqn.s is sat. threshold θ( n ) equations ↑ p = c / n 2

Proposition For some constants c1 and c2, we have ⇒ the planted solution is the unique opt. solution for ML-3LIN instances w.h.p. w.h.p. = prob. 1 - o (1)

Algorithm: simple message passing algo. m = +1 f 1→1 m = 0 f 3→1

problem: Most Likely 3LIN F most likely → a = (a1,...,an) instance generation model: 3LIN( n p q ) noise prob. eqn. density planted solution a* = (+1,..., +1) Summary algHMP( I, MAXSTEP ) 1. Most Likely 2LIN ・ p = c / n for large c. ・ q = 1/2 - ε. ・ I = any one var. 2. Most Likely 3LIN ・ p = c / n log n for large c. ・ q = 1/2 - ε. ・ I = any log n var.s Experiments 平均式数 cn 変数毎登場回数 c 平均式数 cn / log n 変数毎登場回数 cn / log n 2

1. Most Likely 2LIN Problem Algorithm power method

p ・ (1 - q) p・qp・q A = 総和は平均で n ・ p ・ (1 - 2q) = p ・ 2 ε 平均で最大固有値は 2 ε np q = ½ - ε

最大固有ベクトル 1 未満の定数

Most Likely 2LIN Problem Algorithm

3. Most Likely 3 LIN Problem Algorithm

3. Most Likely 3 LIN Problem Algorithm log n correct values 平均式数 cn 変数毎登場回数 c 平均式数 cn / log n 変数毎登場回数 cn / log n 2

Theorem p > c / n log n and q < 1/2 - ε ⇒ the modified algorithm works w.h.p. p > c / n log n too big!? But ... O.W, M.O, & A.Coja-Oghlan

Simple Case Analysis 1. Consider the case q = 0; that is, no perturbation occurs. 2. Execution of the algorithm with MAXSTEP = n and I = all possible { xi, xj } and the correct initial assignment to { xi, xj } When the execution succeeds w.h.p. ? Success threshold for p ? nonzero message is sent to all var.s

Propagation Connectivity new hypergraph connectivity 3LIN instance F Hypergraph H = ( V, HE ) 3LIN( n, p, q ) ⇔ H( n, p ) marking process from { 1, 2 }

Propagation Connectivity - connectivity notion on hypergraphs 3-uniform hyp.graphs marking process from (1,2) Def. hyp.graph H is propagation connected ⇔ ∃ starting vertex pair s.t. the above process marks all vertices from the starting pair

H ( n, p ) : the standard distribution of 3-uniform random hyper graphs where n = # of vertices, and p = hyp.edge prob. whp = with. prob. 1 - o (1) c n ln n = Thm. Let H denote a random graph following the distribution H ( n, p ) for p. Lower Bound: c < 0.16 ⇒ H is not prop. connected whp Upper Bound: c > 0.25 ⇒ H is prop. connected whp. c n ln n =

Proof Outline Lower Bound: c < 0.16 Lemma. For any p = for any c < 0.16, consider the propagation process on n vertices from any fixed vertex pair ( v 1, v 2 ). Then for some d > 0, we have Pr[ ] = o ( n ) -2-2 T ( v 1, v 2 ) > d ln n c n ln n

Proof Outline Upper Bound: c > 0.25 Def. ( v 1, v 2 ) is good ⇔ < T ( v 1, v 2 ) ( ln n ) 3 Lemma. For any p = for any c > 0, consider the propagation process on n vertices from any fixed vertex pair ( v 1, v 2 ). Pr[ < T ( v 1, v 2 ) < n - 1 ] = n ( ln n ) 3 - Ω(1) c n ln n

Proof Outline Upper Bound: c > 0.25 N = # of good starting pairs of vertices Lemma B. If E[ N ] > n, then we have Pr[ N > 0 ] > 1 - o (1). δ V[ N ] is small because two processes from different starting pairs are not so much related during steps. ( ln n ) 3 Lemma A. For any p = for any c > 0.25 consider the propagation process on n vertices. Then for some δ > 0, we have E[ N ] > n. δ c n ln n

some remarks ・ fix our planted solution a* = (a1,...,an) to (+1,...,+1) ・ average # of eqn.s is 3LIN(n,p,q) n : # of variables p : eqn. density param. q : noise prob. Summary [results] Algorithm's success threshold seems

3LIN(n,p,q) n : # of variables p : eqn. density param. q : noise prob. some remarks ・ fix our planted solution a* = (a1,...,an) to (+1,...,+1) ・ average # of eqn.s is Discussions [results] Algorithm's success threshold seems random walk based algorithm, also [R.Montegegro] [Feige-Ofek '05] One step G.elimination + spectral algo. for 2LIN sat. threshold θ( n ) equations ↑ p = c / n 2

References A. Coja-Oghlan, M. Onsjo, and O. Watanabe, Propagation connectivity of random hypergraphs, in Proc. 13th APPROX 2010 and 14th RANDOM 2010, LNCS 6302, , U. Feige and E. Ofek, Spectral techniques applied to sparse random graphs, Random Structures and Algorithms 27, , M. Onsjo and O. Watanabe, A simple message passing algorithm for graph partition problem, in Proc. 17th ISAAC'06, LNCS 4288, , 2006.