Invertible Zero-Error Dispersers and Defective Memory with Stuck-At Errors Ariel Gabizon Ronen Shaltiel.

Slides:



Advertisements
Similar presentations
Optimal Lower Bounds for 2-Query Locally Decodable Linear Codes Kenji Obata.
Advertisements

1+eps-Approximate Sparse Recovery Eric Price MIT David Woodruff IBM Almaden.
PRG for Low Degree Polynomials from AG-Codes Gil Cohen Joint work with Amnon Ta-Shma.
An Introduction to Randomness Extractors Ronen Shaltiel University of Haifa Daddy, how do computers get random bits?
Deterministic Extractors for Small Space Sources Jesse Kamp, Anup Rao, Salil Vadhan, David Zuckerman.
Computing with adversarial noise Aram Harrow (UW -> MIT) Matt Hastings (Duke/MSR) Anup Rao (UW)
Linear-Degree Extractors and the Inapproximability of Max Clique and Chromatic Number David Zuckerman University of Texas at Austin.
Randomness Extractors: Motivation, Applications and Constructions Ronen Shaltiel University of Haifa.
Detection of Algebraic Manipulation with Applications to Robust Secret Sharing and Fuzzy Extractors Ronald Cramer, Yevgeniy Dodis, Serge Fehr, Carles Padro,
Short seed extractors against quantum storage Amnon Ta-Shma Tel-Aviv University 1.
Extracting Randomness From Few Independent Sources Boaz Barak, IAS Russell Impagliazzo, UCSD Avi Wigderson, IAS.
Efficient Non-Malleable Codes and Key-derivations against Poly-size Tampering Circuits PRATYAY MUKHERJEE (Aarhus University) Joint work with Sebastian.
PRATYAY MUKHERJEE Aarhus University Joint work with
How to get more mileage from randomness extractors Ronen Shaltiel University of Haifa.
Approximate List- Decoding and Hardness Amplification Valentine Kabanets (SFU) joint work with Russell Impagliazzo and Ragesh Jaiswal (UCSD)
Simple extractors for all min- entropies and a new pseudo- random generator Ronen Shaltiel Chris Umans.
Quantum Information and the PCP Theorem Ran Raz Weizmann Institute.
Noise, Information Theory, and Entropy (cont.) CS414 – Spring 2007 By Karrie Karahalios, Roger Cheng, Brian Bailey.
parity bit is 1: data should have an odd number of 1's
Applied Algorithmics - week7
Locally Decodable Codes from Nice Subsets of Finite Fields and Prime Factors of Mersenne Numbers Kiran Kedlaya Sergey Yekhanin MIT Microsoft Research.
Probabilistic verification Mario Szegedy, Rutgers www/cs.rutgers.edu/~szegedy/07540 Lecture 4.
Simple Affine Extractors using Dimension Expansion. Matt DeVos and Ariel Gabizon.
Derandomized parallel repetition theorems for free games Ronen Shaltiel, University of Haifa.
Lecture 40 CSE 331 Dec 11, Announcements Solutions to HW 10 and graded HW 9 at end of the lecture Review session on Monday: see blog for details.
Linear-time encodable and decodable error-correcting codes Daniel A. Spielman Presented by Tian Sang Jed Liu 2003 March 3rd.
Quantum Error Correction Michele Mosca. Quantum Error Correction: Bit Flip Errors l Suppose the environment will effect error (i.e. operation ) on our.
The Goldreich-Levin Theorem: List-decoding the Hadamard code
3-source extractors, bi-partite Ramsey graphs, and other explicit constructions Boaz barak rOnen shaltiel Benny sudakov avi wigderson Joint work with GUY.
Derandomizing LOGSPACE Based on a paper by Russell Impagliazo, Noam Nissan and Avi Wigderson Presented by Amir Rosenfeld.
Lossless data compression Lecture 1. Data Compression Lossless data compression: Store/Transmit big files using few bytes so that the original files.
1 Streaming Computation of Combinatorial Objects Ziv Bar-Yossef U.C. Berkeley Omer Reingold AT&T Labs – Research Ronen.
On the Complexity of Approximating the VC Dimension Chris Umans, Microsoft Research joint work with Elchanan Mossel, Microsoft Research June 2001.
Hamming Code Rachel Ah Chuen. Basic concepts Networks must be able to transfer data from one device to another with complete accuracy. Data can be corrupted.
Simulating independence: new constructions of Condensers, Ramsey Graphs, Dispersers and Extractors Boaz Barak Guy Kindler Ronen Shaltiel Benny Sudakov.
Correlation testing for affine invariant properties on Shachar Lovett Institute for Advanced Study Joint with Hamed Hatami (McGill)
Great Theoretical Ideas in Computer Science.
Information and Coding Theory Linear Block Codes. Basic definitions and some examples. Juris Viksna, 2015.
Institute for Experimental Mathematics Ellernstrasse Essen - Germany On STORAGE Systems A.J. Han Vinck June 2004.
Great Theoretical Ideas in Computer Science.
CS717 Algorithm-Based Fault Tolerance Matrix Multiplication Greg Bronevetsky.
§6 Linear Codes § 6.1 Classification of error control system § 6.2 Channel coding conception § 6.3 The generator and parity-check matrices § 6.5 Hamming.
DIGITAL COMMUNICATIONS Linear Block Codes
Information Theory Linear Block Codes Jalal Al Roumy.
The parity bits of linear block codes are linear combination of the message. Therefore, we can represent the encoder by a linear system described by matrices.
Pseudo-random generators Talk for Amnon ’ s seminar.
Error-Correcting Codes and Pseudorandom Projections Luca Trevisan U.C. Berkeley.
Lecture 20 CSE 331 July 30, Longest path problem Given G, does there exist a simple path of length n-1 ?
Richard Cleve DC 2117 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Lecture (2011)
Umans Complexity Theory Lecturess Lecture 11: Randomness Extractors.
Block Coded Modulation Tareq Elhabbash, Yousef Yazji, Mahmoud Amassi.
Hamming Code In 1950s: invented by Richard Hamming
IERG6120 Lecture 22 Kenneth Shum Dec 2016.
General Strong Polarization
Pseudorandomness when the odds are against you
General Strong Polarization
Linear sketching with parities
When are Fuzzy Extractors Possible?
The Curve Merger (Dvir & Widgerson, 2008)
Linear sketching over
When are Fuzzy Extractors Possible?
Non-Malleable Extractors New tools and improved constructions
Linear sketching with parities
Indistinguishability by adaptive procedures with advice, and lower bounds on hardness amplification proofs Aryeh Grinberg, U. Haifa Ronen.
CS151 Complexity Theory Lecture 10 May 2, 2019.
How many deleted bits can one recover?
General Strong Polarization
General Strong Polarization
Error Correction Coding
Zeev Dvir (Princeton) Shachar Lovett (IAS)
Presentation transcript:

Invertible Zero-Error Dispersers and Defective Memory with Stuck-At Errors Ariel Gabizon Ronen Shaltiel

Defective memory with stuck-at errors msg z Enc *0**1*010 Dec 010 Reading proc. Dec must retrieve z, doesn’t know which bits are stuck. n-bit memory with ≤ p∙n “stuck bits” (can’t write) Goal: store msg z. Give explicit schemes.  p: store n-p∙n-log c n bits. Rate > 1-p-o(1). Storage proc. Enc knows which bits are stuck. Earlier work [KuzTsy] shows existence of non-explicit schemes. Error-correcting codes where encoder knows in advance which bits will be corrupted. In standard codes we expect* rate 1-h(2p)-o(1).

We can also handle few additional errs Enc Dec Noise Give explicit schemes. Handle o(n ½ ) unknown errs. Rate > 1-p-o(1). Storage proc. Enc knows which bits are stuck. Reading proc. Dec must retrieve z, doesn’t know which bits are stuck. n-bit memory with ≤ p∙n “stuck bits” (can’t write) Goal: store msg z. msg z *0**1*010

`Defect pattern’:Subset S µ [n] of size ≤p∙n, String a 2 {0,1} p∙n. -Encoding proc. Enc, on message z 2 {0,1} m and pattern S,a produces x 2 {0,1} n with x| S = a. -Decoding proc. Dec such that Dec(x) = z Dec doesn’t depend on S,a m/n is the rate of the scheme. Explicit: Enc expected poly-time *, Dec poly-time. *Simply because that’s what we get. Dfn of memory scheme (for parameter 0<p<1):

Write Once Memory (k-WOM) n-bit memory initialized to zeros. If we raise a bit, it is “stuck-at” 1. Goal: Reuse memory k times. (think k=2). Stuck-at scheme => 2-WOM, gives opt. rate ≈ h(p) + (1-p). Intuitively WOM is easier to achieve. – Defect pattern is not adversarial. – Can pass o(n) bits from enc to dec by reserving first o(n) bits. (errorless channel from enc to dec). Starting point for this work: – Shpilka: Linear seeded extractors => 2-WOM. – We generalize: seedless zero-error disp. for bit-fixing sources  schemes for stuck-at errors

Bit-fixing sources and dispersers: S µ [n] – subset of size p∙n. a – string in {0,1} p∙n X S,a, set of strings in {0,1} n, s.t. x| S =a. Call such a set a bit-fixing source. |X S,a |=2 k for k=n-p∙n (# unfixed bits). Dfn: D:{0,1} n  {0,1} m is a zero-error bit-fixing disperser for threshold k if  S,a as above D(X S,a )={0,1} m. (Explicit: poly-time). Goal: achieve large m. Obviously m ≤ k = n-p∙n. Efficiently invertible: given S,a and output z, can find x 2 X S,a with D(x) = z in expected* poly(n)-time. *Simply because that’s what we get. *0**1**** D

Dispersers => memory schemes (and vice-versa) Bit-fixing disperser D:{0,1} n  {0,1} m. Encoding: Given message z 2 {0,1} m and defect pattern S,a, Enc finds x 2 X S,a with D(x) = z. Decoding: Given x, Dec computes D(x)=z. Output len. m=(1-o(1))∙k => optimal rate. Explicitness of scheme follows if D is explicit and efficiently invertibile. This work: construct such dispersers (for k>log c n). Improve [GabSha08] where m=  (k).

Construction of zero-error disperser

Seeded zero-error dispersers Dfn: E(x,y) is a seeded bit-fixing disperser for threshold k, if  z, and bit-fixing source X S,a with k unfixed bits, there exist: -a seed y, and -x 2 X S,a such that E(x,y) = z. E is efficiently invertible if x,y can be found in expected poly(n)-time given z,S,a. Known [RazReiVad] : m=(1-o(1))∙k, |y|=log 3 n. Shpilka: 2-WOM, Enc and Dec can share o(n) bits. Given z,S,a, Enc finds y and shares it.

Linear seeded extractors => seeded bit- fixing disperser Known: Explicit constructions of E(x,y) s.t.  bit-fixing source * X S,a with k unfixed bits, for most y’s E(X S,a,y) is close to uniform.  y: T(x)=E(x,y) is linear over F 2 n. * In fact,  distribution with min-entropy ≥ k. X S,a is an affine subspace of F 2 n.  E is a seeded bit-fixing source disperser that is efficiently invertible. [Tre99,ShaUma01]: seed O(log n), output √k [RazReiVad00]: seed O(log 3 n), output (1-o(1))k

(Seedless) disperser construction: Step 1: (seedless) disperser with short output m=O(log n) that is efficiently invertible. Step 2: (Using approach of [GabSha08]) Combine with seeded disperser with large output length to get (seedless) and efficiently invertible disperser with large output length.

Thm: Any explicit bit-fixing extractor* D:{0,1} n  {0,1} m with m=O(log n) is efficiently invertible. *Say, Pr X  X S,a [D(X)=z] ¸ 2 -2m = 1/poly(n) Known Explicit constructions: [KempZuck03,GabRazSha04,Rao09]. Proof: Sample random x 2 X S,a and check if D(x) =z. Will take expected poly(n)-time. Step 1: Disperser with m=O(log n)

Step 2: Increasing output length with threshold k’ rather than k => F is efficiently invertible

Step 2: Increasing output length with threshold k’ rather than k affine subspace

Dispersers for affine sources Get: zero-error disperser for affine sources. Used in memory scheme w/ additional errs. Have explicit D only for k=n/(log log n) ½.  Final disperser has k > n/(log log n) ½.  Output length m = k - O(n/(log log n) ½ ). Suffices for achieving rate 1-p-o(1). In paper: Subsource-hitter for bit-fixing. => Final disperser for bit-fixing sources. Works for any k > log c n. Output length m = k - O(log c n). Gives schemes with m = n-p∙n - O(log c n).

Linear seeded extractors Subsource hitter for bit-fixing sources. Subsource hitter (Inspired by [GabRazSha] ) *0**1* X S,a *** seed y *0**1* X S,a *** seed y seed y’ y’ - seed for sampler, selects subset of [n].  y’ that partitions unfixed bits: Most unfixed bits are selected. At least k’ bits are not selected. Apply E(∙,y) only on selected portion of x.

Handling stuck-at memory with noise msg z Enc *0**1*010 Dec Noise Idea: Force Enc to output codeword in code for noise. Need code:  defect pattern  many consistent codewords. Linear code w/ dual dist p∙n. Code ∩ X S,a is affine space. dim n-p∙n-o(n) (for n ½ errs) Use affine source disp. k=n-p∙n-o(n) => m=n-p∙n-o(n). => rate 1-p-o(1). Given corrupted x’ that is close to x: Decode x’ to find x. Compute D(x) to get msg z.

Open questions

That’s it…

Standrad coding theory (unknown errors) msg z Enc 010 Dec Channel If p∙n bits are corrupted, rate can be* ≈ 1-h(p) Achieved by random codes. Not known to be tight. Explicit constructions achieve poorer rate. C(z)

Suppose Enc\Dec know where errors are: msg z Enc *0**1010 Dec 010 Can trivially get get rate 1-p, by using `unstuck’ places Channel replaces ≤p∙n bits by fixed bits.

What if only Enc knows errors? msg z Enc *0**1*010 Dec 010 Dec must retrieve z without knowing the error pattern Channel replaces ≤p∙n bits by fixed bits. [Kuznetsov-Tsybakov:] possible to get any rate R < 1-p. We give first explicit schemes. rate R > 1-p-o(1).

Dfn: A subsource of X S,a is a set X S’,a’ ½ X S,a for -S’ ¾ S with |S’| · 2p ¢ n. *0**1* X S,a : 00**1* X S’,a’ :

Dfn:E(x,y) is a subsource hitter if given z, there exists - a seed y - a subsource X S’,a’ of X S,a such that: E(X S’,a’,y) ´ z E is eff. invertible if y, X S’,a’ can be found in poly(n)-time.

Theorem of [GabSha08] -D is a b.f.d (for sets |S’| · 2p ¢ n). -E is a subsource hitter.  F(x), E(x,D(x)) is a bit-fixing disperser. Proof: -Given z, 9 X S’,a’ with E(X S’,a’,y) ´ z. -Also, 9 x 2 X S’,a’ with D(x)=y. For this x: E(x,D(x))=E(x,y)=z.

Efficient Invertibilty of F Proof: -Given z, 9 X S’,a’ with E(X S’,a’,y) ´ z. -Also, 9 x 2 X S’,a’ with D(x)=y. For this x: E(x,D(x))=E(x,y)=z. E e.i.  can efficiently find X S’,a’,y. D e.i.  can eff. find x 2 X S’,a’ with D(x)=y.

Seeded dispersers  subsource hitters: First, sample a random subset of the input bits. Apply seeded disperser only on this subset. *0**1* Preimage set now includes all settings of non-fixed bits outside of the subset. apply E here

Similar results for `mixed’ scenario ‘stuck errors’ known to encoder+ noise added after encoding uses affine dispersers and error correcting codes rather than bit-fixing dispersers. We get similar improvements for output length of affine dispersers.

Open questions -Invertible dispersers for `Hamming sources’: Find mapping D:{0,1} n  {0,1} m such that for all large enough hamming balls B D(B) = {0,1} m, and there is efficient way of finding preimage of z in B. Thanks

Dispersers imply memory schemes Bit-fixing disperser D:{0,1} n  {0,1} m. - Encoding: Given message z 2 {0,1} m and defect pattern S,a, Enc finds x 2 X S,a with D(x) = z. - Decoding: Given x, Dec computes D(x)=z. Output length m = (1-o(1))∙k => optimal rate. Explicitness of scheme follows if D is explicit and efficiently invertibile.

Fixed constant 0<p<1. Encoding function C:{0,1} m  {0,1} n,n>m. -Message z 2 {0,1} m -Codeword C(z) 2 {0,1} n Decoder must retrieve z from C(z) after p∙n bits of C(z) have been corrupted. (will look at different dfns next) m/n, the rate of C. Coding Theory

What if only Enc knows errors? msg z Enc *0**1*010 Dec 010 [Kuznetsov-Tsybakov:] possible to get any rate R < 1-p. We give first explicit schemes. `defect pattern’- p∙n `stuck’ bits

Memory with n cells. 0<p<1, some constant. p∙n cells are `stuck’ either at 0 or 1. Person writing in memory knows which cells are stuck and to what values. Person reading memory does not. Memory with stuck-at defects:

`Defect pattern’:Subset S µ [n] of size p∙n, String a 2 {0,1} p∙n. Message z 2 {0,1} m. -Encoding function E, given S,a,z, produces x 2 {0,1} n with x| S = a. -Decoding function D such that D(x) = z D doesn’t depend on S,a m/n is the rate of the scheme Formal dfn of memory scheme:

[Kuznetsov-Tsybakov] : Introduced model. Non- explicit schemes with any rate R<1-p [Tsybakov] : Explicit schemes using linear codes with rate <<1-p Our result: Explicit* schemes with rate 1-p- o(1). *-Our encoding is runs in expected polynomial time. Previous results and Ours

Bit-fixing dispersers: *0**1**** Given all strings with a certain `defect pattern’ a bit-fixing disperser produces all possible outputs D

Bit-fixing dispersers: *0**1**** Given all strings with a certain `defect pattern’ a bit-fixing disperser produces all possible outputs D Example: D(x 1,…,x n ),  i x i (mod 2)

Bit-fixing dispersers: S µ [n] – subset of size p∙n. a – string in {0,1} p∙n X S,a, set of strings in {0,1} n, with x| S =a. Call such a set a bit-fixing source. Dfn: D:{0,1} n  {0,1} m is a bit-fixing disperser if for every S,a as above D(X S,a )={0,1} m (this is actually a zero-error disperser)

Bit-fixing dispersers: *0**1**** Another example: Suppose k=n-p∙n. Take the sum of input bits (mod k).. gives us logk output bits. D

Dispersers to memory schemes: Bit-fixing disperser D(x). Message z -Given defect pattern S,a, Enc finds x 2 X S,a with D(x) = z. -Dec simply computes D. Efficiency of scheme requires `efficient invertibility’ of D.

Dfn: We say D is efficiently invertible if given S,a and output z, we can find in expected* poly(n)-time x 2 X S,a with D(x) = z. *Simply because that’s what we get

Our results: Bit-fixing dispersers with output length m=k-o(k), where k = n-p ¢ n. Previous result m=  (k). [GabSha08] Also, our dispersers are efficiently invertible – (previous constructions seem not to be)

Disperser Construction: Step 1: Disperser with short output. Step 2: Seeded disperser with large output length. Step 3: Combining the above to get (seedless) disperser with large output length. (Follows [GabSha08])

Thm: Any explicit bit-fixing extractor* D:{0,1} n  {0,1} m with m=O(log n) is efficiently invertible. *Say, Pr x  X S,a [D(x)=z] ¸ 2 -2m =1/poly(n) Known Explicit constructions: [KempZuck03,GabRazSha04,Rao09]. Proof: Sample random x 2 X S,a and check if D(x) =z. Will take expected poly(n)-time. Step 1: Disperser with m=O(log n)

2 nd step: we construct efficiently invertible seeded bit-fixing dispersers with large output length.

Seeded Bit-fixing dispersers: *0**1**** We `invest’ a short seed to obtain a longer output. E * auxiliary short random input a.k.a `seed’

Dfn:E(x,y) is a seeded bit-fixing disperser if given z, there exist: -a seed y, and -x 2 X S,a such that: E(x,y) = z. E is efficiently invertible if x,y can be found in poly(n)-time...idea is that |y|<<|z|.

Observation: If T(x)=E(x,y) is linear function for every fixed seed y, then E is efficiently invertible: For every y: (assume |y|=O(log n) ) -Check if exists x with -x 2 X S,a -E(x,y) =z (This is a set of affine equations and can be check efficiently)

`Linear seeded extractors’ exactly give us such an E with large output length Such constructions by [Trevisan\RazReingoldVadhan\ShaltielUmans,..]

Step 3: Get rid of seed: -D(x) is a b.f.d. with output length d -E(x,y) is a seeded b.f.d. with seed length d  [GabSha08] : F(x) = E(x,D(x)) is a bit-fixing disperser assuming E is also a `subsource hitter’. Also: If D and E are efficiently invertible so is F.

Subsource hitters [GabSha08] `a seeded bit-fixing disperser where the set of preimages of z contains a bit-fixing source’

Dfn: A subsource of X S,a is a set X S’,a’ ½ X S,a for -S’ ¾ S with |S’| · 2p ¢ n. *0**1* X S,a : 00**1* X S’,a’ :

Dfn:E(x,y) is a subsource hitter if given z, there exists - a seed y - a subsource X S’,a’ of X S,a such that: E(X S’,a’,y) ´ z E is eff. invertible if y, X S’,a’ can be found in poly(n)-time.

Theorem of [GabSha08] -D is a b.f.d (for sets |S’| · 2p ¢ n). -E is a subsource hitter.  F(x), E(x,D(x)) is a bit-fixing disperser. Proof: -Given z, 9 X S’,a’ with E(X S’,a’,y) ´ z. -Also, 9 x 2 X S’,a’ with D(x)=y. For this x: E(x,D(x))=E(x,y)=z.

Efficient Invertibilty of F Proof: -Given z, 9 X S’,a’ with E(X S’,a’,y) ´ z. -Also, 9 x 2 X S’,a’ with D(x)=y. For this x: E(x,D(x))=E(x,y)=z. E e.i.  can efficiently find X S’,a’,y. D e.i.  can eff. find x 2 X S’,a’ with D(x)=y.

Seeded dispersers  subsource hitters: First, sample a random subset of the input bits. Apply seeded disperser only on this subset. *0**1* Preimage set now includes all settings of non-fixed bits outside of the subset. apply E here

Similar results for `mixed’ scenario ‘stuck errors’ known to encoder+ noise added after encoding uses affine dispersers and error correcting codes rather than bit-fixing dispersers. We get similar improvements for output length of affine dispersers.

Open questions -Invertible dispersers for `Hamming sources’: Find mapping D:{0,1} n  {0,1} m such that for all large enough hamming balls B D(B) = {0,1} m, and there is efficient way of finding preimage of z in B. Thanks

Pseudorandomness Course Dispersers are related to Extractors that have applications in TCS, e.g.: -Running randomized algorithms with weak randomness -Generating pseudorandom bits for small space algorithms.

Increasing the output length of dispersers [GabSha08] D – a bit-fixing disperser with output length d E – a subsource hitter with seed length d, output length m  F(x) = E(x,D(x)) is a bit-fixing disperser with output length m.

Efficient invertibility F(x) = E(x,D(x)) D has output length O(log n) Finding preimage of z: -choose random y – find subspace X’ of X S,a with E(X’,y) ´ z (system of affine equations) Find x 2 X’ with D(x) = y. We get F(x) = E(x,D(x)) = E(x,y) = z. (Last equality because x 2 X’)

Efficient invertibility When D has output length O(logn) random sampling will succeed w.h.p in poly(n)-time.

Dispersers to memory schemes: Message z Encoder finds x 2 X S,a with D(x) = z *0**1*010 Decoder computes D(x) to retrieve z 010 must always decode correctly without knowing the defect pattern `defect pattern’- p∙n `stuck’ bits

Standard coding: Message z Encoder 010 Decoder Noise Codeword C(z)