Conditional Computational Entropy Does Pseudo-Entropy = Incompressibility? How to extract more pseudorandom bits? Chun-Yuan Hsiao (Boston University, USA) Joint work with Chi-Jen Lu (Academia Sinica, Taiwan) Leonid Reyzin (Boston University, USA)
Shannon Entropy H(X) Exx [ log ( Pr[X x] ) ] X 2.58 bits Usually in crypto: minimum instead of average (a.k.a. min-entropy H(X) )
Computational Entropy Pseudo-Entropy X has pseudo-entropy k if Y, H(Y) = k and X Y HHILL(X) = k [Håstad,Impagliazzo,Levin,Luby] X means indistinguishable (in polynomial time) PRG (Blum-Micali-Yao) Computational Entropy (version 1: HILL)
Entropy vs Compressibility Shannon's Theorem | X | = 60 H(X) = 40 H(X) X C(X) D(C(X)) = X Compression length C(X) Compress ( C ) Decompress ( D )
Compression-Entropy HYao(X) = k [Yao82] Computational Entropy (version 2: Yao) X has computational entropy k, if we cannot efficiently compress X shorter than k HYao(X) = k [Yao82] [Barak,Shaltiel,Wigderson03] gave min-entropy formulation any subset of the support of X cannot be compressed
Computational Entropy Version 1: HILL HHILL(X) = k, if Y, H(Y) = k and X Y Version 2: Yao HYao(X) = k, if we cannot efficiently compress X shorter than k Question [Impagliazzo99]: Are these equivalent definitions? ? ?
(Pseudo-)Entropy vs Compressibility Recall Shannon’s Theorem: Is computational analogue true? ? pseudo- entropy compression length efficient
Computational Entropy Version 1: HILL HHILL(X) = k, if Y, H(Y) = k and X Y Version 2: Yao HYao(X) = k, if we cannot efficiently compress X shorter than k ?
Cryptographic Motivation pseudo H(X) random bits computational Extractor (Hashing) entropy key Which computational entropy? all extractors work for HHILL(X); some work for HYao(X) [BSW03] e.g. gab If HYao(X) > HHILL(X) may get longer a key (by using the right extractor)
Our results How? 0. New† notion: conditional computational entropy †previously used, but never formalized 1. distribution* X such that HYao(X) > HHILL(X) 2. bits extracted via HYao > bits extracted via HHILL 3. Define computational entropy, version 3: new, unpredictability-based definition *conditional distribution
Our Definition: Conditional Computational Entropy HILL: HHILL(X | Z) = k if Y, H(Y | Z) = k and (X , Z) (Y , Z) Z X Y ?
Our Definition: Conditional Computational Entropy Yao: HYao(X | Z) = k if we cannot efficiently compress X shorter than k Z Z D(C(X , Z) ,Z) =X C( X , Z )
Conditional is Everywhere in Crypto In cryptography, adversaries usually have additional information entropic secret: gab | adversary is given ga, gb entropic secret: x | adversary is given f(x) entropic secret: SignSK(m) | adversary is given PK To make extraction precise, must talk about conditional entropy Conditional computational entropy has been used implicitly in [Gennaro,Krawczyk,Rabin04], but never defined explicitly for HILL and Yao
Our results 0. New† notion: conditional computational entropy †previously used, but never formalized 1. pair (X, Z) such that HYao(X | Z) >> HHILL(X | Z) (where Z is a uniform string) 2. Extract more pseudorandom bits from (X , Z) by considering its Yao-entropy 3. Define computational entropy, version 3: Hunp(X | Z) = k, if efficient M, Pr[ M(Z) = X ] < 2k Allows to talk about entropy of singletons, like x | f(x) Can’t be defined unconditionally
Yao Entropy > HILL Entropy [Wee03] (oracle separation) [this paper] Length increasing random function f PRG G {0,1}n {0,1}3n X Caveat: need uniZK [Lepinski,Micali,Shelat05] X = ( G( Un ) , ) Z = NIZK reference string Non- Interactive Zero- Knowledge Membership oracle m Yes No
Summary Computational Entropy: Conditional Version 1: HHILL (X | Z) Conditional Version 2: HYao (X | Z) Conditional Version 3: Hunp (X | Z) Computational Entropy: Can extract more from Yao than HILL (even unconditionally)
Thank You!