1 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 / RAC 2211 Lecture.

Slides:



Advertisements
Similar presentations
Quantum t-designs: t-wise independence in the quantum world Andris Ambainis, Joseph Emerson IQC, University of Waterloo.
Advertisements

Short seed extractors against quantum storage Amnon Ta-Shma Tel-Aviv University 1.
B ETTI NUMBERS OF RANDOM SIMPLICIAL COMPLEXES MATTHEW KAHLE & ELIZABETH MECKE Presented by Ariel Szapiro.
Foundations of Cryptography Lecture 10 Lecturer: Moni Naor.
Quantum Computing MAS 725 Hartmut Klauck NTU
Information Theory EE322 Al-Sanie.
Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Lecture 11 (2005) Richard Cleve DC 3524
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Entropy in the Quantum World Panagiotis Aleiferis EECS 598, Fall 2001.
Short course on quantum computing Andris Ambainis University of Latvia.
Derandomized DP  Thus far, the DP-test was over sets of size k  For instance, the Z-Test required three random sets: a set of size k, a set of size k-k’
DENSITY MATRICES, traces, Operators and Measurements
Superdense coding. How much classical information in n qubits? Observe that 2 n  1 complex numbers apparently needed to describe an arbitrary n -qubit.
Evaluating Hypotheses
Introduction to Quantum Information Processing Lecture 4 Michele Mosca.
1 Recap (I) n -qubit quantum state: 2 n -dimensional unit vector Unitary op: 2 n  2 n linear operation U such that U † U = I (where U † denotes the conjugate.
Stats & Linear Models.
Introduction to AEP In information theory, the asymptotic equipartition property (AEP) is the analog of the law of large numbers. This law states that.
1 Introduction to Quantum Information Processing QIC 710 / CS 768 / PH 767 / CO 681 / AM 871 Richard Cleve QNC 3129 Lecture 18 (2014)
1 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lecture 14 (2009)
1 Introduction to Quantum Information Processing QIC 710 / CS 678 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 / QNC 3129 Lectures.
1 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lecture 16 (2011)
A Few Simple Applications to Cryptography Louis Salvail BRICS, Aarhus University.
Functions of Random Variables. Methods for determining the distribution of functions of Random Variables 1.Distribution function method 2.Moment generating.
1 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lecture 19 (2009)
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 653 Course.
Channel Capacity.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
1 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lecture 20 (2009)
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
8.4.2 Quantum process tomography 8.5 Limitations of the quantum operations formalism 量子輪講 2003 年 10 月 16 日 担当:徳本 晋
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 653 Lecture.
Quantum Two 1. 2 Angular Momentum and Rotations 3.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Image Processing Architecture, © Oleh TretiakPage 1Lecture 4 ECE-C490 Winter 2004 Image Processing Architecture Lecture 4, 1/20/2004 Principles.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Fidelity of a Quantum ARQ Protocol Alexei Ashikhmin Bell Labs  Classical Automatic Repeat Request (ARQ) Protocol  Quantum Automatic Repeat Request (ARQ)
Basic Concepts of Information Theory A measure of uncertainty. Entropy. 1.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Richard Cleve DC 2117 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Lecture (2011)
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 2117 Lecture.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
1 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lectures
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Richard Cleve DC 3524 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Lecture.
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Lecture.
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 16 (2009) Richard.
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 15 (2009) Richard.
Introduction to Quantum Computing Lecture 1 of 2
Linear sketching with parities
The Curve Merger (Dvir & Widgerson, 2008)
Quantum Information Theory Introduction
Richard Cleve DC 2117 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Lecture 15 (2011)
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 24 (2009) Richard.
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 18 (2009) Richard.
6.3 Sampling Distributions
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 16 (2009) Richard.
Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Lecture 17 (2005) Richard Cleve DC 3524
Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Lecture 4 (2005) Richard Cleve DC 653
A quantum machine learning algorithm based on generative models
Richard Cleve DC 3524 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Lecture.
Presentation transcript:

1 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 / RAC 2211 Lecture 24 (2012)

2 Entropy and compression

Shannon Entropy 3 Let p = ( p 1,…, p d ) be a probability distribution on a set {1,…,d} Then the (Shannon) entropy of p is H( p 1,…, p d ) Intuitively, this turns out to be a good measure of “how random” the distribution p is: vs. H(p) = log dH(p) = 0 Operationally, H(p) is the number of bits needed to store the outcome (in a sense that will be made formal shortly)

Von Neumann Entropy 4 For a density matrix , it turns out that S(  ) = − Tr  log  is a good quantum analogue of entropy Note: S(  ) = H( p 1,…, p d ), where p 1,…, p d are the eigenvalues of  (with multiplicity) Operationally, S(  ) is the number of qubits needed to store  (in a sense that will be made formal later on) Both the classical and quantum compression results pertain to the case of large blocks of n independent instances of data: probability distribution p  n in the classical case, and quantum state   n in the quantum case

Classical compression (1) 5 Let p = ( p 1,…, p d ) be a probability distribution on a set {1,…,d} where n independent instances are sampled: ( j 1,…, j n )  {1,…,d} n ( d n possibilities, n log d bits to specify one) Theorem*: for all  > 0, for sufficiently large n, there is a scheme that compresses the specification to n (H(p) +  ) bits while introducing an error with probability at most  A nice way to prove the theorem, is based on two cleverly defined random variables … Intuitively, there is a subset of {1,…,d} n, called the “typical sequences”, that has size 2 n (H(p) +  ) and probability 1 −  * “Plain vanilla” version that ignores, for example, the tradeoffs between n and 

Classical compression (2) 6 Define the random variable f :{1,…,d}  R as f ( j ) = − log p j Note that Also, g ( j 1,…, j n ) Define g :{1,…,d} n  R as Thus

Classical compression (3) 7 By standard results in statistics, as n  , the observed value of g ( j 1,…, j n ) approaches its expected value, H(p) More formally, call ( j 1,…, j n )  {1,…,d} n  - typical if Then, the result is that, for all  > 0, for sufficiently large n, Pr [ ( j 1,…, j n ) is  - typical ]  1 −  We can also bound the number of these  - typical sequences: By definition, each such sequence has probability  2 − n (H(p) +  ) Therefore, there can be at most 2 n (H(p) +  ) such sequences

Classical compression (4) 8 In summary, the compression procedure is as follows: The input data is ( j 1,…, j n )  {1,…,d} n, each independently sampled according the probability distribution p = ( p 1,…, p d ) The compression procedure is to leave ( j 1,…, j n ) intact if it is  - typical and otherwise change it to some fixed  - typical sequence, say, ( j,…, j) (which will result in an error) Since there are at most 2 n (H(p) +  )  - typical sequences, the data can then be converted into n (H(p) +  ) bits The error probability is at most , the probability of an atypical input arising

Quantum compression (1) 9 The scenario: n independent instances of a d -dimensional state are randomly generated according some distribution:  φ 1  prob. p 1     φ r  prob. p r Goal: to “compress” this into as few qubits as possible so that the original state can be reconstructed with small error in the following sense …  -good: No procedure can distinguish between these two states (a)compressing and then uncompressing the data (b)the original data left as is with probability more than ½ + ¼   0  prob. ½  +  prob. ½ Example:

Quantum compression (2) 10 Express  in its eigenbasis as Define Theorem: for all  > 0, for sufficiently large n, there is a scheme that compresses the data to n (S(  ) +  ) qubits, that is 2 √  -good For the aforementioned example,  0.6 n qubits suffices With respect to this basis, we will define an  - typical subspace of dimension 2 n (S(  ) +  ) = 2 n (H(q) +  ) The compression method:

Quantum compression (3) 11 The  - typical subspace is that spanned by where ( j 1,…, j n ) is  - typical with respect to ( q 1,…, q d ) By the same argument as in the classical case, the subspace has dimension  2 n (S(  ) +  ) and Tr(  typ   n )  1 −  Define  typ as the projector into the  - typical subspace  1  prob. q 1     d  prob. q d This is because  is the density matrix of

Quantum compression (4) 12 Calculation of the expected fidelity: Using ||  −  || tr  √1 − F( ,  ) 2, we can upper bound the expected trace distance by √ 2  Abbreviations: p i 1  i n = p i 1  p i n  φ i 1  i n  =  φ i 1    φ i n 