Richard Cleve DC 2117 cleve@cs.uwaterloo.ca Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 16 (2009) Richard.

Slides:



Advertisements
Similar presentations
Quantum t-designs: t-wise independence in the quantum world Andris Ambainis, Joseph Emerson IQC, University of Waterloo.
Advertisements

Quantum Computing MAS 725 Hartmut Klauck NTU
Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Lecture 11 (2005) Richard Cleve DC 3524
Entropy in the Quantum World Panagiotis Aleiferis EECS 598, Fall 2001.
1 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 / RAC 2211 Lecture.
DENSITY MATRICES, traces, Operators and Measurements
Superdense coding. How much classical information in n qubits? Observe that 2 n  1 complex numbers apparently needed to describe an arbitrary n -qubit.
Introduction to Quantum Information Processing Lecture 4 Michele Mosca.
Visual Recognition Tutorial
1 Recap (I) n -qubit quantum state: 2 n -dimensional unit vector Unitary op: 2 n  2 n linear operation U such that U † U = I (where U † denotes the conjugate.
Stats & Linear Models.
Introduction to AEP In information theory, the asymptotic equipartition property (AEP) is the analog of the law of large numbers. This law states that.
1 Introduction to Quantum Information Processing QIC 710 / CS 768 / PH 767 / CO 681 / AM 871 Richard Cleve QNC 3129 Lecture 18 (2014)
1 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lecture 14 (2009)
1 Introduction to Quantum Information Processing QIC 710 / CS 678 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 / QNC 3129 Lectures.
1 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lecture 16 (2011)
A Few Simple Applications to Cryptography Louis Salvail BRICS, Aarhus University.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 2117 Lecture.
1 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lecture 19 (2009)
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Channel Capacity.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 2117 Lecture.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
1 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lecture 20 (2009)
8.4.2 Quantum process tomography 8.5 Limitations of the quantum operations formalism 量子輪講 2003 年 10 月 16 日 担当:徳本 晋
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 653 Lecture.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Fidelity of a Quantum ARQ Protocol Alexei Ashikhmin Bell Labs  Classical Automatic Repeat Request (ARQ) Protocol  Quantum Automatic Repeat Request (ARQ)
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
Richard Cleve DC 2117 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Lecture (2011)
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 2117 Lecture.
1 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lectures
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Quantum Two 1. 2 Angular Momentum and Rotations 3.
Richard Cleve DC 3524 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Lecture.
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Lecture.
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 16 (2009) Richard.
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 15 (2009) Richard.
Postulates of Quantum Mechanics
Markov Chains Mixing Times Lecture 5
General Strong Polarization
Quantum Two.
Shadow Tomography of Quantum States
Quantum Foundations Lecture 15
Quantum One.
Hidden Markov Models Part 2: Algorithms
CONCEPTS OF ESTIMATION
Lecture on Linear Algebra
The Curve Merger (Dvir & Widgerson, 2008)
Quantum Information Theory Introduction
Quantum One.
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 22 (2009) Richard.
Summarizing Data by Statistics
Welcome to the wonderful world of Probability
Richard Cleve DC 2117 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Lecture 15 (2011)
Generally Discriminant Analysis
Symmetric Matrices and Quadratic Forms
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 24 (2009) Richard.
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 18 (2009) Richard.
Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Lecture 17 (2005) Richard Cleve DC 3524
Vector Spaces RANK © 2012 Pearson Education, Inc..
Symmetric Matrices and Quadratic Forms
Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Lecture 4 (2005) Richard Cleve DC 653
A quantum machine learning algorithm based on generative models
Richard Cleve DC 3524 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 667 / Phys 767 C&O 481 / C&O 681 Lecture.
Presentation transcript:

Richard Cleve DC 2117 cleve@cs.uwaterloo.ca Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 16 (2009) Richard Cleve DC 2117 cleve@cs.uwaterloo.ca

Distinguishing between two arbitrary quantum states

Holevo-Helstrom Theorem (1) Theorem: for any two quantum states  and , the optimal measurement procedure for distinguishing between them succeeds with probability ½ + ¼ || −  ||tr (equal prior probs.) Proof* (the attainability part): Since  −  is Hermitian, its eigenvalues are real Let + be the projector onto the positive eigenspaces Let − be the projector onto the non-positive eigenspaces Take the POVM measurement specified by + and − with the associations +   and −   * The other direction of the theorem (optimality) is omitted here

Holevo-Helstrom Theorem (2) Claim: this succeeds with probability ½ + ¼ || −  ||tr Proof of Claim: A key observation is Tr(+ − −) ( −  ) = || −  ||tr The success probability is ps = ½Tr(+  ) + ½Tr(− ) & the failure probability is pf = ½Tr(+  ) + ½Tr(− ) Therefore, ps − pf = ½Tr(+ − −) ( −  ) = ½|| −  ||tr From this, the result follows

Purifications & Ulhmann’s Theorem Any density matrix , can be obtained by tracing out part of some larger pure state: a purification of  Ulhmann’s Theorem*: The fidelity between  and  is the maximum of φψ taken over all purifications ψ and φ * See [Nielsen & Chuang, pp. 410-411] for a proof of this Recall our previous definition of fidelity as F(,  ) = Tr√ 1/2  1/2  ||1/2  1/2||tr

Relationships between fidelity and trace distance 1 − F(, )  || −  ||tr  √1 − F(, )2 See [Nielsen & Chuang, pp. 415-416] for more details

Entropy and compression

Shannon Entropy Let p = (p1,…, pd) be a probability distribution on a set {1,…,d} Then the (Shannon) entropy of p is H(p1,…, pd) Intuitively, this turns out to be a good measure of “how random” the distribution p is: vs. vs. vs. H(p) = log d H(p) = 0 Operationally, H(p) is the number of bits needed to store the outcome (in a sense that will be made formal shortly)

Von Neumann Entropy For a density matrix , it turns out that S() = − Tr log is a good quantum analog of entropy Note: S() = H(p1,…, pd), where p1,…, pd are the eigenvalues of  (with multiplicity) Operationally, S() is the number of qubits needed to store  (in a sense that will be made formal later on) Both the classical and quantum compression results pertain to the case of large blocks of n independent instances of data: probability distribution pn in the classical case, and quantum state  n in the quantum case

Classical compression (1) Let p = (p1,…, pd) be a probability distribution on a set {1,…,d} where n independent instances are sampled: ( j1,…, jn) {1,…,d}n (d n possibilities, n log d bits to specify one) Theorem*: for all  > 0, for sufficiently large n, there is a scheme that compresses the specification to n(H(p) + ) bits while introducing an error with probability at most  Intuitively, there is a subset of {1,…,d}n, called the “typical sequences”, that has size 2n(H(p) + ) and probability 1 −  A nice way to prove the theorem, is based on two cleverly defined random variables … * “Plain vanilla” version that ignores, for example, the tradeoffs between n and 

Classical compression (2) Define the random variable f :{1,…,d}  R as f ( j ) = − log pj Note that Define g :{1,…,d}n  R as Thus Also, g ( j1,…, jn)

Classical compression (3) By standard results in statistics, as n  , the observed value of g ( j1,…, jn) approaches its expected value, H(p) More formally, call ( j1,…, jn) {1,…,d}n -typical if Then, the result is that, for all  > 0, for sufficiently large n, Pr[( j1,…, jn) is -typical]  1−  We can also bound the number of these -typical sequences: By definition, each such sequence has probability  2−n(H(p) + ) Therefore, there can be at most 2n(H(p) + ) such sequences

Classical compression (4) In summary, the compression procedure is as follows: The input data is ( j1,…, jn) {1,…,d}n, each independently sampled according the probability distribution p = (p1,…, pd) The compression procedure is to leave ( j1,…, jn) intact if it is -typical and otherwise change it to some fixed -typical sequence, say, ( j ,…, j) (which will result in an error) Since there are at most 2n(H(p) + ) -typical sequences, the data can then be converted into n(H(p) + ) bits The error probability is at most , the probability of an atypical input arising

Quantum compression (1) The scenario: n independent instances of a d -dimensional state are randomly generated according some distribution: φ1  prob. p1    φr  prob. pr 0 prob. ½ + prob. ½ Example: Goal: to “compress” this into as few qubits as possible so that the original state can be reconstructed with small error in the following sense … The expected* trace distance between the reconstructed state and the state that was actually generated should be small * Defined as the expected value of the trace distance, taken with respect to the randomness of the generation procedure

Quantum compression (2) Define Theorem: for all  > 0, for sufficiently large n, there is a scheme that compresses the data to n(S() + ) qubits, with expected trace distance  √2 For the aforementioned example,  0.6n qubits suffices The compression method: Express  in its eigenbasis as With respect to this basis, we will define an -typical subspace of dimension 2n(S() + ) = 2n(H(q) + )

Quantum compression (3) The -typical subspace is that spanned by where ( j1,…, jn) is -typical with respect to (q1,…, qd) Define typ as the projector into the -typical subspace By the same argument as in the classical case, the subspace has dimension  2n(S() + ) and Tr(typ  n)  1−  This is because  is the density matrix of 1 prob. q1    d prob. qd

Quantum compression (4) Calculation of the expected fidelity: Abbreviations: pi1in = pi1 pin φi1in = φi1φin Using || −  ||tr  √1 − F(, )2, we can upper bound the expected trace distance by √2