Quantum Information Theory Patrick Hayden (McGill) 4 August 2005, Canadian Quantum Information Summer School.

Slides:



Advertisements
Similar presentations
Quantum Computing MAS 725 Hartmut Klauck NTU
Advertisements

I NFORMATION CAUSALITY AND ITS TESTS FOR QUANTUM COMMUNICATIONS I- Ching Yu Host : Prof. Chi-Yee Cheung Collaborators: Prof. Feng-Li Lin (NTNU) Prof. Li-Yi.
Information Theory EE322 Al-Sanie.
Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Lecture 11 (2005) Richard Cleve DC 3524
Quantum Information Stephen M. Barnett University of Strathclyde The Wolfson Foundation.
Erasing correlations, destroying entanglement and other new challenges for quantum information theory Aram Harrow, Bristol Peter Shor, MIT quant-ph/
Entropy in the Quantum World Panagiotis Aleiferis EECS 598, Fall 2001.
Michael A. Nielsen University of Queensland Quantum entropy Goals: 1.To define entropy, both classical and quantum. 2.To explain data compression, and.
1 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 / RAC 2211 Lecture.
Quantum information as high-dimensional geometry Patrick Hayden McGill University Perspectives in High Dimensions, Cleveland, August 2010.
Chain Rules for Entropy
Quantum Computing MAS 725 Hartmut Klauck NTU
Chapter 6 Information Theory
Classical capacities of bidirectional channels Charles Bennett, IBM Aram Harrow, MIT/IBM, Debbie Leung, MSRI/IBM John Smolin,
DENSITY MATRICES, traces, Operators and Measurements
Quantum communication from Alice to Bob Andreas Winter, Bristol quant-ph/ Aram Harrow, MIT Igor Devetak, USC.
Fundamental limits in Information Theory Chapter 10 :
Superdense coding. How much classical information in n qubits? Observe that 2 n  1 complex numbers apparently needed to describe an arbitrary n -qubit.
Avraham Ben-Aroya (Tel Aviv University) Oded Regev (Tel Aviv University) Ronald de Wolf (CWI, Amsterdam) A Hypercontractive Inequality for Matrix-Valued.
A Family of Quantum Protocols Igor Devetak, IBM Aram Harrow, MIT Andreas Winter, Bristol quant-ph/ IEEE Symposium on Information Theory June 28,
Introduction to Quantum Information Processing Lecture 4 Michele Mosca.
Coherent Classical Communication Aram Harrow (MIT) quant-ph/
CSI Uncertainty in A.I. Lecture 201 Basic Information Theory Review Measuring the uncertainty of an event Measuring the uncertainty in a probability.
Erasing correlations, destroying entanglement and other new challenges for quantum information theory Aram Harrow, Bristol Peter Shor, MIT quant-ph/
EECS 598 Fall ’01 Quantum Cryptography Presentation By George Mathew.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 13 June 22, 2005
Noise, Information Theory, and Entropy
Quantum Shannon Theory Patrick Hayden (McGill) 17 July 2005, Q-Logic Meets Q-Info.
Introduction to Quantum Shannon Theory Patrick Hayden (McGill University) 12 February 2007, BIRS Quantum Structures Workshop | 
Noise, Information Theory, and Entropy
1 Statistical NLP: Lecture 5 Mathematical Foundations II: Information Theory.
Some basic concepts of Information Theory and Entropy
§1 Entropy and mutual information
1 Introduction to Quantum Information Processing QIC 710 / CS 768 / PH 767 / CO 681 / AM 871 Richard Cleve QNC 3129 Lecture 18 (2014)
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
1 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lecture 14 (2009)
1 Introduction to Quantum Information Processing QIC 710 / CS 667 / PH 767 / CO 681 / AM 871 Richard Cleve DC 2117 Lecture 16 (2011)
ENTROPIC CHARACTERISTICS OF QUANTUM CHANNELS AND THE ADDITIVITY PROBLEM A. S. Holevo Steklov Mathematical Institute, Moscow.
A Few Simple Applications to Cryptography Louis Salvail BRICS, Aarhus University.
The Operational Meaning of Min- and Max-Entropy
§4 Continuous source and Gaussian channel
Channel Capacity.
COMMUNICATION NETWORK. NOISE CHARACTERISTICS OF A CHANNEL 1.
JHU CS /Jan Hajic 1 Introduction to Natural Language Processing ( ) Essential Information Theory I AI-lab
The private capacities of a secret shared reference frame Patrick Hayden (McGill) with: PRA 75: (2005) ??? Stephen Bartlett Robert Spekkens arXiv:quant-ph/
The Operational Meaning of Min- and Max-Entropy Christian Schaffner – CWI Amsterdam, NL joint work with Robert König – Caltech Renato Renner – ETH Zürich,
Counterexamples to the maximal p -norm multiplicativity conjecture Patrick Hayden (McGill University) || | | N(½)N(½) p C&QIC, Santa Fe 2008.
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
Coherent Communication of Classical Messages Aram Harrow (MIT) quant-ph/
Quantum correlations with no causal order OgnyanOreshkov, Fabio Costa, ČaslavBrukner Bhubaneswar arXiv: December2011 Conference on Quantum.
Quantum correlations with no causal order OgnyanOreshkov, Fabio Costa, ČaslavBrukner Bhubaneswar arXiv: December2011 Conference on Quantum.
The Classically Enhanced Father Protocol
Nawaf M Albadia
Detecting Incapacity Graeme Smith IBM Research Joint Work with John Smolin QECC 2011 December 6, 2011 TexPoint fonts used in EMF. Read the TexPoint manual.
Optimal Trading of Classical Communication, Quantum Communication, and Entanglement Mark M. Wilde arXiv: ERATO-SORST Min-Hsiu Hsieh 4 th Workshop.
A generalization of quantum Stein’s Lemma Fernando G.S.L. Brandão and Martin B. Plenio Tohoku University, 13/09/2008.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
1 Lecture 7 System Models Attributes of a man-made system. Concerns in the design of a distributed system Communication channels Entropy and mutual information.
Coherent Classical Communication Aram Harrow, MIT Quantum Computing Graduate Research Fellow Objective Objective ApproachStatus Determine.
1 Introduction to Quantum Information Processing CS 467 / CS 667 Phys 467 / Phys 767 C&O 481 / C&O 681 Richard Cleve DC 3524 Course.
Coherent Communication of Classical Messages Aram Harrow (MIT) quant-ph/
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
(C) 2000, The University of Michigan 1 Language and Information Handout #2 September 21, 2000.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 16 (2009) Richard.
Introduction to Information theory
COT 5611 Operating Systems Design Principles Spring 2012
COT 5611 Operating Systems Design Principles Spring 2014
Quantum Information Theory Introduction
Richard Cleve DC 2117 Introduction to Quantum Information Processing CS 667 / PH 767 / CO 681 / AM 871 Lecture 16 (2009) Richard.
Presentation transcript:

Quantum Information Theory Patrick Hayden (McGill) 4 August 2005, Canadian Quantum Information Summer School

Overview  Part I:  What is information theory?  Entropy, compression, noisy coding and beyond  What does it have to do with quantum mechanics?  Noise in the quantum mechanical formalism  Density operators, the partial trace, quantum operations  Some quantum information theory highlights  Part II:  Resource inequalities  A skeleton key

Information (Shannon) theory  A practical question:  How to best make use of a given communications resource?  A mathematico-epistemological question:  How to quantify uncertainty and information?  Shannon:  Solved the first by considering the second.  A mathematical theory of communication [1948] The

Quantifying uncertainty  Entropy: H(X) = -  x p(x) log 2 p(x)  Proportional to entropy of statistical physics  Term suggested by von Neumann (more on him later)  Can arrive at definition axiomatically:  H(X,Y) = H(X) + H(Y) for independent X, Y, etc.  Operational point of view…

X1X1 X 2 … XnXn Compression Source of independent copies of X {0,1} n : 2 n possible strings 2 nH(X) typical strings If X is binary: About nP(X=0) 0’s and nP(X=1) 1’s Can compress n copies of X to a binary string of length ~nH(X)

Typicality in more detail  Let x n = x 1,x 2,…,x n with x j 2 X  We say that x n is  -typical with respect to p(x) if  For all a 2 X with p(a)>0, |1/n N(a|x n ) – p(a) | <  / | X |  For all a 2 X with p(a) = 0, N(a|x n )=0.  For  >0, the probability that a random string X n is  -typical goes to 1.  If x n is  -typical, 2 -n[H(X)+  ] · p(x n ) · 2 -n[H(X)-  ]  The number of  -typical strings is bounded above by 2 n[H(X)+ 

H(Y) Quantifying information H(X) H(Y|X) Information is that which reduces uncertainty I(X;Y) H(X|Y) Uncertainty in X when value of Y is known H(X|Y)= H(X,Y)-H(Y) = E Y H(X|Y=y) I(X;Y) = H(X) – H(X|Y) = H(X)+H(Y)-H(X,Y) H(X,Y)

Sending information through noisy channels Statistical model of a noisy channel: ´ m Encoding Decoding m’ Shannon’s noisy coding theorem: In the limit of many uses, the optimal rate at which Alice can send messages reliably to Bob through  is given by the formula

Data processing inequality Alice Bob time I(X;Y) ¸ I(Z;Y) I(X;Y) X Y p(z|x) Z Y I(Z;Y)

Optimality in Shannon’s theorem m Encoding Decoding m’ Shannon’s noisy coding theorem: In the limit of many uses, the optimal rate at which Alice can send messages reliably to Bob through  is given by the formula Assume there exists a code with rate R and perfect decoding. Let M be the random variable corresponding to the uniform distribution over messages. XnXn YnYn nR = H(M) = I(M;M’) · I(M;Y n ) · I(X n ;Y n ) ·  j=1 n I(X j,Y j ) · n ¢ max p(x) I(X;Y) M has nR bits of entropy Perfect decoding: M = M ’ Data processing Some fiddling Term by term

Shannon theory provides  Practically speaking:  A holy grail for error-correcting codes  Conceptually speaking:  A operationally-motivated way of thinking about correlations  What’s missing (for a quantum mechanic)?  Features from linear structure: Entanglement and non-orthogonality

Quantum Shannon Theory provides  General theory of interconvertibility between different types of communications resources: qubits, cbits, ebits, cobits, sbits…  Relies on a  Major simplifying assumption: Computation is free  Minor simplifying assumption: Noise and data have regular structure

Before we get going: Some unavoidable formalism  We need quantum generalizations of:  Probability distributions (density operators)  Marginal distributions (partial trace)  Noisy channels (quantum operations)

Mixing quantum states: The density operator Draw |  x i with probability p(x) Perform a measurement {|0 i,|1 i }: Probability of outcome j: q j =  x p(x) | h j|  x i | 2 =  x p(x) tr[|j ih j|  x ih  x |] = tr[ |j ih j|  ], where Outcome probability is linear in 

Properties of the density operator   is Hermitian:   y = [  x p(x) |  x ih  x |] y =  x p(x) [|  x ih  x |] y =    is positive semidefinite:  h  |  |  i =  x p(x) h  |  x ih  x |  i¸ 0  tr[  ] = 1:  tr[  ] =  x p(x) tr[|  x ih  x |] =  x p(x) = 1  Ensemble ambiguity:  I/2 = ½[|0ih 0| + |1ih 1|] = ½[|+ih+| + |-ih-|]

The density operator: examples Which of the following are density operators?

  The partial trace  Suppose that  AB is a density operator on A­B  Alice measures {M k } on A  Outcome probability is q k = tr[ (M k ­ I B )  AB ]  Define  A = tr B [  AB ] =  j B hj|  AB |ji B.  Then q k = tr[ M k  A ]   A describes outcome statistics for all possible experiments by Alice alone {M k } 

 i Purification  Suppose that  A is a density operator on A  Diagonalize  A =  i i |  i ih  i |  Let |  i =  i i 1/2 |  i i A |ii B  Note that  A = tr B [  ]  |  i is a purification of   Symmetry:   =   and   have same non-zero eigenvalues 

Quantum (noisy) channels: Analogs of p(y|x) What reasonable constraints might such a channel  :A ! B satisfy? 1)Take density operators to density operators 2)Convex linearity: a mixture of input states should be mapped to a corresponding mixture of output states All such maps can, in principle, be realized physically Must be interpreted very strictly Require that (  ­ I C )(  AC ) always be a density operator too Doesn’t come for free! Let T be the transpose map on A. If |  i = |00 i AC + |11 i AC, then (T ­ I C )(|  ih  |) has negative eigenvalues The resulting set of transformations on density operators are known as trace-preserving, completely positive maps

Quantum channels: examples  Adjoining ancilla:    ­ |0ih0|  Unitary transformations:   U  U y  Partial trace:  AB  tr B [  AB ]  That’s it! All channels can be built out of these operations: U  |0 i 

Further examples  The depolarizing channel:   (1-p)  + p I/2  The dephasing channel    j hj|  |ji  |0 i  Equivalent to measuring {|j i } then forgetting the outcome

One last thing you should see...  What happens if a measurement is preceded by a general quantum operation?  Leads to more general types of measurements: Positive Operator- Valued Measures (forevermore POVM)  {M k } such that M k ¸ 0,  k M k = 1  Probability of outcome k is tr[M k  ]

POVM’s: What are they good for? Try to distinguish |  0 i =|0 i and |  1 i = |+ i = (|0 i +|1 i )/2 1/2 States are non-orthogonal, so projective measurements won’t work. Let N = 1/(1+1/2 1/2 ). Exercise: M 0 = N |1 ih 1|, M 1 = N |- ih -|, M 2 = I – M 0 – M 1 is a POVM Note:* Outcome 0 implies |  1 i * Outcome 1 implies |  0 i * Outcome 2 is inconclusive Instead of imperfect distinguishability all of the time, the POVM provides perfect distinguishability some of the time.

Notions of distinguishability Basic requirement: quantum channels do not increase “distinguishability” FidelityTrace distance F( ,  )=max | h   |   i | 2 T( ,  )=|  -  | 1 F( ,  )=[Tr(  1/2  1/2 )] 2 F=0 for perfectly distinguishable F=1 for identical T=2 for perfectly distinguishable T=0 for identical T( ,  )=2max|p(k=0|  )-p(k=0|  )| where max is over measurements {M k } F(  (  ),  (  )) ¸ F( ,  )T( ,  ) ¸ T(  ( ,  (  )) Statements made today hold for both measures

Back to information theory!

Quantifying uncertainty  Let  =  x p(x) |  x ih  x | be a density operator  von Neumann entropy: H(  ) = - tr [  log   Equal to Shannon entropy of  eigenvalues  Analog of a joint random variable:   AB describes a composite system A ­ B  H(A)  = H(  A ) = H( tr B  AB )

Quantifying uncertainty: examples  H(|  ih  |) = 0  H(I/2) = 1  H(  ­  ) = H(  ) + H(  )  H(I/2 n ) = n  H(p  © (1-p)  ) = H(p,1-p) + pH(  ) + (1-p)H(  )

­­ ­ …­ … ­  Compression Source of independent copies of   : B ­ n dim(Effective supp of  B ­ n ) ~ 2 nH(B) Can compress n copies of B to a system of ~nH(B) qubits while preserving correlations with A No statistical assumptions: Just quantum mechanics! A A A BBB (aka typical subspace) [Schumacher, Petz]

The typical subspace  Diagonalize  =  x p(x) |e x ihe x |  Then  ­n =  x n p(x n ) |e x n ihe x n |  The  -typical projector  t is the projector onto the span of the |e x n ihe x n | such that x n is typical  tr[  ­ n  t ] ! 1 as n ! 1

H(B)  Quantifying information H(A)  H(B|A)  H(A|B)  Uncertainty in A when value of B is known? H(A|B)= H(AB)-H(B) |  i AB =|0 i A |0 i B +|1 i A |1 i B  B = I/2 H(A|B)  = 0 – 1 = -1 Conditional entropy can be negative ! H(AB) 

H(B)  Quantifying information H(A)  H(B|A)  Information is that which reduces uncertainty I(A;B)  H(A|B)  Uncertainty in A when value of B is known? H(A|B)= H(AB)-H(B) I(A;B) = H(A) – H(A|B) = H(A)+H(B)-H(AB) ¸ 0 H(AB) 

Sending classical information through noisy channels Physical model of a noisy channel: (Trace-preserving, completely positive map) m Encoding (  state) Decoding (measurement) m’ HSW noisy coding theorem: In the limit of many uses, the optimal rate at which Alice can send messages reliably to Bob through  is given by the ( regularization of the ) formula where

Sending classical information through noisy channels m Encoding (  state) Decoding (measurement) m’ B ­ n 2 nH(B) X 1,X 2,…,X n 2 nH(B|A)

Sending classical information through noisy channels m Encoding (  state) Decoding (measurement) m’ B ­ n 2 nH(B) X 1,X 2,…,X n 2 nH(B|A) Distinguish using well-chosen POVM

Data processing inequality (Strong subadditivity) Alice Bob time U I(A;B)  I(A;B)    I(A;B)  ¸ I(A;B) 

Optimality in the HSW theorem Assume there exists a code with rate R with perfect decoding. Let M be the random variable corresponding to the uniform distribution over messages. nR = H(M) = I(M;M’) · I(A;B)  M has nR bits of entropy Perfect decoding: M = M ’ Data processing m Encoding (  state) Decoding (measurement) m’ where mm

Sending quantum information through noisy channels Physical model of a noisy channel: (Trace-preserving, completely positive map) |  i 2 C d Encoding (TPCP map) Decoding (TPCP map) ‘‘ LSD noisy coding theorem: In the limit of many uses, the optimal rate at which Alice can reliably send qubits to Bob (1/n log d) through  is given by the ( regularization of the ) formula where Conditional entropy!

All x Random 2 n(I(X;Y)-  ) x Entanglement and privacy: More than an analogy p(y,z|x) x = x 1 x 2 … x n y=y 1 y 2 … y n z = z 1 z 2 … z n How to send a private message from Alice to Bob? AC93 Can send private messages at rate I(X;Y)-I(X;Z) Sets of size 2 n(I(X;Z)+  )

All x Random 2 n(I(X:A)-  ) x Entanglement and privacy: More than an analogy U A’->BE ­ n |  x i A’ |  i BE = U ­ n |  x i How to send a private message from Alice to Bob? D03 Can send private messages at rate I(X:A)-I(X:E) Sets of size 2 n(I(X:E)+  )

All x Random 2 n(I(X:A)-  ) x Entanglement and privacy: More than an analogy U A’->BE ­ n  x p x 1/2 |x i A |  x i A’  x p x 1/2 |x i A |  x i BE How to send a private message from Alice to Bob? SW97 D03 Can send private messages at rate I(X:A)-I(X:E)=H(A)-H(E) Sets of size 2 n(I(X:E)+  ) H(E)=H(AB)

Conclusions: Part I  Information theory can be generalized to analyze quantum information processing  Yields a rich theory, surprising conceptual simplicity  Operational approach to thinking about quantum mechanics:  Compression, data transmission, superdense coding, subspace transmission, teleportation

Some references: Part I: Standard textbooks: * Cover & Thomas, Elements of information theory. * Nielsen & Chuang, Quantum computation and quantum information. (and references therein) * Devetak, The private classical capacity and quantum capacity of a quantum channel, quant-ph/ Part II: Papers available at arxiv.org: * Devetak, Harrow & Winter, A family of quantum protocols, quant-ph/ * Horodecki, Oppenheim & Winter, Quantum information can be negative, quant-ph/