Michael A. Nielsen University of Queensland Quantum entropy Goals: 1.To define entropy, both classical and quantum. 2.To explain data compression, and.

Slides:



Advertisements
Similar presentations
Pretty-Good Tomography Scott Aaronson MIT. Theres a problem… To do tomography on an entangled state of n qubits, we need exp(n) measurements Does this.
Advertisements

Solving Hard Problems With Light Scott Aaronson (Assoc. Prof., EECS) Joint work with Alex Arkhipov vs.
Lecture 2: Basic Information Theory TSBK01 Image Coding and Data Compression Jörgen Ahlberg Div. of Sensor Technology Swedish Defence Research Agency (FOI)
Entropy and Information Theory
Evaluating Classifiers
Information Theory EE322 Al-Sanie.
An introduction to Data Compression
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Entropy in the Quantum World Panagiotis Aleiferis EECS 598, Fall 2001.
SIMS-201 Compressing Information. 2  Overview Chapter 7: Compression Introduction Entropy Huffman coding Universal coding.
Chapter 6 Information Theory
ENGS Lecture 8 ENGS 4 - Lecture 8 Technology of Cyberspace Winter 2004 Thayer School of Engineering Dartmouth College Instructor: George Cybenko,
SWE 423: Multimedia Systems
UCB Claude Shannon – In Memoriam Jean Walrand U.C. Berkeley
Fundamental limits in Information Theory Chapter 10 :
Superdense coding. How much classical information in n qubits? Observe that 2 n  1 complex numbers apparently needed to describe an arbitrary n -qubit.
Avraham Ben-Aroya (Tel Aviv University) Oded Regev (Tel Aviv University) Ronald de Wolf (CWI, Amsterdam) A Hypercontractive Inequality for Matrix-Valued.
Information Theory and Security. Lecture Motivation Up to this point we have seen: –Classical Crypto –Symmetric Crypto –Asymmetric Crypto These systems.
1 Chapter 1 Introduction. 2 Outline 1.1 A Very Abstract Summary 1.2 History 1.3 Model of the Signaling System 1.4 Information Source 1.5 Encoding a Source.
Lecture 2: Basic Information Theory Thinh Nguyen Oregon State University.
Review of Probability and Random Processes
Information Theory and Security
Noise, Information Theory, and Entropy
Probability and Statistics Dr. Saeid Moloudzadeh Axioms of Probability/ Basic Theorems 1 Contents Descriptive Statistics Axioms of Probability.
1 Introduction to Quantum Information Processing QIC 710 / CS 768 / PH 767 / CO 681 / AM 871 Richard Cleve QNC 3129 Lecture 18 (2014)
Information Theory & Coding…
Introduction to the Theory of Computation
ENTROPIC CHARACTERISTICS OF QUANTUM CHANNELS AND THE ADDITIVITY PROBLEM A. S. Holevo Steklov Mathematical Institute, Moscow.
All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.
Dr.-Ing. Khaled Shawky Hassan
A Few Simple Applications to Cryptography Louis Salvail BRICS, Aarhus University.
Binomial Distributions Calculating the Probability of Success.
Probability Rules!. ● Probability relates short-term results to long-term results ● An example  A short term result – what is the chance of getting a.
Theory of Computing Lecture 15 MAS 714 Hartmut Klauck.
What is being plotted?. Answer: Number of papers with “quantum entanglement” in title or abstract N. D. Mermin, Phys. Rev. Lett. (1990) Entanglement is.
Channel Capacity.
COMMUNICATION NETWORK. NOISE CHARACTERISTICS OF A CHANNEL 1.
1 New Coins from old: Computing with unknown bias Elchanan Mossel, U.C. Berkeley
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Summer 2004CS 4953 The Hidden Art of Steganography A Brief Introduction to Information Theory  Information theory is a branch of science that deals with.
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
Information & Communication INST 4200 David J Stucki Spring 2015.
Grand Paradigms. How do you see your place in the world?
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Coding Theory Efficient and Reliable Transfer of Information
Michael A. Nielsen University of Queensland Quantum Mechanics II: Examples Goals: 1.To apply the principles introduced in the last lecture to some illustrative.
Basic Principles (continuation) 1. A Quantitative Measure of Information As we already have realized, when a statistical experiment has n eqiuprobable.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Entropy (YAC- Ch. 6)  Introduce the thermodynamic property called Entropy (S)  Entropy is defined using the Clausius inequality  Introduce the Increase.
Binomial Distributions Chapter 5.3 – Probability Distributions and Predictions Mathematics of Data Management (Nelson) MDM 4U.
Presented by Minkoo Seo March, 2006
Quantum Cryptography Antonio Acín
1 CSCD 433 Network Programming Fall 2013 Lecture 5a Digital Line Coding and other...
Rate Distortion Theory. Introduction The description of an arbitrary real number requires an infinite number of bits, so a finite representation of a.
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
Chapter 6 Large Random Samples Weiqi Luo ( 骆伟祺 ) School of Data & Computer Science Sun Yat-Sen University :
1 Review of Probability and Random Processes. 2 Importance of Random Processes Random variables and processes talk about quantities and signals which.
Binomial Distributions Chapter 5.3 – Probability Distributions and Predictions Mathematics of Data Management (Nelson) MDM 4U Authors: Gary Greer (with.
3/7/20161 Now it’s time to look at… Discrete Probability.
UNIT I. Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity.
PROBABILITY AND COMPUTING RANDOMIZED ALGORITHMS AND PROBABILISTIC ANALYSIS CHAPTER 1 IWAMA and ITO Lab. M1 Sakaidani Hikaru 1.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Shannon Entropy Shannon worked at Bell Labs (part of AT&T)
Information Theory Michael J. Watts
A Brief Introduction to Information Theory
Quantum Information Theory Introduction
Econometric Models The most basic econometric model consists of a relationship between two variables which is disturbed by a random error. We need to use.
Lecture 2: Basic Information Theory
Presentation transcript:

Michael A. Nielsen University of Queensland Quantum entropy Goals: 1.To define entropy, both classical and quantum. 2.To explain data compression, and its connection with entropy. 3.To explain some of the basic properties of entropy, both classical and quantum.

What is an information source? We need a simple model of an information source. The model might not be realistic, but it should give rise to a theory of information that can be applied to realistic situations.

Discrete iid sources Definition: Each output from a discrete information source comes from a finite set. We will mostly be concerned with the case where the alphabet consists of 0 and … More generally, there is no loss of generality in supposing that the alphabet is 0,…,n-1.

Discrete iid sources We will model sources using a probability distribution for the output of the source … Definition: Each output from an iid (independent and identically distributed) source is independent of the other outputs, and each output has the same distribution. Example: A sequence of coin tosses of a biased coin with probability p of heads, and 1-p of tails. More generally, the distribution on alphabet symbols is denoted p 0,p 1,…,p n.

What other sources are discrete iid? Most interesting sources are not. However, lots of sources can be approximated as iid – even with English text this is not a bad approximation. “What a piece of work is a man! how noble in reason! how infinite in faculties! in form and moving how express and admirable! in action how like an angel! in apprehension how like a god! the beauty of the world, the paragon of animals! And yet to me what is this quintessence of dust?” Many sources can be described as stationary, ergodic sequences of random variables, and similar results apply. Research problem: Find a good quantum analogue of “stationary, ergodic sources” for, and extend quantum information theory to those sources. (Quantum Shannon-Macmillan-Breiman theorem?)

How can we quantify the rate at which information is being produced by a source? Two broad approaches Axiomatic approach: Write down desirable axioms which a measure of information “should” obey, and find such a measure. Operational approach: Based on the “fundamental program” of information science. How many bits are needed to store the output of the source, so the output can be reliably recovered?

Historical origin of data compression “He can compress the most words into the smallest ideas of any man I ever met.”

Data compression abcde… { n uses nR bits compressdecompress abcde… What is the minimal value of R that allows reliable decompression? We will define the minimal value to be the information content of the source.

Data compression Suppose we flip coins, getting heads with probability p, and tails with probability 1-p. For large values of n, it is very likely that we will get roughly np heads, and n(1-p) tails.

Data compression

Data compression: the algorithm The two critical facts n+1 bits nH(p,1-p)+1 bits On average, only H(p,1-p) bits were required to store the compressed string, per use of the source. 1. x 1 2. x 2 3. x 3 4. x 4 …

Variants on the data compression algorithm Our algorithm is for large n, gives variable-length output that achieves the Shannon entropy on average. The algorithm never makes an error in recovery. Algorithms for small n can be designed that do almost as well. Fixed-length compression Errors must always occur in a fixed-length scheme, but it does work with probability approaching one.

Why it’s impossible to compress below the Shannon rate

Basic properties of the entropy

Why’s this notion called entropy, anyway? “When the American scientist Claude Shannon found that the mathematical formula of Boltzmann defined a useful quantity in information theory, he hesitated to name this newly discovered quantity entropy because of its philosophical baggage. The mathematician John Von [sic] Neumann encouraged Shannon to go ahead with the name entropy, however, since`no one knows what entropy is, so in a debate you will always have the advantage.’ ” From the American Heritage Book of English Usage (1996):

What else can be done with the Shannon entropy? Quantum processes teleportation communication cryptography theory of entanglement Shor’s algorithm quantum error-correction Complexity quantum phase transitions 1.Identify a physical resource – energy, time, bits, space, entanglement. 2.Identify an information processing task – data compression, information transmission, teleportation. 3. Identify a criterion for success. How much of 1 do I need to achieve 2, while satisfying 3?

What else can be done with the Shannon entropy? Classical processes data compression networks cryptography thermodynamics reliable communication in the presence of noise Complexity gambling quantum information

What is a quantum information source? Example: “Semiclassical coin toss” Example: “Quantum coin toss”

Quantum data compression decompressioncompression

What’s the best possible rate for quantum data compression? “Semiclassical coin toss” “Quantum coin toss”

Quantum entropy

Basic properties of the von Neumann entropy

The typical subspace

Outline of Schumacher’s data compression

Recall classical to quantum circuits

How to measure P, Q

Outline of Schumacher’s data compression

Schumacher compression

The idea of the proof is similar to Shannon’s proof. Two known proofs: One is a complicated kludge, done from first principles. The other proof is an elegant “easy” proof that relies on other deep theorems.