Pretty-Good Tomography Scott Aaronson MIT. Theres a problem… To do tomography on an entangled state of n qubits, we need exp(n) measurements Does this.

Slides:



Advertisements
Similar presentations
1+eps-Approximate Sparse Recovery Eric Price MIT David Woodruff IBM Almaden.
Advertisements

Closed Timelike Curves Make Quantum and Classical Computing Equivalent
How Much Information Is In A Quantum State? Scott Aaronson MIT.
How Much Information Is In Entangled Quantum States? Scott Aaronson MIT |
The Learnability of Quantum States Scott Aaronson University of Waterloo.
Quantum t-designs: t-wise independence in the quantum world Andris Ambainis, Joseph Emerson IQC, University of Waterloo.
Quantum Versus Classical Proofs and Advice Scott Aaronson Waterloo MIT Greg Kuperberg UC Davis | x {0,1} n ?
Quantum Software Copy-Protection Scott Aaronson (MIT) |
The Future (and Past) of Quantum Lower Bounds by Polynomials Scott Aaronson UC Berkeley.
The Learnability of Quantum States Scott Aaronson University of Waterloo.
How Much Information Is In A Quantum State?
Multilinear Formulas and Skepticism of Quantum Computing Scott Aaronson UC Berkeley IAS.
Limitations of Quantum Advice and One-Way Communication Scott Aaronson UC Berkeley IAS Useful?
How Much Information Is In A Quantum State? Scott Aaronson MIT |
Quantum Double Feature Scott Aaronson (MIT) The Learnability of Quantum States Quantum Software Copy-Protection.
A Full Characterization of Quantum Advice Scott Aaronson Andrew Drucker.
New Evidence That Quantum Mechanics Is Hard to Simulate on Classical Computers Scott Aaronson Parts based on joint work with Alex Arkhipov.
The Complexity of Agreement A 100% Quantum-Free Talk Scott Aaronson MIT.
QMA/qpoly PSPACE/poly: De-Merlinizing Quantum Protocols Scott Aaronson University of Waterloo.
The Equivalence of Sampling and Searching Scott Aaronson MIT.
The Computational Complexity of Linear Optics Scott Aaronson and Alex Arkhipov MIT vs.
When Qubits Go Analog A Relatively Easy Problem in Quantum Information Theory Scott Aaronson (MIT)
Solving Hard Problems With Light Scott Aaronson (Assoc. Prof., EECS) Joint work with Alex Arkhipov vs.
University of Queensland
Shortest Vector In A Lattice is NP-Hard to approximate
Approximate List- Decoding and Hardness Amplification Valentine Kabanets (SFU) joint work with Russell Impagliazzo and Ragesh Jaiswal (UCSD)
THE QUANTUM COMPLEXITY OF TIME TRAVEL Scott Aaronson (MIT)
Incremental Linear Programming Linear programming involves finding a solution to the constraints, one that maximizes the given linear function of variables.
Evaluating Classifiers
Mean, Proportion, CLT Bootstrap
Least squares CS1114
October 1999 Statistical Methods for Computer Science Marie desJardins CMSC 601 April 9, 2012 Material adapted.
Machine Learning Week 2 Lecture 2.
Computational Learning Theory PAC IID VC Dimension SVM Kunstmatige Intelligentie / RuG KI2 - 5 Marius Bulacu & prof. dr. Lambert Schomaker.
Northwestern University Winter 2007 Machine Learning EECS Machine Learning Lecture 13: Computational Learning Theory.
Department of Computer Science & Engineering University of Washington
Quantum Error Correction Michele Mosca. Quantum Error Correction: Bit Flip Errors l Suppose the environment will effect error (i.e. operation ) on our.
Probably Approximately Correct Model (PAC)
Quantum Algorithms II Andrew C. Yao Tsinghua University & Chinese U. of Hong Kong.
Experts and Boosting Algorithms. Experts: Motivation Given a set of experts –No prior information –No consistent behavior –Goal: Predict as the best expert.
Part I: Classification and Bayesian Learning
Multiplicative Weights Algorithms CompSci Instructor: Ashwin Machanavajjhala 1Lecture 13 : Fall 12.
Estimation and Hypothesis Testing. The Investment Decision What would you like to know? What will be the return on my investment? Not possible PDF for.
Probability theory: (lecture 2 on AMLbook.com)
Statistical Decision Theory
10/14/ Algorithms1 Algorithms - Ch2 - Sorting.
Introduction to Machine Learning Supervised Learning 姓名 : 李政軒.
Quantum Computing MAS 725 Hartmut Klauck NTU
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Concept learning, Regression Adapted from slides from Alpaydin’s book and slides by Professor Doina Precup, Mcgill University.
CSEP 590tv: Quantum Computing Dave Bacon July 20, 2005 Today’s Menu n Qubit registers Begin Quantum Algorithms Administrivia Superdense Coding Finish Teleportation.
CS 8751 ML & KDDComputational Learning Theory1 Notions of interest: efficiency, accuracy, complexity Probably, Approximately Correct (PAC) Learning Agnostic.
MACHINE LEARNING 3. Supervised Learning. Learning a Class from Examples Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Can small quantum systems learn? NATHAN WIEBE & CHRISTOPHER GRANADE, DEC
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
Learning From Observations Inductive Learning Decision Trees Ensembles.
CS-424 Gregory Dudek Lecture 14 Learning –Inductive inference –Probably approximately correct learning.
1 CS 391L: Machine Learning: Computational Learning Theory Raymond J. Mooney University of Texas at Austin.
Estimating standard error using bootstrap
PAC-Learning and Reconstruction of Quantum States
Introduction to Machine Learning
CH. 2: Supervised Learning
Background: Lattices and the Learning-with-Errors problem
Four approaches to Shor
More about Posterior Distributions
Computational Learning Theory
Computational Learning Theory
Scott Aaronson (UT Austin) UNM, Albuquerque, October 18, 2018
Lecture 14 Learning Inductive inference
Gentle Measurement of Quantum States and Differential Privacy *
Presentation transcript:

Pretty-Good Tomography Scott Aaronson MIT

Theres a problem… To do tomography on an entangled state of n qubits, we need exp(n) measurements Does this mean that a generic state of (say) 10,000 particles can never be learned within the lifetime of the universe? If so, this is certainly a practical problembut to me, its a conceptual problem as well

What is a quantum state? A state of the world? A state of knowledge? Whatever else it is, should at least be a useful hypothesis that encapsulates previous observations and lets us predict future ones How useful is a hypothesis that takes bits even to write down? Seems to bolster the arguments of quantum computing skeptics who think quantum mechanics will break down in the large N limit

Really were talking about Humes Problem of Induction… You see 500 ravens. Every one is black. Why does that give you any grounds whatsoever for expecting the next raven to be black? ? The answer, according to computational learning theory: In practice, we always restrict attention to some class of hypotheses vastly smaller than the class of all logically conceivable hypotheses

Probably Approximately Correct (PAC) Learning Set S called the sample space Probability distribution D over S Class C of hypotheses: functions from S to {0,1} Unknown function f C Goal: Given x 1,…,x m drawn independently from D, together with f(x 1 ),…,f(x m ), output a hypothesis h C such that with probability at least 1- over x 1,…,x m

Valiant 1984: If the hypothesis class C is finite, then any hypothesis consistent with random samples will also be consistent with a 1- fraction of future data, with probability at least 1- over the choice of samples Compression implies prediction Occams Razor Theorem And even if we discretize, its still doubly exponential in the number of qubits! But the number of quantum states is infinite!

Theorem [A. 2004]: Any n-qubit quantum state can be simulated using O(n log n log m) classical bits, where m is the number of (binary) measurements whose outcomes we care about. bits, from which Tr(E i ) can be estimated to within additive error given any E i (without knowing ). A Hint of Whats Possible… Let E=(E 1,…,E m ) be two-outcome POVMs on an n- qubit state. Then given (classical descriptions of) E and, we can produce a classical string of

Let be an n-qubit state, and let D be a distribution over two-outcome measurements. Suppose we draw measurements E 1,…,E m independently from D, and then find a hypothesis state that minimizes Quantum Occams Razor Theorem [A. 2006] Then with probability at least 1- over E 1,…,E m, provided (b i = outcome of E i ) (C a constant)

Beyond the Bayesian and Max-Lik creeds: a third way? Were not assuming any prior over states Removes a lot of problems! Instead we assume a distribution over measurements Why might that be preferable for some applications? We can control which measurements to apply, but not what the state is

Extension to process tomography? No! Suppose U|x =(-1) f(x) |x, for some random Boolean function f:{0,1} n {0,1} Then the values of f(x) constitute 2 n independently accessible bits to be learned about Yet each measurement provides at most n of the bits Hence, no analogue of my learning theorem is possible

Extension to k-outcome measurements? Sure, if we increase the number of sample measurements m by a poly(k) factor Note that theres no hope of learning to simulate 2 n -outcome measurements (i.e. measurements on all n qubits) after poly(n) sample measurements

How do we actually find ? This is a convex programming problem, which can be solved in time polynomial in the Hilbert space dimension N=2 n In general, we cant hope for better than thisfor basic computational complexity reasons Let b 1,…,b m be the binary outcomes of measurements E 1,…,E m Then choose a hypothesis state to minimize

Custom Convex Programming Method [E. Hazan, 2008] Theorem (Hazan): This algorithm returns an -optimal solution after only log(m)/ 2 iterations. Let Set S 0 := I/N For t:=0 to Compute smallest eigenvector v t of f(S t ) Compute step size t that minimizes f(S t + t (v t v t *-S t )) Set S t+1 := S t + t (v t v t *-S t )

Implementation [A. & Dechter 2008] We implemented Hazans algorithm in MATLAB Code available on request Using MITs computing cluster, we then did numerical simulations to check experimentally that the learning theorem is true

Experiments We Ran 1. Classical States (sanity check). States have form =|x x|, measurements check if i th bit is 1 or 0, distribution over measurements is uniform. 2. Linear Cluster States. States are n qubits, prepared by starting with |+ n and then applying conditional phase (P|xy =(-1) xy |xy ) to each neighboring pair. Measurements check three randomly-chosen neighboring qubits, in a basis like {|0 |+ |0,|1 |+ |1,|0 |- |1 }. Acceptance probability is always ¾.

3. Z 2 n Subgroup States. Let H be a subgroup of G=Z 2 n of order 2 n-1. States =|H H| are equal superpositions over H. Theres a measurement E g for each element g G, which checks whether g H: where U g |h =|gh for all h G. E g accepts with probability 1 if g H, or ½ if g H. Inspired by [Watrous 2000]; meant to showcase pretty-good tomography with non-commuting measurements.

Open Problems Find more convincing applications of our learning theorem Find special classes of states for which learning can be done using computation time polynomial in the number of qubits Improve the parameters of the learning theorem Experimental demonstration! DUNCE DUNCE