Quantum Boltzmann Machine

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Nathan Wiebe, Ashish Kapoor and Krysta Svore Microsoft Research ASCR Workshop Washington DC Quantum Deep Learning.
Spin chains and channels with memory Martin Plenio (a) & Shashank Virmani (a,b) quant-ph/ , to appear prl (a)Institute for Mathematical Sciences.
CS 678 –Boltzmann Machines1 Boltzmann Machine Relaxation net with visible and hidden units Learning algorithm Avoids local minima (and speeds up learning)
1 Optimization Algorithms on a Quantum Computer A New Paradigm for Technical Computing Richard H. Warren, PhD Optimization.
Approaches to Data Acquisition The LCA depends upon data acquisition Qualitative vs. Quantitative –While some quantitative analysis is appropriate, inappropriate.
Boris Altshuler Columbia University Anderson Localization against Adiabatic Quantum Computation Hari Krovi, Jérémie Roland NEC Laboratories America.
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
Adiabatic Quantum Computation with Noisy Qubits Mohammad Amin D-Wave Systems Inc., Vancouver, Canada.
Visual Recognition Tutorial
Complexity of simulating quantum systems on classical computers Barbara Terhal IBM Research.
Lattice regularized diffusion Monte Carlo
Schrödinger’s Elephants & Quantum Slide Rules A.M. Zagoskin (FRS RIKEN & UBC) S. Savel’ev (FRS RIKEN & Loughborough U.) F. Nori (FRS RIKEN & U. of Michigan)
Chapter 14 Simulation. Monte Carlo Process Statistical Analysis of Simulation Results Verification of the Simulation Model Computer Simulation with Excel.
Quantum Mechanics from Classical Statistics. what is an atom ? quantum mechanics : isolated object quantum mechanics : isolated object quantum field theory.
Efficient Quantum State Tomography using the MERA in 1D critical system Presenter : Jong Yeon Lee (Undergraduate, Caltech)
1 Assessment of Imprecise Reliability Using Efficient Probabilistic Reanalysis Farizal Efstratios Nikolaidis SAE 2007 World Congress.
Introduction to Monte Carlo Methods D.J.C. Mackay.
Monte Carlo Simulation of Ising Model and Phase Transition Studies By Gelman Evgenii.
Monte Carlo Simulation of Interacting Electron Models by a New Determinant Approach Mucheng Zhang (Under the direction of Robert W. Robinson and Heinz-Bernd.
1 IE 607 Heuristic Optimization Simulated Annealing.
Chapter 7 Other Important NN Models Continuous Hopfield mode (in detail) –For combinatorial optimization Simulated annealing (in detail) –Escape from local.
Topic Models in Text Processing IR Group Meeting Presented by Qiaozhu Mei.
Chapter 14 Monte Carlo Simulation Introduction Find several parameters Parameter follow the specific probability distribution Generate parameter.
Number of Blocks per Pole Diego Arbelaez. Option – Number of Blocks per Pole Required magnetic field tolerance of ~10 -4 For a single gap this can be.
CSC2535: Computation in Neural Networks Lecture 11: Conditional Random Fields Geoffrey Hinton.
Ran El-Yaniv and Dmitry Pechyony Technion – Israel Institute of Technology, Haifa, Israel Transductive Rademacher Complexity and its Applications.
Merlin-Arthur Games and Stoquastic Hamiltonians B.M. Terhal, IBM Research based on work with Bravyi, Bessen, DiVincenzo & Oliveira. quant-ph/ &
9. Convergence and Monte Carlo Errors. Measuring Convergence to Equilibrium Variation distance where P 1 and P 2 are two probability distributions, A.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
Learning Lateral Connections between Hidden Units Geoffrey Hinton University of Toronto in collaboration with Kejie Bao University of Toronto.
Geoffrey Hinton CSC2535: 2013 Lecture 5 Deep Boltzmann Machines.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Markov Random Fields Probabilistic Models for Images
Adiabatic quantum computation and stoquastic Hamiltonians B.M. Terhal IQI, RWTH Aachen A review of joint work with S. Bravyi, D.P.DiVincenzo and R. Oliveira.
8. Selected Applications. Applications of Monte Carlo Method Structural and thermodynamic properties of matter [gas, liquid, solid, polymers, (bio)-macro-
The Group Lasso for Logistic Regression Lukas Meier, Sara van de Geer and Peter Bühlmann Presenter: Lu Ren ECE Dept., Duke University Sept. 19, 2008.
CSC 2535 Lecture 8 Products of Experts Geoffrey Hinton.
: Chapter 3: Maximum-Likelihood and Baysian Parameter Estimation 1 Montri Karnjanadecha ac.th/~montri.
1 Dorit Aharonov Hebrew Univ. & UC Berkeley Adiabatic Quantum Computation.
A Study of Error-Correcting Codes for Quantum Adiabatic Computing Omid Etesami Daniel Preda CS252 – Spring 2007.
CSC2535 Lecture 4 Boltzmann Machines, Sigmoid Belief Nets and Gibbs sampling Geoffrey Hinton.
General ideas to communicate Dynamic model Noise Propagation of uncertainty Covariance matrices Correlations and dependencs.
Workshop on Optimization in Complex Networks, CNLS, LANL (19-22 June 2006) Application of replica method to scale-free networks: Spectral density and spin-glass.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 18 Learning Boltzmann Machines Geoffrey Hinton.
Adiabatic Quantum Computation with Noisy Qubits M.H.S. Amin D-Wave Systems Inc., Vancouver, Canada.
Dd Generalized Optimal Kernel-based Ensemble Learning for HS Classification Problems Generalized Optimal Kernel-based Ensemble Learning for HS Classification.
Expectation-Maximization (EM) Algorithm & Monte Carlo Sampling for Inference and Approximation.
Discrete optimisation problems on an adiabatic quantum computer
Copyright© 2012, D-Wave Systems Inc. 1 Quantum Boltzmann Machine Mohammad Amin D-Wave Systems Inc.
Determining the Complexity of the Quantum Adiabatic Algorithm using Monte Carlo Simulations A.P. Young, University of California Santa Cruz
Adiabatic Quantum Computing Josh Ball with advisor Professor Harsh Mathur Problems which are classically difficult to solve may be solved much more quickly.
Physical Fluctuomatics (Tohoku University) 1 Physical Fluctuomatics 13th Quantum-mechanical extensions of probabilistic information processing Kazuyuki.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 17: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Beginner’s Guide to Quantum Computing Graduate Seminar Presentation Oct. 5, 2007.
Physical Fluctuomatics 13th Quantum-mechanical extensions of probabilistic information processing Kazuyuki Tanaka Graduate School of Information Sciences,
Some Slides from 2007 NIPS tutorial by Prof. Geoffrey Hinton
D-Wave Quantum Computing Quantum Computing for the Real World Today
Statistical Quality Control, 7th Edition by Douglas C. Montgomery.
Theoretical Investigations at
CSC321: Neural Networks Lecture 19: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Multimodal Learning with Deep Boltzmann Machines
Structure learning with deep autoencoders
Unfolding Problem: A Machine Learning Approach
Regulation Analysis using Restricted Boltzmann Machines
Boltzmann Machine (BM) (§6.4)
Chapter 14 Monte Carlo Simulation
Computational approaches for quantum many-body systems
A quantum machine learning algorithm based on generative models
Presentation transcript:

Quantum Boltzmann Machine Mohammad Amin D-Wave Systems Inc. 1

Not the only use of QA Maybe not the best use of QA

Adiabatic Quantum Computation H(t) = (1-s)HD + sHP , s = t/tf energy levels gmin tf ~ (1/gmin)2 Solution Initial state s 1

Thermal Noise System Bath Interaction Dynamical freeze-out P0 s 1 energy levels kBT Dynamical freeze-out P0 s 1

Open quantum calculations of a 16 qubit random problem Classical energies

Equilibration Can Cause Correlation Correlation with simulated annealing Hen et al., PRA 92, 042325 (2015)

Equilibration Can Cause Correlation Boixo et al., Nature Phys. 10, 218 (2014) Correlation with Quantum Monte Carlo

Equilibration Can Cause Correlation Correlation with spin vector Monte Carlo Shin et al., arXiv:1401.7087 SVMC SVMC

Equilibration Can Mask Quantum Speedup Brooke et al., Science 284, 779 (1999) Quantum advantage is expected to be dynamical

Equilibration Can Mask Quantum Speedup Ronnow et al., Science 345, 420 (2014) Hen et al., arXiv:1502.01663 King et al., arXiv:1502.02098 Equilibrated probability!!! Computation time is independent of dynamics!

Residual Energy vs Annealing Time 50 random problems, 100 samples per problem per annealing time Bimodal (J=-1, +1 , h=0) Mean residual energy Lowest residual energy Annealing time (ms)

Residual Energy vs Annealing Time 50 random problems, 100 samples per problem per annealing time Frustrated loops (a=0.25) Bimodal (J=-1, +1 , h=0) Annealing time (ms) Annealing time (ms)

Boltzmann sampling is #P Quantum Boltzmann Distribution? harder than NP What can we do with a Quantum Boltzmann Distribution?

arXiv:1601.02036 Evgeny Andriyash Jason Rolfe Bohdan Kulchytskyy Roger Melko

Machine Learning in our Daily Life

Introduction to Machine Learning Data Model 3 Model Unseen data

Data Model Probabilistic Models Training: Tune q such that Probability distribution Data Variables Parameters q Model Training: Tune q such that

Boltzmann distribution (b =1) Boltzmann Machine Data Variables Parameters q Model Boltzmann distribution (b =1)

Boltzmann Machine Ising model: spins parameters

za Fully Visible BM Hz only has O(N2) parameters needs O(2N) parameters to be fully described

Adding Hidden Variables z i zn visible hidden za = (zn , zi) hidden visible

We need an efficient way to calculate Training a BM Tune such that We need an efficient way to calculate Maximize log-likelihood: Or minimize: gradient descent technique training rate

Calculating the Gradient Average with clamped visibles Unclamped average

Training Ising Hamiltonian Parameters Clamped average Unclamped average Gradients can be estimated using sampling!

Is it possible to train a quantum Boltzmann machine? Question: Is it possible to train a quantum Boltzmann machine? Ising Hamiltonian Transverse Ising Hamiltonian

Transverse Ising Hamiltonian

Quantum Boltzmann Distribution Boltzmann probability distribution: Density matrix: Projection operator Identity matrix

Gradient Descent = Classically: Clamped average Unclamped average =

Calculating the Gradient Gradient cannot be estimated using sampling! ≠ ≠ Clamped average Unclamped average

Two Useful Properties of Trace Golden-Thompson inequality: For Hermitian matrices A and B

Finding lower bounds Golden-Thompson inequality

Finding lower bounds Lower bound for log-likelihood Golden-Thompson inequality Lower bound for log-likelihood

Calculating the Gradients Minimize the upper bound ? Unclamped average

Visible qubits are clamped to their classical values given by the data Clamped Hamiltonian for Infinite energy penalty for states different from v Visible qubits are clamped to their classical values given by the data

We can now use sampling to estimate the steps Estimating the Steps Clamped average Unclamped average We can now use sampling to estimate the steps

Training the Transverse Field (Ga) Minimizing the upper bound: for all visible qubits, thus cannot be estimated from measurements Two problems: Gn cannot be trained using the bound

Example: 10-Qubit QBM Graph: fully connected (K10), fully visible

Example: 10-Qubit QBM Training set: M-modal distribution p = 0.9 M = 8 Random spin orientation Single mode: Hamming distance p = 0.9 Multi-mode: M = 8

Exact Diagonalization Results KL-divergence: Classical BM Bound gradient D=2 Exact gradient (D is trained) D final = 2.5

Sampling from D-Wave Probabilities cross at the anticrossing Dickson et al., Nat. Commun. 4, 1903 (2013) Probabilities cross at the anticrossing

Conclusions: A quantum annealer can provide fast samples of quantum Boltzmann distribution QBM can be trained by sampling QBM may learn some distributions better than classical BM See arXiv:1601.02036