Neural Networks for Quantum Simulation

Slides:



Advertisements
Similar presentations
EE 690 Design of Embodied Intelligence
Advertisements

Regularization David Kauchak CS 451 – Fall 2013.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Classification for High Dimensional Problems Using Bayesian Neural Networks and Dirichlet Diffusion Trees Radford M. Neal and Jianguo Zhang the winners.
Three kinds of learning
MACHINE LEARNING 12. Multilayer Perceptrons. Neural Networks Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
EVALUATION David Kauchak CS 451 – Fall Admin Assignment 3 - change constructor to take zero parameters - instead, in the train method, call getFeatureIndices()
Radial Basis Function Networks
Hybrid AI & Machine Learning Systems Using Ne ural Network and Subsumption Architecture Libraries By Logan Kearsley.
IE 585 Introduction to Neural Networks. 2 Modeling Continuum Unarticulated Wisdom Articulated Qualitative Models Theoretic (First Principles) Models Empirical.
Outline What Neural Networks are and why they are desirable Historical background Applications Strengths neural networks and advantages Status N.N and.
CSC321: Neural Networks Lecture 2: Learning with linear neurons Geoffrey Hinton.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 9: Ways of speeding up the learning and preventing overfitting Geoffrey Hinton.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
CS 189 Brian Chu Slides at: brianchu.com/ml/
CSC321: Lecture 7:Ways to prevent overfitting
Neural Networks Vladimir Pleskonjić 3188/ /20 Vladimir Pleskonjić General Feedforward neural networks Inputs are numeric features Outputs are in.
CSC321 Lecture 5 Applying backpropagation to shape recognition Geoffrey Hinton.
Collaborative Filtering via Euclidean Embedding M. Khoshneshin and W. Street Proc. of ACM RecSys, pp , 2010.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 23: Linear Support Vector Machines Geoffrey Hinton.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Chapter 11 – Neural Nets © Galit Shmueli and Peter Bruce 2010 Data Mining for Business Intelligence Shmueli, Patel & Bruce.
MicroArray Data Analysis Candice Quadros & Amol Kothari.
An SVM Learning Approach to Robotic Grasping Raphael Pelossof December Advanced Machine Learning.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning Supervised Learning Classification and Regression
Big data classification using neural network
Learning to Compare Image Patches via Convolutional Neural Networks
Analysis of Sparse Convolutional Neural Networks
RNNs: An example applied to the prediction task
Quantum Simulation Neural Networks
Data Mining, Neural Network and Genetic Programming
ECE 5424: Introduction to Machine Learning
Computer Science and Engineering, Seoul National University
Deep Learning in HEP Large number of applications:
Lecture 24: Convolutional neural networks
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Policy Compression for MDPs
Convolutional Networks
Deep Learning Convoluted Neural Networks Part 2 11/13/
Master’s Thesis defense Ming Du Advisor: Dr. Yi Shang
Bird-species Recognition Using Convolutional Neural Network
RNNs: Going Beyond the SRN in Language Prediction
ECE408 / CS483 Applied Parallel Programming Lecture 23: Application Case Study – Electrostatic Potential Calculation.
CS 4501: Introduction to Computer Vision Training Neural Networks II
Logistic Regression & Parallel SGD
Chap. 7 Regularization for Deep Learning (7.8~7.12 )
Optimization for Fully Connected Neural Network for FPGA application
Image Classification Painting and handwriting identification
Convolutional networks
Neural Networks Geoff Hulten.
Overfitting and Underfitting
Vinit Shah, Joseph Picone and Iyad Obeid
Tuning CNN: Tips & Tricks
Convolutional Neural Networks
Artificial Intelligence 10. Neural Networks
Machine learning overview
Image Classification & Training of Neural Networks
EE 193: Parallel Computing
David Kauchak CS51A Spring 2019
Model Compression Joseph E. Gonzalez
Introduction to Neural Networks
Batch Normalization.
Image recognition.
Pose Estimation in hockey videos using convolutional neural networks
Overall Introduction for the Lecture
Evaluation David Kauchak CS 158 – Fall 2019.
Presentation transcript:

Neural Networks for Quantum Simulation Brian Barch

Problem: Quantum Simulation Want to find motion of atoms in a system Done by calculating energy, then finding force Classically easy, but not for quantum wavefunctions Want to learn a function from atomic position to system energy Regression problem Poliovirus, used in 108 atom simulation of infection

Hydrogen Dataset Configurations of hydrogen atoms and associated energy produced by iterative simulation methods 8 sets, 3000 samples each Each sample: 54 H atom positions in (x, y, z) and total energy Preprocessing: Previously: Calculated inter-atomic distances and angles, and from there symmetry function values. Normalized results Currently: Just inter-atomic distances and angles. Not normalized.

Base Model: Atomic NN structure Describe environment of each atom with symmetry functions Atomic NN uses this vector to predict energy for that atom Atomic NNs share weights atom positions total energy preprocess Separate NNs with shared weights

New Model: Atomic NN structure Reformulated atomic NN structure as convolution Used 1D convolutional layers, with atoms as the input dimension and symmetry function values as the channels Rather than preprocessing, first layer calculates symmetry function values from distances and angles Second is batch normalization layer Allows application to arbitrary numbers of atoms and GPU CNN optimization Con: much slower on complicated symmetry functions

New Model: Atomic NNs Ê Preprocess NN + Sum of atomic energies angles ѳ + Ê D distances Sum of atomic energies Sym. Func. Layer Batch normalization Atoms Conv layers

Base Model: Symmetry Functions Represent atomic environment in a form invariant to changes that don’t affect total energy (i.e. shifts, rotation, translation) Currently manually selected and used to preprocess data - NNs are trained on saved sets of symmetry functions  

New Model: Symmetry Functions Calculated per batch by first layer - Made possible by the miracle of Theano tensor broadcasting Slower than preprocessing, but… allows training of sym. func. parameters - Don’t need to rely on hand-picking values anymore - Still need to hand pick initial values for now - Used Keras’ built in analytic gradient tools Once good values are found, can use them for preprocessing

New Model: Results Wasn’t trying to optimize parameters as much as apply a new NN structure to this problem Angle-based symmetry functions required triple sum terms - too expensive to compute on the fly for my computer - haven’t had a chance to run on cluster yet Current results use a subset of symmetry functions, so aren’t comparable to previous results That being said...

Results: trainable vs static parameters Trainable symmetry function parameters had a purely beneficial effect on accuracy Reduced MSE by ~10%, faster training, less overfitting Static sym funcs Trainable sym funcs

Results continued Improved accuracy more for worse accuracies – e.g. for worse hyperparameters parameters often remained near initial value - likely because other weights optimized for those parameters Other times there was significant deviation, i.e. multiple parameters converging to nearly the same value Future work will focus on avoiding such redundancies - tried l1 regularization but it was insufficient

Lessons and Challenges From checkpoint 1: Visualizing can be challenging but is very important New Lessons: Implementation efficient surprisingly important, because it allows testing on the fly during development Linux is much easier to run code in than Windows 10 is New Challenges: Interpreting whether trained sym func parameters are meaningful Making angle based sym funcs efficient enough to run