A brief introduction to neural network

Slides:



Advertisements
Similar presentations
Slides from: Doug Gray, David Poole
Advertisements

Neural networks Introduction Fitting neural networks
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Artificial Neural Networks ECE 398BD Instructor: Shobha Vasudevan.
Supervised and Unsupervised learning and application to Neuroscience Cours CA6b-4.
Neural Network Introduction Hung-yi Lee. Review: Supervised Learning Training: Pick the “best” Function f * Training Data Model Testing: Hypothesis Function.
Non-Bayes classifiers. Linear discriminants, neural networks.
Neural Networks Vladimir Pleskonjić 3188/ /20 Vladimir Pleskonjić General Feedforward neural networks Inputs are numeric features Outputs are in.
Object Recognizing. Deep Learning Success in 2012 DeepNet and speech processing.
Convolutional Neural Network
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
Intro. ANN & Fuzzy Systems Lecture 11. MLP (III): Back-Propagation.
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Neural networks.
Neural networks and support vector machines
Welcome deep loria !.
Big data classification using neural network
Sentiment analysis using deep learning methods
Deep Learning: What is it good for? R. Burgmann
Deep Feedforward Networks
CS 6501: 3D Reconstruction and Understanding Convolutional Neural Networks Connelly Barnes.
Summary of “Efficient Deep Learning for Stereo Matching”
Deep Learning Amin Sobhani.
Machine Learning & Deep Learning
Data Mining, Neural Network and Genetic Programming
ECE 5424: Introduction to Machine Learning
Computer Science and Engineering, Seoul National University
DeepCount Mark Lenson.
第 3 章 神经网络.
COMP24111: Machine Learning and Optimisation
Mastering the game of Go with deep neural network and tree search
Applications of Deep Learning and how to get started with implementation of deep learning Presentation By : Manaswi Advisor : Dr.Chinmay.
Restricted Boltzmann Machines for Classification
Deep Learning Libraries
CSE 473 Introduction to Artificial Intelligence Neural Networks
Neural Networks CS 446 Machine Learning.
Classification with Perceptrons Reading:
Intelligent Information System Lab
Classification of Hand-Written Digits Using Scattering Convolutional Network Dongmian Zou Advisor: Professor Radu Balan.
ECE 6504 Deep Learning for Perception
State-of-the-art face recognition systems
Introduction to Neural Networks
Goodfellow: Chap 6 Deep Feedforward Networks
CS 4501: Introduction to Computer Vision Training Neural Networks II
Artificial Intelligence Chapter 3 Neural Networks
Smart Robots, Drones, IoT
network of simple neuron-like computing elements
[Figure taken from googleblog
A Proposal Defense On Deep Residual Network For Face Recognition Presented By SAGAR MISHRA MECE
Neural Networks Geoff Hulten.
Artificial Intelligence Chapter 3 Neural Networks
Overview of Neural Network Architecture Assignment Code
Neural Networks II Chen Gao Virginia Tech ECE-5424G / CS-5824
Artificial Intelligence Chapter 3 Neural Networks
Artificial Intelligence 10. Neural Networks
Machine learning overview
实习生汇报 ——北邮 张安迪.
Artificial Intelligence Chapter 3 Neural Networks
Neural Networks II Chen Gao Virginia Tech ECE-5424G / CS-5824
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Introduction to Neural Networks
Deep Learning Libraries
EE 193/Comp 150 Computing with Biological Parts
Artificial Intelligence Chapter 3 Neural Networks
Principles of Back-Propagation
Outline Announcement Neural networks Perceptrons - continued
An introduction to neural network and machine learning
Overall Introduction for the Lecture
Presentation transcript:

A brief introduction to neural network

Machine learning/neural network in physics research “Identifying quantum phase transitions using artificial neural network on experimental data,” arxiv:1809.05519, B. S. Rem, et al. “Galaxy Zoo: reproducing galaxy morphologies via machine learning,” M. Banerji, et al, Monthly Notices …, 406, 342, (2010). “Prediction of thermal boundary resistance by the machine learning method,” T Zhan, et al, Sci Rep 7, 7109 (2017). “Searching for exotic particles in high-energy physics with deep learning,” P. Baldi, et al, Nature Comm 5, 4308 (2014).

Biological neuron Picture from http://cs231n.github.io/neural-networks-1/

A “mathematical” neuron x1 w1 w2 output x2 inputs F(.) y … wN xN F (ReLU)   The function F is called activation, and the particular form F=0 if S < 0 and F=S is called rectified linear unit or reLU. S  

(feedforward) Neural network Y1 Y2 X (input layer) x1 Y (output layer) x2 x3

Supervised learning Determine the W(i) and b(i) with a training set of inputs {x} to minimize the predicted differences. Least square errors: we minimize (y(j) is the output of j-th sample and d(j) is expected value):

Classification problems Hand written digits in 28x28 black/white pixels With 10 output neutrons answering the question: is it 0? is it 1? …, is it 9? 60000 for a training set, 10000 examples for testing set. From MNIST dataset.

Network x1 y0 y1 x2 The input “2” is a 28x28 bitmap of xi of 0 and 1 of 784 numbers. … … … y9 x784 The predicted digit is j such that yj is a maximum, i.e., y gives a score for each of the 10 possibilities. The last step does not apply the F function.

Hinge loss function Example: Given an image for 2, the 10 outputs (scores), let’s say, are 10, 2, 8, …, 13 for j = 0, 1, .., 9. Clearly, j = 2 should be the correct answer. Let take Δ=1. Then the loss is max(0, 10-8+1) + max(0,2-8+1) + … + max(0,13-8+1) = 3 + 6 = 9. {Incorrect scores get a large penalty} The learning algorithm tries to minimize total L summed over each sample i with a “regularization” term: Lambda is called super-parameter and is not changed.

Softmax or cross-entropy loss Softmax method for judging the correctness of result is given by the following formula for the i-th sample. We can interpret Pj as a probability of having value j.

Update the network The steepest descent or (stochastic) gradient descent To evaluate the gradient efficiently we use something called back propagation on the network. W

The gradient

Preventing under-fit and over-fit by adjust λ From “Deep learning”, Goodfellow, et al, page 119.

Convolutional network Convolutional networks are simply neural networks that use convolution in place of general matrix multiplication (Wx) in at least one of their layers. Pooling: replace the results by some static From Figure 9.8 in Goodfellow, et al, page 358.

Convolution x1 Convolution in math sense x2 x3 Each neuron is connected to only three inputs based on locality. Three weights w1, w2, w3 are the same on all of the neurons. … x784

Max Pool From http://cs231n.github.io/convolutional-networks/ This is very much like the real space RG transform in physics.

Other Topics not covered Recurrent network Boltzmann Machine/statistical mechanics etc

Tensorflow TensorFlow is an open source software library from google for high performance numerical computation. an open-source machine learning library for research and production. In Python, C++, javaScript

Example codes import tensorflow as tf mnist = tf.keras.datasets.mnist (x_train, y_train),(x_test, y_test) = mnist.load_data() x_train, x_test = x_train / 255.0, x_test / 255.0 model = tf.keras.models.Sequential([ tf.keras.layers.Flatten(), tf.keras.layers.Dense(512, activation=tf.nn.relu), tf.keras.layers.Dropout(0.2), tf.keras.layers.Dense(10, activation=tf.nn.softmax) ]) model.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['accuracy']) model.fit(x_train, y_train, epochs=5) model.evaluate(x_test, y_test)

Research Project Can we use a convolutional neutral network to determine the Tc accurately? When the network is trained with only low (ferromagnetic phase) and high temperature (paramagnetic phase) spin configurations for the two-dimensional Ising model.

References Stanford Univ CS231n “Convolutional Neural Networks for Visual Recognition,” http://cs231n.github.io/ “Deep Learning,” Goodfellow, Bengio, and Courville, MIT press (2016). “Neural Networks”, Haykin, 3rd ed, Pearson (2008).