Parallel Methods for Neuron Network

Slides:



Advertisements
Similar presentations
Neural networks Introduction Fitting neural networks
Advertisements

1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Navneet Goyal, BITS-Pilani Perceptrons. Labeled data is called Linearly Separable Data (LSD) if there is a linear decision boundary separating the classes.
Optimization Tutorial
Tuomas Sandholm Carnegie Mellon University Computer Science Department
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Artificial Neural Networks
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
CES 514 – Data Mining Lecture 8 classification (contd…)
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Lecture 4 Neural Networks ICS 273A UC Irvine Instructor: Max Welling Read chapter 4.
Introduction to Recurrent neural networks (RNN), Long short-term memory (LSTM) Wenjie Pei In this coffee talk, I would like to present you some basic.
Artificial Neural Networks
ICS 273A UC Irvine Instructor: Max Welling Neural Networks.
KOHONEN SELF ORGANISING MAP SEMINAR BY M.V.MAHENDRAN., Reg no: III SEM, M.E., Control And Instrumentation Engg.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Adaptive CSMA under the SINR Model: Fast convergence using the Bethe Approximation Krishna Jagannathan IIT Madras (Joint work with) Peruru Subrahmanya.
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Machine Learning Chapter 4. Artificial Neural Networks
Discriminant Functions
An informal description of artificial neural networks John MacCormick.
Learning a Small Mixture of Trees M. Pawan Kumar Daphne Koller Aim: To efficiently learn a.
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
M Machine Learning F# and Accord.net. Alena Dzenisenka Software architect at Luxoft Poland Member of F# Software Foundation Board of Trustees Researcher.
Non-Bayes classifiers. Linear discriminants, neural networks.
 Based on observed functioning of human brain.  (Artificial Neural Networks (ANN)  Our view of neural networks is very simplistic.  We view a neural.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Intro. ANN & Fuzzy Systems Lecture 20 Clustering (1)
An Evolutionary Algorithm for Neural Network Learning using Direct Encoding Paul Batchis Department of Computer Science Rutgers University.
4 Starting Tips to Keep Your Car in Top Condition
CS 388: Natural Language Processing: Neural Networks
Artificial Neural Networks
Fall 2004 Perceptron CS478 - Machine Learning.
A Fast Trust Region Newton Method for Logistic Regression
COMP24111: Machine Learning and Optimisation
Matt Gormley Lecture 16 October 24, 2016
A Simple Artificial Neuron
ICS 491 Big Data Analytics Fall 2017 Deep Learning
Classification with Perceptrons Reading:
Deep Learning Qing LU, Siyuan CAO.
Structure learning with deep autoencoders
Prof. Carolina Ruiz Department of Computer Science
Limitations of Traditional Deep Network Architectures
Gradient Checks for ANN
Variations on Backpropagation.
Matrix Methods Summary
Convolutional networks
Neural Networks Geoff Hulten.
Neural Networks ICS 273A UC Irvine Instructor: Max Welling
Machine Learning: Lecture 4
Ch4: Backpropagation (BP)
Machine Learning: UNIT-2 CHAPTER-1
Artificial Neural Networks
Neural networks (1) Traditional multi-layer perceptrons
Artificial Intelligence 10. Neural Networks
CS639: Data Management for Data Science
COSC 4368 Machine Learning Organization
Variations on Backpropagation.
Chapter 8: Recurrent associative networks and episodic memory
The Updated experiment based on LSTM
Ch4: Backpropagation (BP)
ADMM and DSO.
CSC 578 Neural Networks and Deep Learning
CRCV REU 2019 Aaron Honculada.
Artificial Neural Network learning
Prof. Carolina Ruiz Department of Computer Science
Patterson: Chap 1 A Review of Machine Learning
Presentation transcript:

Parallel Methods for Neuron Network Haihao Lu and Yuanchu Dang

Machine Learning and Artificial Neuron Network

Problem of Interest: ANN W is the weight b is the bias ϕ is activation function x is feature y is label l is loss function

Drawback of SGD SGD is a first-order method, thus it converges slow. SGD suffers from vanishing gradient problem Most importantly, it is hard to parallelize SGD

Equivalent Problem Relaxed Problem

Alternating Minimization W-update z-update one or more steps of damped Newton cheap to compute Hessian parallel computing

RNN

RNN Equivalent Problem Relaxed Problem

With L_2 Regulation W-update z-update one or more steps of damped Newton cheap to compute Hessian parallel computing

Code neuralNetwork.jl https://github.com/Yuanchu/neuralNetwor k

Numerical Test: Auto-Encoder