READINGS IN DEEP LEARNING 4 Sep 2013. ADMINSTRIVIA New course numbers (11-785/786) are assigned – Should be up on the hub shortly Lab assignment 1 up.

Slides:



Advertisements
Similar presentations
© Negnevitsky, Pearson Education, Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Advertisements

Accessing I/O Devices Processor Memory BUS I/O Device 1 I/O Device 2.
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
NEURAL NETWORKS Perceptron
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Artificial Neural Network
Neural Network Intro Slides
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Artificial neural networks:
A Parallel Computational Model for Heterogeneous Clusters Jose Luis Bosque, Luis Pastor, IEEE TRASACTION ON PARALLEL AND DISTRIBUTED SYSTEM, VOL. 17, NO.
Scaling Distributed Machine Learning with the BASED ON THE PAPER AND PRESENTATION: SCALING DISTRIBUTED MACHINE LEARNING WITH THE PARAMETER SERVER – GOOGLE,
September 7, 2010Neural Networks Lecture 1: Motivation & History 1 Welcome to CS 672 – Neural Networks Fall 2010 Instructor: Marc Pomplun Instructor: Marc.
Slides are based on Negnevitsky, Pearson Education, Lecture 8 Artificial neural networks: Unsupervised learning n Introduction n Hebbian learning.
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
Design, Implementation, and Evaluation of Differentiated Caching Services Ying Lu, Tarek F. Abdelzaher, Avneesh Saxena IEEE TRASACTION ON PARALLEL AND.
AN INTERACTIVE TOOL FOR THE STOCK MARKET RESEARCH USING RECURSIVE NEURAL NETWORKS Master Thesis Michal Trna
1 Pendahuluan Pertemuan 1 Matakuliah: T0293/Neuro Computing Tahun: 2005.
Lecture 4 Neural Networks ICS 273A UC Irvine Instructor: Max Welling Read chapter 4.
Dr.-Ing. Erwin Sitompul President University Lecture 1 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 1/1
September 28, 2010Neural Networks Lecture 7: Perceptron Modifications 1 Adaline Schematic Adjust weights i1i1i1i1 i2i2i2i2 inininin …  w 0 + w 1 i 1 +
NEURAL NETWORKS Introduction
Artificial Neural Network
SYSTEMS SUPPORT FOR GRAPHICAL LEARNING Ken Birman 1 CS6410 Fall /18/2014.
Artificial neural networks:
CS-485: Capstone in Computer Science Artificial Neural Networks and their application in Intelligent Image Processing Spring
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Operating System Review September 10, 2012Introduction to Computer Security ©2004 Matt Bishop Slide #1-1.
Artificial Neural Network Unsupervised Learning
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Example: Sorting on Distributed Computing Environment Apr 20,
09/21/2010CS4961 CS4961 Parallel Programming Lecture 9: Red/Blue and Introduction to Data Locality Mary Hall September 21,
Feed-Forward Neural Networks 主講人 : 虞台文. Content Introduction Single-Layer Perceptron Networks Learning Rules for Single-Layer Perceptron Networks – Perceptron.
ADVANCED PERCEPTRON LEARNING David Kauchak CS 451 – Fall 2013.
PARALLEL APPLICATIONS EE 524/CS 561 Kishore Dhaveji 01/09/2000.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 8: Neural Networks.
Features of Biological Neural Networks 1)Robustness and Fault Tolerance. 2)Flexibility. 3)Ability to deal with variety of Data situations. 4)Collective.
Neural Networks in Computer Science n CS/PY 231 Lab Presentation # 1 n January 14, 2005 n Mount Union College.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
PARALLELIZATION OF ARTIFICIAL NEURAL NETWORKS Joe Bradish CS5802 Fall 2015.
Slide-1 Multicore Theory MIT Lincoln Laboratory Theory of Multicore Algorithms Jeremy Kepner and Nadya Bliss MIT Lincoln Laboratory HPEC 2008 This work.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
IE 585 History of Neural Networks & Introduction to Simple Learning Rules.
1 Perceptron as one Type of Linear Discriminants IntroductionIntroduction Design of Primitive UnitsDesign of Primitive Units PerceptronsPerceptrons.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Where are we? What’s left? HW 7 due on Wednesday Finish learning this week. Exam #4 next Monday Final Exam is a take-home handed out next Friday in class.
BA Week 10 Course overview, conversations about this class and the IS field’s future Dave Sullivan.
| presented by Vasileios Zois CS at USC 09/20/2013 Introducing Scalability into Smart Grid 1.
Clouds , Grids and Clusters
Dual-Channel Architecture
Artificial neural networks:
Real Neurons Cell structures Cell body Dendrites Axon
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
CSE P573 Applications of Artificial Intelligence Neural Networks
CSC 578 Neural Networks and Deep Learning
Closing Remarks Cyrus M. Vahid, Principal Solutions Architect,
Perceptron as one Type of Linear Discriminants
CSE 573 Introduction to Artificial Intelligence Neural Networks
Introduction to Neural Networks and Fuzzy Logic
Introduction to Neural Networks and Fuzzy Logic
CSSE463: Image Recognition Day 13
Artificial Neural Networks
Artificial Neural Networks
Introduction to Neural Network

PYTHON Deep Learning Prof. Muhammad Saeed.
Facts About High-Performance Computing
Presentation transcript:

READINGS IN DEEP LEARNING 4 Sep 2013

ADMINSTRIVIA New course numbers (11-785/786) are assigned – Should be up on the hub shortly Lab assignment 1 up – Due date: 2 weeks from today Google group: is everyone on? Website issues.. – Wordpress not yet an option (CMU CS setup) – Piazza?

Poll for next 2 classes Monday, Sep 9 – The perceptron: A probabilistic model for information storage and organization in the brain Rosenblatt Not really about the logistic perceptron, more about the probabilistic interpretation of learning in connectionist networks – Organization of behavior Donald Hebb About the Hebbian learning rule

Poll for next 2 classes Wed, Sep 11 – Optimal unsupervised learning in a single-layer linear feedforward neural network. Terence Sanger Generalized Hebbian learning rule – The Widrow Hoff learning rule Widrow and Hoff Will be presented by Pallavi Baljekar

Notices Success of course depends on good presentations Please send in your slides 1-2 days before the presentations – So that we can ensure they are OK You are encouraged to discuss your papers with us/your classmates while preparing for them – Use the google group for discussion

A new project Distributed large scale training of NNs.. Looking for volunteers

The Problem: Distributed data Training enormous networks – Billions of units from large amounts of data – Billions or Trillions of instances – Data may be localized.. – Or distributed

The problem: Distributed computing A single computer will not suffice – Need many processors – Tens or hundreds or thousands of computers Of possibly varying types and capacity

Challenge Getting the data to the computers – Tons of data to many computers Bandwidth problems Timing issues – Synchronizing the learning

Logistic Challenges How to transfer vast amounts of data to processors Which processor gets how much data.. – Not all processors equally fast – Not all data take equal amounts of time to process.. and which data – Data locality

Learning Challenges How to transfer parameters to processors – Networks are large, billions or trillions of parameters – Each processor must have the latest copy of parameters How to receive updates from processors – Each processor learns on local data – Updates from all processors must be pooled

Learning Challenges Synchronizing processor updates – Some processors slower than others – Inefficient to wait for slower ones In order to update parameters at all processors Requires asynchronous updates – Each processor updates when done – Problem: Different processors now have different set of parameters Other processors may have updated parameters already Requires algorithmic changes – How to update asynchronously – Which updates to trust

Current Solutions Faster processors GPUs – GPU programming required Large simple clusters – Simple distributed programming Large heterogeneous clusters – Techniques for asynchronous learning

Current Solutions Still assume data distribution not a major problem Assume relatively fast connectivity – Gigabit ethernet Fundamentally cluster-computing based – Local area network

New project Distributed learning Wide area network – Computers distributed across the world

New project Supervisor/Worker architecture One or more supervisors – May be a hierarchy A large number of workers Supervisors in charge of resource and task allocation, gathering and redistributing updates, synchronization

New project Challenges Data allocation – Optimal policy for data distribution Minimal latency Maximum locality

New project Challenges Computation allocation – Optimal policy for learning Compute load proportional to compute capacity Reallocation of data/task as required

New project Challenges Parameter allocation – Do we have to distribute all parameters – Can learning be local

New project Challenges Trustable updates – Different processors/LANs have different speeds – How do we trust their updates Do we incorporate or reject?

New project Optimal resychronization: how much do we transmit – Should not have to retransmit everything – Entropy coding? – Bit-level optimization?

Possibilities Massively parallel learning Never ending learning Multimodal learning GAIA..

Asking for Volunteers Will be an open source project Write to Anders

Today Bain’s theory: Lars Mahler – Linguist, mathematician, philosopher – One of the earliest people to propose connectionist architecture – Anticipated much of modern ideas McCulloch and Pitts: Kartik Goyal – Early model of neuron: Threshold gates – Earliest model to consider excitation and inhibition