Introduction to Neural Networks Eric Wong Martin Ho Eddy Li Kitty Wong.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Artificial Intelligence 12. Two Layer ANNs
Beyond Linear Separability
Slides from: Doug Gray, David Poole
Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 안순길.
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Tuomas Sandholm Carnegie Mellon University Computer Science Department
Kostas Kontogiannis E&CE
Artificial Neural Networks
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Prof. Carla P. Gomes Module: Neural Networks: Concepts (Reading:
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Data Mining with Neural Networks (HK: Chapter 7.5)
CS 484 – Artificial Intelligence
Neural Networks William Lai Chris Rowlett. What are Neural Networks? A type of program that is completely different from functional programming. Consists.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
Machine Learning. Learning agent Any other agent.
Backpropagation Networks
Some more Artificial Intelligence
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Networks
Data Mining and Neural Networks Danny Leung CS157B, Spring 2006 Professor Sin-Min Lee.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
IE 585 Introduction to Neural Networks. 2 Modeling Continuum Unarticulated Wisdom Articulated Qualitative Models Theoretic (First Principles) Models Empirical.
Explorations in Neural Networks Tianhui Cai Period 3.
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
NEURAL NETWORKS FOR DATA MINING
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
Classification / Regression Neural Networks 2
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Neural Networks II By Jinhwa Kim. 2 Neural Computing is a problem solving methodology that attempts to mimic how human brain function Artificial Neural.
© 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang 12-1 Chapter 12 Advanced Intelligent Systems.
Akram Bitar and Larry Manevitz Department of Computer Science
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
Introduction to Neural Networks Freek Stulp. 2 Overview Biological Background Artificial Neuron Classes of Neural Networks 1. Perceptrons 2. Multi-Layered.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Neural networks.
Neural Network Architecture Session 2
Artificial Neural Networks
Other Classification Models: Neural Network
Learning with Perceptrons and Neural Networks
Learning in Neural Networks
Artificial Intelligence (CS 370D)
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Classification / Regression Neural Networks 2
Chapter 12 Advanced Intelligent Systems
Artificial Intelligence Methods
CSE 573 Introduction to Artificial Intelligence Neural Networks
Neural Networks ICS 273A UC Irvine Instructor: Max Welling
David Kauchak CS158 – Spring 2019
Presentation transcript:

Introduction to Neural Networks Eric Wong Martin Ho Eddy Li Kitty Wong

Presentation Outline 1) Introduction to Neural Networks 2) What is Neural Networks? 3) Perceptrons 4) Introduction to Backpropagation 5) Applications of Neural Networks 6) Summary Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Background - Neural Networks can be : - Biological models - Artificial models - Desire to produce artificial systems capable of sophisticated computations similar to the human brain. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

How Does the Brain Work ? (1) NEURON - The cell that perform information processing in the brain. - Fundamental functional unit of all nervous system tissue. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

How Does the Brain Work ? (2) Each consists of : SOMA, DENDRITES, AXON, and SYNAPSE. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Brain vs. Digital Computers (1) Computers require hundreds of cycles to simulate a firing of a neuron. - The brain can fire all the neurons in a single step. Parallelism - Serial computers require billions of cycles to perform some tasks but the brain takes less than a second. e.g. Face Recognition Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Brain vs. Digital Computers (2) Future : combine parallelism of the brain with the switching speed of the computer. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Definition of Neural Network A Neural Network is a system composed of many simple processing elements operating in parallel which can acquire, store, and utilize experiential knowledge. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Neurons vs. Units (1) - Each element is a node called unit. - Units are connected by links. - Each link has a numeric weight. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Neurons vs. Units (2) Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Computing Elements A typical unit: Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Presentation Outline 1) Introduction to Neural Networks 2) What is Neural Networks? 3) Perceptrons 4) Introduction to Backpropagation 5) Applications of Neural Networks 6) Summary Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Planning in building a Neural Network Decisions on: - The number of units to use. - The type of units required. - Connection between the units. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Learning a Task - Initializing the weights. - Use of a learning algorithm. - Set of training examples. - Encode the examples as inputs. - Convert output into meaningful results. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Neural Network Example Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Simple Computations - 2 components: Linear and Non-linear. - Linear: Input function - calculate weighted sum of all inputs. - Non-linear: Activation function - transform sum into activation level. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Calculations Input function: Activation function: Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

A Computing Unit Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Activation Functions - Use different functions to obtain different models. - 3 most common choices : 1) Step function 2) Sign function 3) Sigmoid function - An output of 1 represents firing of a neuron down the axon. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

3 Activation Functions Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Units in Action - Individual units representing Boolean functions Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Network Structures Feed-forward : Links can only go in one direction. Recurrent : Links can go anywhere and form arbitrary topologies. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Feed-forward Networks - Arranged in layers. - Each unit is linked only in the unit in next layer. No units are linked between the same layer, back to the previous layer or skipping a layer. - Computations can proceed uniformly from input to output units. - No internal state exists. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Feed-Forward Example H3 H5 I1 O7 I2 H4 H6 t = -0.5 t = 1.5 W57 = 1 I1 W25 = 1 O7 t = 0.5 W16 = 1 I2 W67 = 1 W24= -1 t = -0.5 t = 1.5 W46 = 1 H4 H6 Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Multi-layer Networks and Perceptrons - Have one or more layers of hidden units. - With two possibly very large hidden layers, it is possible to implement any function. - Networks without hidden layer are called perceptrons. - Perceptrons are very limited in what they can represent, but this makes their learning problem much simpler. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Recurrent Network (1) - The brain is not and cannot be a feed-forward network. - Allows activation to be fed back to the previous unit. - Internal state is stored in its activation level. - Can become unstable or oscillate. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Recurrent Network (2) - May take long time to compute a stable output. - Learning process is much more difficult. - Can implement more complex designs. - Can model certain systems with internal states. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Presentation Outline 1) Introduction to Neural Networks 2) What is Neural Networks? 3) Perceptrons 4) Introduction to Backpropagation 5) Applications of Neural Networks 6) Summary Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Perceptrons - First studied in the late 1950s. - Also known as Layered Feed-Forward Networks. - The only efficient learning element at that time was for single-layered networks. - Today, used as a synonym for a single-layer, feed-forward network. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Perceptrons Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

What can Perceptrons Represent ? - Some complex Boolean function can be represented. For example: Majority function - will be covered in the next presentation. - Limited in the Boolean functions they can represent. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Linear Separability in Perceptrons Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Learning Linearly Separable Functions (1) What can these functions learn ? Bad news: - There are not many linearly separable functions. Good news: - There is a perceptron algorithm that will learn any linearly separable function, given enough training examples. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Learning Linearly Separable Functions (2) Most neural network learning algorithms, including the perceptrons learning method, follow the current-best- hypothesis (CBH) scheme. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Learning Linearly Separable Functions (3) - Initial network has a randomly assigned weights. - Done by making small adjustments in the weights to reduce the difference between the observed and predicted values. - Main difference from the logical algorithms is need to repeat the update phase several times in order to achieve convergence. - Updating process is divided into epochs, each epoch updates all the weights of the process. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Neural-Network-Learning Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Presentation Outline 1) Introduction to Neural Networks 2) What is Neural Networks? 3) Perceptrons 4) Introduction to Backpropagation 5) Applications of Neural Networks 6) Summary Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Introduction to Backpropagation - In 1969 a method for learning in multi-layer network, Backpropagation, was invented by Bryson and Ho. - The Backpropagation algorithm is a sensible approach for dividing the contribution of each weight. - Works basically the same as perceptrons Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Backpropagation Learning There are two differences for the updating rule : 1) The activation of the hidden unit is used instead of the input value. 2) The rule contains a term for the gradient of the activation function. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Backpropagation Algorithm(1) Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Backpropagation Algorithm(2) The ideas of the algorithm can be summarized as follows : - Computes the error term for the output units using the observed error. - From output layer, repeat propagating the error term back to the previous layer and updating the weights between the two layers until the earliest hidden layer is reached. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Examples of Backpropagation Learning Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Examples of Feed-Forward Learning Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Examples of Backpropagation Learning Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Presentation Outline 1) Introduction to Neural Networks 2) What is Neural Networks? 3) Perceptrons 4) Introduction to Backpropagation 5) Applications of Neural Networks 6) Summary Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Neural Network Applications - May provide a model for massive parallel computation. - More successful approach of “parallelizing” traditional serial algorithms. - Can compute any computable function. - Can do everything a normal digital computer can do. - Can do even more under some impractical assumptions. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Neural Network in Practice For classification and function approximation or mapping problems which are: - Tolerant of some imprecision. - Have lots of training data available. - Hard and fast rules cannot easily be applied. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Neural Network Approaches ALVINN - Autonomous Land Vehicle In a Neural Network Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Neural Network Approaches - Developed in 1993. - Performs driving with Neural Networks. - An intelligent VLSI image sensor for road following. - Learns to filter out image details not relevant to driving. Output units Hidden layer Input units Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Neural Network Approaches Input Array Hidden Units Output units Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Actual Products Available Enterprise Miner: - Single multi-layered feed-forward neural networks. - Provides business solutions for data mining. Nestor: - Uses Nestor Learning System (NLS). - Several multi-layered feed-forward neural networks. - Intel has made such a chip - NE1000 for VLSI. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Enterprise Miner - Based on SEMMA (Sample, Explore, Modify, Model, Access) methodology. - Statistical tools include : Clustering, decision trees, linear and logistic regression and neural networks. - Data preparation tools include : Outliner detection, variable transformation, random sampling, and partition of data sets (into training, testing and validation data sets). Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Nestor - With low connectivity within each layer. - Minimized connectivity within each layer results in rapid training and efficient memory utilization, ideal for VLSI. - Composed of multiple neural networks, each specializing in a subset of information about the input patterns. - Real time operation without the need of special computers or custom hardware. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Presentation Outline 1) Introduction to Neural Networks 2) What is Neural Networks? 3) Perceptrons 4) Introduction to Backpropagation 5) Applications of Neural Networks 6) Summary Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Summary - Neural network is a computational model that simulate some properties of the human brain. - The connections and nature of units determine the behavior of a neural network. - Perceptrons are feed-forward networks that can only represent linearly separable functions. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

Summary - Given enough units, any function can be represented by Multi-layer feed-forward networks. - Backpropagation learning works on multi-layer feed-forward networks. - Neural Networks are widely used in developing artificial learning systems. Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000

References - Russel, S. and P. Norvig (1995). Artificial Intelligence - A Modern Approach. Upper Saddle River, NJ, Prentice Hall. - Sarle, W.S., ed. (1997), Neural Network FAQ, part 1 of 7: Introduction, periodic posting to the Usenet newsgroup comp.ai.neural-nets, URL: ftp://ftp.sas.com/pub/neural/FAQ.html Presented by Martin Ho, Eddy Li, Eric Wong and Kitty Wong - Copyright© 2000