Introduction to Neural Networks CMSC475/675

Slides:



Advertisements
Similar presentations
KULIAH II JST: BASIC CONCEPTS
Advertisements

Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 안순길.
Introduction to Artificial Neural Networks
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Introduction to Neural Networks Computing
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Artificial Neural Network
Artificial Neural Network Motivation By Dr. Rezaeian Modified: Vali Derhami Yazd University, Computer Department HomePage:
Biological and Artificial Neurons Michael J. Watts
Artificial Neural Networks - Introduction -
Artificial Neural Networks - Introduction -
Neural NetworksNN 11 Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Neural Networks.
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Prof. Carla P. Gomes Module: Neural Networks: Concepts (Reading:
PERCEPTRON. Chapter 3: The Basic Neuron  The structure of the brain can be viewed as a highly interconnected network of relatively simple processing.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Rutgers CS440, Fall 2003 Neural networks Reading: Ch. 20, Sec. 5, AIMA 2 nd Ed.
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
AN INTERACTIVE TOOL FOR THE STOCK MARKET RESEARCH USING RECURSIVE NEURAL NETWORKS Master Thesis Michal Trna
1 Pendahuluan Pertemuan 1 Matakuliah: T0293/Neuro Computing Tahun: 2005.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
From Neuroinformatics to Bioinformatics: Methods for Data Analysis David Horn Spring 2006 Weizmann Institute of Science Course website:
Lecture 4 Neural Networks ICS 273A UC Irvine Instructor: Max Welling Read chapter 4.
ICS 273A UC Irvine Instructor: Max Welling Neural Networks.
Artificial Neural Network
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Artificial neural networks:
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Neural Networks Chapter 18.7
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Introduction to Neural Networks. Neural Networks in the Brain Human brain “computes” in an entirely different way from conventional digital computers.
Explorations in Neural Networks Tianhui Cai Period 3.
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
Artificial Intelligence Neural Networks ( Chapter 9 )
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
Neural Networks Kasin Prakobwaitayakit Department of Electrical Engineering Chiangmai University EE459 Neural Networks The Structure.
NEURAL NETWORKS FOR DATA MINING
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Artificial Neural Networks An Introduction. What is a Neural Network? A human Brain A porpoise brain The brain in a living creature A computer program.
Introduction to Artificial Intelligence (G51IAI) Dr Rong Qu Neural Networks.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 8: Neural Networks.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Lecture 5 Neural Control
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Artificial Neural Networks Chapter 4 Perceptron Gradient Descent Multilayer Networks Backpropagation Algorithm 1.
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
Artificial Intelligence Methods Neural Networks Lecture 1 Rakesh K. Bissoondeeal Rakesh K.
Perceptrons Michael J. Watts
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Where are we? What’s left? HW 7 due on Wednesday Finish learning this week. Exam #4 next Monday Final Exam is a take-home handed out next Friday in class.
Intro. ANN & Fuzzy Systems Lecture 3 Basic Definitions of ANN.
Lecture 12. Outline of Rule-Based Classification 1. Overview of ANN 2. Basic Feedforward ANN 3. Linear Perceptron Algorithm 4. Nonlinear and Multilayer.
Introduction to Neural Networks
Neural Networks.
Artificial Intelligence (CS 370D)
Artificial neural networks:
Real Neurons Cell structures Cell body Dendrites Axon
Joost N. Kok Universiteit Leiden
Dr. Unnikrishnan P.C. Professor, EEE
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang.
Introduction to Neural Network
David Kauchak CS158 – Spring 2019
Presentation transcript:

Introduction to Neural Networks CMSC475/675

Chapter 1 Introduction

Introduction Why ANN Some tasks can be done easily (effortlessly) by humans but are hard by conventional paradigms on Von Neumann machine with algorithmic approach Pattern recognition (old friends, hand-written characters) Content addressable recall Approximate, common sense reasoning (driving, playing piano, baseball player) These tasks are often ill-defined, experience based, hard to apply logic

Introduction Human Brain Von Neumann machine -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- One or a few high speed (ns) processors with considerable computing power One or a few shared high speed buses for communication Sequential memory access by address Problem-solving knowledge is separated from the computing component Hard to be adaptive Human Brain -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Large # (1011) of low speed processors (ms) with limited computing power Large # (1015) of low speed connections Content addressable recall (CAM) Problem-solving knowledge resides in the connectivity of neurons Adaptation by changing the connectivity

Biological neural activity Each neuron has a body, an axon, and many dendrites Can be in one of the two states: firing and rest. Neuron fires if the total incoming stimulus exceeds the threshold Synapse: thin gap between axon of one neuron and dendrite of another. Signal exchange Synaptic strength/efficiency

Introduction What is an (artificial) neural network A set of nodes (units, neurons, processing elements) Each node has input and output Each node performs a simple computation by its node function Weighted connections between nodes Connectivity gives the structure/architecture of the net What can be computed by a NN is primarily determined by the connections and their weights A very much simplified version of networks of neurons in animal nerve systems

Introduction Bio NN NN Nodes Cell body input -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Nodes input output node function Connections connection strength Bio NN -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Cell body signals from other neurons firing frequency firing mechanism Synapses synaptic strength Highly parallel, simple local computation (at neuron level) achieves global results as emerging property of the interaction (at network level) Pattern directed (meaning of individual nodes only in the context of a pattern) Fault-tolerant/graceful degrading Learning/adaptation plays important role.

History of NN Pitts & McCulloch (1943) Hebb (1949) First mathematical model of biological neurons All Boolean operations can be implemented by these neuron-like nodes (with different threshold and excitatory/inhibitory connections). Competitor to Von Neumann model for general purpose computing device in 1940’s and early 50’s Origin of automata theory. Hebb (1949) Hebbian rule of learning: increase the connection strength between neurons i and j whenever both i and j are activated. Or increase the connection strength between nodes i and j whenever both nodes are simultaneously ON or OFF.

History of NN Early booming (50’s – early 60’s) Rosenblatt (1958) x1 x2 xn Early booming (50’s – early 60’s) Rosenblatt (1958) Perceptron: network of threshold nodes for pattern classification Perceptron learning rule Percenptron convergence theorem: everything that can be represented by a perceptron can be learned Widrow and Hoff (1960, 1962) Learning rule based on gradient descent (with differentiable unit) Minsky’s attempt to build a general purpose machine with Pitts/McCullock units

History of NN The setback (mid 60’s – late 70’s) Serious problems with perceptron model (Minsky’s book 1969) Single layer perceonptrons cannot represent (learn) simple functions such as XOR Multi-layer of non-linear units may have greater power but there is no learning rule for such nets Scaling problem: connection weights may grow infinitely The first two problems overcame by latter effort in 80’s, but the scaling problem persists Death of Rosenblatt (1964) Striving of Von Neumann machine and AI

History of NN Renewed enthusiasm and flourish (80’s – present) New techniques Backpropagation learning for multi-layer feed forward nets (with non-linear, differentiable node functions) Thermodynamic models (Hopfield net, Boltzmann machine, etc.) Unsupervised learning Impressive application (character recognition, speech recognition, text-to-speech transformation, process control, associative memory, etc.) Traditional approaches face difficult challenges Caution: Don’t underestimate difficulties and limitations Poses more problems than solutions

Weighted input summation ANN Neuron Models Each node has one or more inputs from other nodes, and one output to other nodes Input/output values can be Binary {0, 1} Bipolar {-1, 1} Continuous (bounded or not) All inputs to one node come in at the same time and remain activated until the output is produced Weights associated with links General neuron model Weighted input summation

Node Function Step (threshold) function where c is called the threshold Ramp function Step function Ramp function

Node Function Sigmoid function When y = 0 and z = 0: S-shaped Continuous and everywhere differentiable Rotationally symmetric about some point (net = c) Asymptotically approaches saturation points Examples: Sigmoid function When y = 0 and z = 0: a = 0, b = 1, c = 0. When y = 0 and z = -0.5 a = -0.5, b = 0.5, c = 0. Larger x gives steeper curve

Node Function Gaussian function Bell-shaped (radial basis) Continuous f(net) asymptotically approaches 0 (or some constant) when |net| is large Single maximum (when net = ) Example: Gaussian function

Network Architecture (Asymmetric) Fully Connected Networks Every node is connected to every other node Connection may be excitatory (positive), inhibitory (negative), or irrelevant ( 0). Most general Symmetric fully connected nets: weights are symmetric (wij = wji) Input nodes: receive input from the environment Output nodes: send signals to the environment Hidden nodes: no direct interaction to the environment

Network Architecture Layered Networks Nodes are partitioned into subsets, called layers. No connections that lead from nodes in layer j to those in layer k if j > k. Inputs from the environment are applied to nodes in layer 0 (input layer). Nodes in input layer are place holders with no computation occurring (i.e., their node functions are identity function)

Network Architecture Feedforward Networks A connection is allowed from a node in layer i only to nodes in layer i + 1. Most widely used architecture. Conceptually, nodes at higher levels successively abstract features from preceding layers

Network Architecture Acyclic Networks Recurrent Networks Modular nets Connections do not form directed cycles Multi-layered feedforward nets are acyclic Recurrent Networks Nets with directed cycles Much harder to analyze than acyclic nets. Modular nets Consists of several modules, each of which is itself a neural net for a particular sub-problem Sparse connections between modules