Soft computing Lecture 6 Introduction to neural networks.

Slides:



Advertisements
Similar presentations
Perceptron Lecture 4.
Advertisements

Slides from: Doug Gray, David Poole
Introduction to Artificial Neural Networks
Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
Perceptron Learning Rule
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Navneet Goyal, BITS-Pilani Perceptrons. Labeled data is called Linearly Separable Data (LSD) if there is a linear decision boundary separating the classes.
Artificial Neural Networks
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Lecture 14 – Neural Networks
Neural NetworksNN 11 Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Simple Neural Nets For Pattern Classification
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Neural Networks Basic concepts ArchitectureOperation.
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
20.5 Nerual Networks Thanks: Professors Frank Hoffmann and Jiawei Han, and Russell and Norvig.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Pattern Recognition using Hebbian Learning and Floating-Gates Certain pattern recognition problems have been shown to be easily solved by Artificial neural.
An Illustrative Example
Artificial Neural Networks (ANNs)
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Data Mining with Neural Networks (HK: Chapter 7.5)
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
Soft Computing Colloquium 2 Selection of neural network, Hybrid neural networks.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Data Mining and Neural Networks Danny Leung CS157B, Spring 2006 Professor Sin-Min Lee.
Neural NetworksNN 11 Neural netwoks thanks to: Basics of neural network theory and practice for supervised and unsupervised.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
IE 585 Introduction to Neural Networks. 2 Modeling Continuum Unarticulated Wisdom Articulated Qualitative Models Theoretic (First Principles) Models Empirical.
Chapter 9 Neural Network.
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
NEURAL NETWORKS FOR DATA MINING
Lecture 3 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 3/1 Dr.-Ing. Erwin Sitompul President University
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
1 Pattern Classification X. 2 Content General Method K Nearest Neighbors Decision Trees Nerual Networks.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
So Far……  Clustering basics, necessity for clustering, Usage in various fields : engineering and industrial fields  Properties : hierarchical, flat,
Features of Biological Neural Networks 1)Robustness and Fault Tolerance. 2)Flexibility. 3)Ability to deal with variety of Data situations. 4)Collective.
Features of Biological Neural Networks 1)Robustness and Fault Tolerance. 2)Flexibility. 3)Ability to deal with variety of Data situations. 4)Collective.
© 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang 12-1 Chapter 12 Advanced Intelligent Systems.
Soft computing Lecture 7 Multi-Layer perceptrons.
Artificial Intelligence, Expert Systems, and Neural Networks Group 10 Cameron Kinard Leaundre Zeno Heath Carley Megan Wiedmaier.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Supervised learning network G.Anuradha. Learning objectives The basic networks in supervised learning Perceptron networks better than Hebb rule Single.
Artificial Neural Networks Chapter 4 Perceptron Gradient Descent Multilayer Networks Backpropagation Algorithm 1.
Perceptrons Michael J. Watts
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
CSC321: Neural Networks Lecture 1: What are neural networks? Geoffrey Hinton
1 Neural networks 2. 2 Introduction: Neural networks The nervous system contains 10^12 interconnected neurons.
“Principles of Soft Computing, 2 nd Edition” by S.N. Sivanandam & SN Deepa Copyright  2011 Wiley India Pvt. Ltd. All rights reserved. CHAPTER 2 ARTIFICIAL.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Artificial Neural Networks By: Steve Kidos. Outline Artificial Neural Networks: An Introduction Frank Rosenblatt’s Perceptron Multi-layer Perceptron Dot.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Introduction to Machine Learning, its potential usage in network area,
Learning in Neural Networks
Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang.
Chapter 12 Advanced Intelligent Systems
Chapter 3. Artificial Neural Networks - Introduction -
Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang.
Creating Data Representations
Lecture Notes for Chapter 4 Artificial Neural Networks
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Presentation transcript:

Soft computing Lecture 6 Introduction to neural networks

Disadvantages of fuzzy systems Difficulties of formalized of fuzzy sets and linguistic variables –Its description is subjective, –Its description depend on context, features of situation and inference, Fuzzy systems are keeping knowledge base systems with its main disadvantage – orientation on formalizing of knowledge by anybody (expert) and unable to learn Way out is to use neural networks instead of its or together with its

visual cortex motor cortex association cortex to motor output

neurons 3 km wires 1mm Signal: action potential (spike) action potential

Formal neuron by MacCallock-Pitts x i – binary signal

Non-binary inputs or

Geometric interpretation of TLU action

Training of TLU repeat for each training vector pair (v, t) evaluate the output y when v is input to the TLU if y ≠ t then form a new weight vector w ’ according to formulas above else do nothing end if end for until y = t for all vectors w’ i = w i + ∆w i

Perceptron of Rosenblatt The A-units can be assigned any arbitrary Boolean functionality but are fixed - they do not learn.

Classification

1. The four classes may separated by 2-hyperplanes 2. (A,B) was linearly separable from (C,D) and (A,D) was linearly separable from (B,C).

Rule of training of Widrow-Hoff w ij (t+1)=w ij (t)+x j (d i -y i ) Task of minimization of function: It was implemented in ADALINE (Adaptive Linear Elements)

Classification of models of neural networks Tutoring –Supervised learning –Unsupervised learning –Reinforcement learning Structure –Forward or recurrent networks –With regular or not links –Describes by full-links graph or no –Static or dynamic (constructive learning) Signals (inputs or outputs, hidden) –Binary –Analog Time –Discrete –Continuous Kind of giving of inputs and getting of outputs –State of synapses –State of neurons –Weights of synapses

Tasks solved by neural networks Classification –In diagnostic systems –In monitoring systems –In recognition systems of robots –In speech recognition –In security systems for authentication Clusterization –In Data Mining for extracting of knowledge –In search systems for indexing of documents Prediction –In financial analyzing –In control systems of mobile robots Approximation –In control systems of technological processes

Comparison of Computing Approaches

Comparisons of Expert Systems and Neural Networks

Advantages of neural computing Clearly the style of processing is completely dierent - it is more akin to signal processing than symbol processing. The combining of signals and producing new ones is to be contrasted with the execution of instructions stored in a memory. Information is stored in a set of weights rather than a program. The weights are supposed to adapt when the net is shown examples from a training set. Nets are robust in the presence of noise: small changes in an input signal will not drastically aect a node's output. Nets are robust in the presence of hardware failure: a change in a weight may only aect the output for a few of the possible input patterns. High level concepts will be represented as a pattern of activity across many nodes rather than as the contents of a small portion of computer memory. The net can deal with `unseen' patterns and generalise from the training set. Nets are good at `perceptual' tasks and associative recall. These are just the tasks that the symbolic approach has diculties with.