So Far……  Clustering basics, necessity for clustering, Usage in various fields : engineering and industrial fields  Properties : hierarchical, flat,

Slides:



Advertisements
Similar presentations
Artificial Intelligence 12. Two Layer ANNs
Advertisements

Perceptron Lecture 4.
Beyond Linear Separability
Slides from: Doug Gray, David Poole
Artificial Neural Networks (1)
Perceptron Learning Rule
NEURAL NETWORKS Perceptron
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Artificial neural networks:
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Artificial Neural Networks
Simple Neural Nets For Pattern Classification
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Prof. Carla P. Gomes Module: Neural Networks: Concepts (Reading:
20.5 Nerual Networks Thanks: Professors Frank Hoffmann and Jiawei Han, and Russell and Norvig.
November 9, 2010Neural Networks Lecture 16: Counterpropagation 1 Unsupervised Learning So far, we have only looked at supervised learning, in which an.
September 14, 2010Neural Networks Lecture 3: Models of Neurons and Neural Networks 1 Visual Illusions demonstrate how we perceive an “interpreted version”
An Illustrative Example
August 12, 2003 IV. FUZZY SET METHODS - CLUSTER ANALYSIS: Math Clinic Fall IV. FUZZY SET METHODS for CLUSTER ANALYSIS and (super brief) NEURAL NETWORKS.
Data Mining with Neural Networks (HK: Chapter 7.5)
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Aula 4 Radial Basis Function Networks
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Radial Basis Function (RBF) Networks
Radial-Basis Function Networks
Radial Basis Function Networks
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
Radial Basis Function Networks
Neurons, Neural Networks, and Learning 1. Human brain contains a massively interconnected net of (10 billion) neurons (cortical cells) Biological.
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
1 Chapter 6: Artificial Neural Networks Part 2 of 3 (Sections 6.4 – 6.6) Asst. Prof. Dr. Sukanya Pongsuparb Dr. Srisupa Palakvangsa Na Ayudhya Dr. Benjarath.
Mestrado em Ciência de Computadores Mestrado Integrado em Engenharia de Redes e Sistemas Informáticos VC 14/15 – TP19 Neural Networks & SVMs Miguel Tavares.
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
NEURAL NETWORKS FOR DATA MINING
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
1 Pattern Classification X. 2 Content General Method K Nearest Neighbors Decision Trees Nerual Networks.
Neural Networks and Machine Learning Applications CSC 563 Prof. Mohamed Batouche Computer Science Department CCIS – King Saud University Riyadh, Saudi.
Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 29 Nov 11, 2005 Nanjing University of Science & Technology.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Supervised learning network G.Anuradha. Learning objectives The basic networks in supervised learning Perceptron networks better than Hebb rule Single.
November 21, 2013Computer Vision Lecture 14: Object Recognition II 1 Statistical Pattern Recognition The formal description consists of relevant numerical.
NEURAL NETWORKS LECTURE 1 dr Zoran Ševarac FON, 2015.
1 Learning Bias & Clustering Louis Oliphant CS based on slides by Burr H. Settles.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Linear Models & Clustering Presented by Kwak, Nam-ju 1.
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
March 31, 2016Introduction to Artificial Intelligence Lecture 16: Neural Network Paradigms I 1 … let us move on to… Artificial Neural Networks.
Artificial Neural Networks By: Steve Kidos. Outline Artificial Neural Networks: An Introduction Frank Rosenblatt’s Perceptron Multi-layer Perceptron Dot.
CLUSTERING EE Class Presentation. TOPICS  Clustering basic and types  K-means, a type of Unsupervised clustering  Supervised clustering type.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Today’s Lecture Neural networks Training
What is an ANN ? The inventor of the first neuro computer, Dr. Robert defines a neural network as,A human brain like system consisting of a large number.
Machine Learning. Support Vector Machines A Support Vector Machine (SVM) can be imagined as a surface that creates a boundary between points of data.
Data Mining with Neural Networks (HK: Chapter 7.5)
Neuro-Computing Lecture 4 Radial Basis Function Network
of the Artificial Neural Networks.
Machine Learning. Support Vector Machines A Support Vector Machine (SVM) can be imagined as a surface that creates a boundary between points of data.
Machine Learning. Support Vector Machines A Support Vector Machine (SVM) can be imagined as a surface that creates a boundary between points of data.
Capabilities of Threshold Neurons
Artificial Intelligence Chapter 3 Neural Networks
Computer Vision Lecture 19: Object Recognition III
Sanguthevar Rajasekaran University of Connecticut
Presentation transcript:

So Far……  Clustering basics, necessity for clustering, Usage in various fields : engineering and industrial fields  Properties : hierarchical, flat, iterative, hard, soft, disjunctive  Types: Supervised and unsupervised  K-means algorithm – for unsupervised clustering  Vector quantization – for supervised clustering  CONTINUE with fuzzy and neural net algorithms – supervised clustering

Topics for today  Basics of fuzzy systems, and fields of application  Residual analysis methods  Fuzzy C-means algorithm – Matlab illustration  Artificial neural network – basics  Structure and working of artificial neuron  Hyper plane analysis of the output

FUZZY LOGIC SYSTEMS IN SUPERVISED CLUSTERING

Fuzzy Systems  Last decade increase in Fuzzy system implementation  More popular in the field of control systems and pattern recognition  Consumer products (washing machine,camcorders, palm pilot…) and industrial systems (to provide decision support and expert system with powerful reasoning capabilities bound by a minimum of rules)  Classical set (non-fuzzy) – either belongs to or does not belong to the set (crisp membership)  Fuzzy set – allows degree of membership for each element to range over the unit interval [0, 1]

 Fuzzy membership represent similarities of objects to imprecisely defined properties whereas, probabilities convey information about relative frequencies  Probability : some kind of likelihood or degree of certainty or if it the outcome of clearly defined but randomly occurring events  Major feature of fuzzy : expresses the amount of ambiguity in human thinking  When to use fuzzy logic ?  Continuous phenomenon, not easily breakable into discrete segments  Cannot model a process

Residual Analysis  Fuzzy reasoning :IF-THEN reasoning based on the sign of the residual Ex: IF residual-1 is positive and residual-2 is negative THEN fault1 is Present IF residual-1 is zero and residual-2 is zero THEN system is fault free :  Fuzzy Clustering : each data point belongs to all classes with a certain degree of membership. The degree is dependant upon the distance to all cluster centers. For fault diagnosis, each class could correspond to a particular fault

(1) Fuzzy centers computed by minimizing the following partition formula: Fuzzy C-means algorithm subject to C - Number of clusters N - Number of data points - fuzzy membership of the k-th point to the i-th cluster - Euclidean distance between the data point and the cluster center fuzzy weighting factor which defines the degree of fuzziness of the results (normally chosen m =2, to get analytical solution) -

(2)The cluster centers v, (centroids or prototypes) are defined as the fuzzy weighted center of gravity of the data, (3)The minimization of the partition functional (1) will give the following expression for the membership (4) Euclidean distance defined as :

Two steps of Fuzzy Clustering  Off-line phase: Learning phase, determines cluster centers of the classes (this is done by iteratively calculating membership function). A learning data set is necessary, which must contain residuals for all known faults.  On-line phase: Calculates the membership degree of the current residuals to each of the known classes.

1. Choose the number of classes C, ; Chose m=2, Initialize ( start with some arbitrary values for cluster centers and corresponding partition matrix values ) 2. Calculate the cluster centers using Eq. in (2) 3. Calculate new partition matrix using Eq. in (3) 4. Compare and. If the variation of the membership degree, calculated with an appropriate norm, is smaller than a given threshold, stop the algorithm, otherwise go back to step 2. Off-line phase Cluster centers determination

Matlab – fuzzy logic tool box – illustration of fuzzy C-means algorithm

 For the incoming data, calculate the degree of membership to all the centers using the following formula On-line phase - fuzzy membership of the k-th point to the i-th cluster - Euclidean distance between the data point and the cluster center

Artificial Neural Net systems

 Based on low level microscopic biological models  Originated from modeling of human brain and evolution  Collective behavior of NN, like a human brain, demonstrates the ability to learn, recall and generalize from training patterns of data  Consists of large number of highly interconnected processing elements (nodes)  Application areas: speech recognition, speech to text conversion, image processing, investing, trading……

 Model specified by three basic elements: Models of the processing element Models of interconnections and structures (network topology) Learning rules (the way information is stored in the network)  Each node collects values from all its input connections, performs a predefined mathematical operation and produces a single output  Net input to the node = integration function (typically dot product) combining all inputs  Each input connection is weighed. These weights (can be positive or negative values) are determined during learning process. Because of this adjustments, NN is able to learn.  Activation output of node = activation function ( net input ), usually a non-linear function

Ex: Activation function Unipolar ( output takes the value 0 or 1 ) Output y = 0, if sum < threshold (b) = 1,if sum > threshold (b) sum is dot product of input vector x= ( ) and weight vector w = ( )

Some activation functions (unipolar and bipolar)

 Input vector x = ( )  Linearly combined with weights  Then s is activated by a threshold function T(-) to produce the output y = T(s) = 1 when s > 0, else y = T(s) = -1.  Then all the input vectors x such that forms a Hyperplane H in the input vector space. H partitions the feature vector space into right and left half spaces, H+ (when sum is > b) and H- (when sum < b) The Perceptron as Hyperplane separator, where b is the threshold

Ex : consider a single perceptron with two inputs Let w1 = 2 andw2 = -1, b=0, then 2x1 - x2 = 0 determines H the points (0,0) and (1,2) belong to H The feature vector x = (x1,x2) = (2,3) is summed into S = 2(2) - 1(3) = 1 > 0, so that the activated output is y = T(1) = 1 (corresponds to H+ in the plane, i.e right half) (x1,x2) = (0,2) activates the output y = T(2(0) - 1(2)) = T(-1) = -1, which indicates that (0,2) is in the left halfspace H-. The figure shows these points.

Mapping in Hyperplane. (example of linear mapping between input and output

Non-linear mapping between input and output Example : XOR logic function or 2- bit parity problem. N = 2 inputs, M = 1 output, and Q = 4 sample vector (input/output) pairs for training, and K= 2 clusters (even and odd).

Hyperplane diagram for 2-bit parity problem

XOR function implementation in three layered network Take: result is two parallel hyperplanes that yield three convex regions. The hyperplanes are determined by

The threshold at the first neuron in the hidden layer yields The threshold at the second hidden neuron yields

hyperplanes yield three convex regions. The four sets of above inputs yield the three unique vectors Corresponding to three regions in the hyperplane

Hyperplanes showing three regions for 2-bit parity problem Regions 1 and 3make up the odd parity (Class 2),while Region 3 is even parity (Class 1).

For the second layer (output layer), the equations are as follows: Choosing Threshold = 1/2