Seoul National University Neural Network Modeling for Intelligent Novelty Detection 제 2 차 뇌신경정보학 Workshop 일시 : 2002 년 2 월 27 일 ( 수 ) 10:00-18:00 장소 : KAIST.

Slides:



Advertisements
Similar presentations
Introduction to Support Vector Machines (SVM)
Advertisements

1 Image Classification MSc Image Processing Assignment March 2003.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Support Vector Machines
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
An Overview of Machine Learning
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
The Perceptron CS/CMPE 333 – Neural Networks. CS/CMPE Neural Networks (Sp 2002/2003) - Asim LUMS2 The Perceptron – Basics Simplest and one.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 15: Introduction to Artificial Neural Networks Martin Russell.
November 2, 2010Neural Networks Lecture 14: Radial Basis Functions 1 Cascade Correlation Weights to each new hidden node are trained to maximize the covariance.
Greg GrudicIntro AI1 Introduction to Artificial Intelligence CSCI 3202: The Perceptron Algorithm Greg Grudic.
Optimal Adaptation for Statistical Classifiers Xiao Li.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Aula 4 Radial Basis Function Networks
Image Compression Using Neural Networks Vishal Agrawal (Y6541) Nandan Dubey (Y6279)
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks II PROF. DR. YUSUF OYSAL.
Radial Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
An Introduction to Support Vector Machines Martin Law.
Where We’re At Three learning rules  Hebbian learning regression  LMS (delta rule) regression  Perceptron classification.
Radial Basis Function Networks
1 Introduction to Artificial Neural Networks Andrew L. Nelson Visiting Research Faculty University of South Florida.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Multi-Layer Perceptrons Michael J. Watts
Introduction to Neural Networks Debrup Chakraborty Pattern Recognition and Machine Learning 2006.
Chapter 9 Neural Network.
Appendix B: An Example of Back-propagation algorithm
Matlab Matlab Sigmoid Sigmoid Perceptron Perceptron Linear Linear Training Training Small, Round Blue-Cell Tumor Classification Example Small, Round Blue-Cell.
Machine Learning Lecture 11 Summary G53MLE | Machine Learning | Dr Guoping Qiu1.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
Classification / Regression Neural Networks 2
Artificial Intelligence Techniques Multilayer Perceptrons.
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Applying Neural Networks Michael J. Watts
An Introduction to Support Vector Machines (M. Law)
Digital Media Lab 1 Data Mining Applied To Fault Detection Shinho Jeong Jaewon Shim Hyunsoo Lee {cinooco, poohut,
Multi-Layer Perceptron
Non-Bayes classifiers. Linear discriminants, neural networks.
CSSE463: Image Recognition Day 14 Lab due Weds, 3:25. Lab due Weds, 3:25. My solutions assume that you don't threshold the shapes.ppt image. My solutions.
CS621 : Artificial Intelligence
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Neural Networks Demystified by Louise Francis Francis Analytics and Actuarial Data Mining, Inc.
Lecture 5 Neural Control
Speech Lab, ECE, State University of New York at Binghamton  Classification accuracies of neural network (left) and MXL (right) classifiers with various.
Chapter 8: Adaptive Networks
Support-Vector Networks C Cortes and V Vapnik (Tue) Computational Models of Intelligence Joon Shik Kim.
Neural Networks 2nd Edition Simon Haykin
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Web-Mining Agents Prof. Dr. Ralf Möller Universität zu Lübeck Institut für Informationssysteme Tanya Braun (Übungen)
Landmark-Based Speech Recognition: Spectrogram Reading, Support Vector Machines, Dynamic Bayesian Networks, and Phonology Mark Hasegawa-Johnson
SUPPORT VECTOR MACHINES Presented by: Naman Fatehpuria Sumana Venkatesh.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Neural network based hybrid computing model for wind speed prediction K. Gnana Sheela, S.N. Deepa Neurocomputing Volume 122, 25 December 2013, Pages 425–429.
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Neural networks.
CSSE463: Image Recognition Day 14
CS 9633 Machine Learning Support Vector Machines
Deep Feedforward Networks
CSE 473 Introduction to Artificial Intelligence Neural Networks
Self organizing networks
Neuro-Computing Lecture 4 Radial Basis Function Network
Multi-Layer Perceptron
Introduction to Radial Basis Function Networks
What is Artificial Intelligence?
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Seoul National University Neural Network Modeling for Intelligent Novelty Detection 제 2 차 뇌신경정보학 Workshop 일시 : 2002 년 2 월 27 일 ( 수 ) 10:00-18:00 장소 : KAIST LG Semicon Hall 1 층 강당

제 2 차 뇌신경 정보학 Workshop Introduction Concept of Novelty Detection – Typical Classification Class A Class B Classifier A New Instance Class A or Class B Training Classification Class A Novelty Detector A New Instance Class A or NOT Training Classification – Novelty Detection

제 2 차 뇌신경 정보학 Workshop Introduction Applications of Novelty Detection – Authentication of computer system – Detection of counterfeit – Detection of phase transformation in a financial market – Fault detection in a mechanical system Algorithms for Novelty Detection – Principal Component Analysis (PCA) – Auto-Associative Multi-Layer Perceptron (AAMLP) – Probability Density Estimator : SOM, Mixture of Kernels, etc.

제 2 차 뇌신경 정보학 Workshop Characteristics of 2-Layer AAMLP Structure – – Auto-Association  Input vectors = Target vectors  Normal patterns  Smaller errors Novel patterns  Larger errors – Training Algorithms  Same as a typical MLP  Back-propagation,  Gradient-descent,  Levenberg-Marquadt, etc. Hidden Layer Output Layer nodesMLP with

제 2 차 뇌신경 정보학 Workshop Characteristics of 2-Layer AAMLP Properties (Lee, Hwang, Cho; submitted) – A 2-layer AAMLP defines an output-constrained hyperplane – The hyperplane is bounded – The hyperplane lies within the training input vector area – It is trained so that the association error, the sum of distances between the hyperplane and the input vectors, is minimized

제 2 차 뇌신경 정보학 Workshop Experiments with 2-Layer AAMLP Comparison with 2-Layer AAMLP (Linear Ftn.) (=PCA) – AAMLP with Linear Ftn. – AAMLP with Linear Ftn.

제 2 차 뇌신경 정보학 Workshop Experiments with 2-Layer AAMLP Comparison with 2-Layer AAMLP (Linear Ftn.) (=PCA) – AAMLP with Linear Ftn.

제 2 차 뇌신경 정보학 Workshop Experiments with 2-Layer AAMLP 2-Layer AAMLP with Saturated Linear Ftn. Equivalent to 2-layer AAMLP with sigmoid ftn.

제 2 차 뇌신경 정보학 Workshop Experiments with 2-Layer AAMLP Limitations of 2-Layer AAMLP Inability to model a non-linear data

제 2 차 뇌신경 정보학 Workshop Characteristics of 4-Layer AAMLP 4-Layer AAMLP (Kramer, 1991) – – Non-Linear PCA – Limitations of NLPCA (Malthouse, 1998)  Good at interpolation  But not at extrapolation  A desirable property as a novelty detector Mapping Layer De-Mapping Layer Bottleneck Layer Output Layer

제 2 차 뇌신경 정보학 Workshop Characteristics of 4-Layer AAMLP Novelty Detection in A Non-Linear Data – 4-Layer AAMLP with Sigmoid Ftn. Ability to model a non-linear data

제 2 차 뇌신경 정보학 Workshop Characteristics of 4-Layer AAMLP Novelty Detection in A Non-Linear Data – 4-Layer AAMLP with Saturated Linear Ftn. Ability to model a non-linear data

제 2 차 뇌신경 정보학 Workshop Characteristics of 4-Layer AAMLP Novelty Detection in A Non-Linear Data – Similarity of Sigmoid and Saturated Linear Function – AAMLP with Sigmoid and Saturated Linear Function Assumption AAMLP-SatLinis similar to 2-layer AAMLP in the output characteristics &is similar to 4-layer AAMLP in the shape of output vectors Hypothesis AAMLP-Sigmoiddefines an output-constrained hypersurface &can minimizes the association error, at least, locally

제 2 차 뇌신경 정보학 Workshop Characteristics of 4-Layer AAMLP Novelty Detection in A Multi-Modal Data – 4-Layer AAMLP with Sigmoid Ftn. Inability to model multi-modal data

제 2 차 뇌신경 정보학 Workshop Characteristics of 4-Layer AAMLP Novelty Detection in A Multi-Modal Data – Clustering + 4-Layer AAMLP with Sigmoid Ftn. Ability to model a multi-modal data

제 2 차 뇌신경 정보학 Workshop Characteristics of 4-Layer AAMLP Properties – A 4-layer AAMLP with sigmoid ftn. defines a hypersurface – A one with saturated linear ftn. defines a set of segments – A one with saturated linear ftn. can model a non-linear data in a way similar to that of 2-layer AAMLP – It can be argued that a one with sigmoid ftn. can model a non-linear data & minimizes the association error – A one with sigmoid ftn. cannot model a multi-modal data – A one with sigmoid ftn. can model a multi-modal data if it is preceded by proper clustering algorithms

제 2 차 뇌신경 정보학 Workshop Conclusion – A 2-layer AAMLP can be a novelty detector for a linear data – A 4-layer one can be a novelty detector for some non-linear data, which a 4-layer one cannot – A 4-layer one can be a novelty detector for a multi-modal data, if the data is pre-clustered

제 2 차 뇌신경 정보학 Workshop Future Works – Probability Density Estimators : SOM, Mixture of Kernels, etc.  Comparison of AAMLP and the density estimators in performances  Sophisticated clustering methods using the density estimators – AAMLP Ensemble  The unstableness of AAMLP  An AAMLP ensemble may overcome  Bagging & Majority voting