Neural Computing Group Department of Computing University of Surrey In-situ Learning in Multi-net Systems 25 th August 2004

Slides:



Advertisements
Similar presentations
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Advertisements

Computer Science Department FMIPA IPB 2003 Neural Computing Yeni Herdiyeni Computer Science Dept. FMIPA IPB.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
School of Cybernetics, School of Systems Engineering, University of Reading Presentation Skills Workshop March 22, ‘11 Diagnosis of Breast Cancer by Modular.
POSTER TEMPLATE BY: Multi-Sensor Health Diagnosis Using Deep Belief Network Based State Classification Prasanna Tamilselvan.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Integrated Learning in Multi-net Systems Neural Computing Group Department of Computing University of Surrey Matthew Casey 6 th February 2004
You wanted to know what the Matrix is? Matthew Casey.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Three kinds of learning
L ++ An Ensemble of Classifiers Approach for the Missing Feature Problem Using learn ++ IEEE Region 2 Student Paper Contest University of Maryland Eastern.
1 Combining Multiple Modes of Information using Unsupervised Neural Classifiers Neural Computing Group, Department.
Who am I and what am I doing here? Allan Tucker A brief introduction to my research
Can We Count on Neural Networks? Modelling the ‘Action of the Brain’ Matthew Casey
Handwritten Character Recognition Using Block wise Segmentation Technique (BST) in Neural Network 47th Annual Convention of the Computer Society of India.
Statistical Learning: Pattern Classification, Prediction, and Control Peter Bartlett August 2002, UC Berkeley CIS.
Neural Computing Group Department of Computing University of Surrey Matthew Casey 21 st April 2004
JYC: CSM17 Bioinformatics Week 9: Simulations #3: Neural Networks biological neurons natural neural networks artificial neural networks applications.
Rotation Forest: A New Classifier Ensemble Method 交通大學 電子所 蕭晴駿 Juan J. Rodríguez and Ludmila I. Kuncheva.
Information Fusion Yu Cai. Research Article “Comparative Analysis of Some Neural Network Architectures for Data Fusion”, Authors: Juan Cires, PA Romo,
Breast Cancer Diagnosis A discussion of methods Meena Vairavan.
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Are we still talking about diversity in classifier ensembles? Ludmila I Kuncheva School of Computer Science Bangor University, UK.
Intrusion Detection Using Hybrid Neural Networks Vishal Sevani ( )
Multi Layer NN and Bit-True Modeling of These Networks SILab presentation Ali Ahmadi September 2007.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Full model selection with heuristic search: a first approach with PSO Hugo Jair Escalante Computer Science Department, Instituto Nacional de Astrofísica,
Parallel Artificial Neural Networks Ian Wesley-Smith Frameworks Division Center for Computation and Technology Louisiana State University
Introduction to Neural Networks Debrup Chakraborty Pattern Recognition and Machine Learning 2006.
Chapter 9 Neural Network.
NEURAL NETWORKS FOR DATA MINING
ACTIVE LEARNING USING CONFORMAL PREDICTORS: APPLICATION TO IMAGE CLASSIFICATION HypHyp Introduction HypHyp Conceptual overview HypHyp Experiments and results.
Artificial Intelligence Techniques Multilayer Perceptrons.
Data Dependence in Combining Classifiers Mohamed Kamel PAMI Lab University of Waterloo.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Lecture 10: 8/6/1435 Machine Learning Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Machine Learning Using Support Vector Machines (Paper Review) Presented to: Prof. Dr. Mohamed Batouche Prepared By: Asma B. Al-Saleh Amani A. Al-Ajlan.
Ensemble Methods: Bagging and Boosting
Classifier ensembles: Does the combination rule matter? Ludmila Kuncheva School of Computer Science Bangor University, UK
Artificial Neural Networks Bruno Angeles McGill University – Schulich School of Music MUMT-621 Fall 2009.
Cellular Automata Machine For Pattern Recognition Pradipta Maji 1 Niloy Ganguly 2 Sourav Saha 1 Anup K Roy 1 P Pal Chaudhuri 1 1 Department of Computer.
Neural Networks Steven Le. Overview Introduction Architectures Learning Techniques Advantages Applications.
Behavioral Fault Model for Neural Networks A. Ahmadi, S. M. Fakhraie, and C. Lucas Silicon Intelligence and VLSI Signal Processing Laboratory, School of.
AUTOMATIC TARGET RECOGNITION AND DATA FUSION March 9 th, 2004 Bala Lakshminarayanan.
Ivica Dimitrovski 1, Dragi Kocev 2, Suzana Loskovska 1, Sašo Džeroski 2 1 Faculty of Electrical Engineering and Information Technologies, Department of.
Ping-Tsun Chang Intelligent Systems Laboratory Computer Science and Information Engineering National Taiwan University Combining Unsupervised Feature Selection.
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
By Akhilesh K. Sinha Nishant Singh Supervised by Prof. Amitabha Mukerjee Video Surveillance of Basketball Matches and Goal Detection Indian Institute of.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Artificial Neural Networks By: Steve Kidos. Outline Artificial Neural Networks: An Introduction Frank Rosenblatt’s Perceptron Multi-layer Perceptron Dot.
NEURAL NETWORKS. An extremely simplified model of the brain !
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
A presentation to El Paso del Norte Software Association
Deep Learning Amin Sobhani.
A Pool of Deep Models for Event Recognition
Restricted Boltzmann Machines for Classification
Face Recognition using Artificial Neural Network
Neural Networks Dr. Peter Phillips.
What is an ANN ? The inventor of the first neuro computer, Dr. Robert defines a neural network as,A human brain like system consisting of a large number.
COMP61011 : Machine Learning Ensemble Models
Artificial Neural Networks
What is Pattern Recognition?
Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang.
Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang.
network of simple neuron-like computing elements
IEEE World Conference on Computational Intelligence – WCCI 2010
Evolutionary Ensembles with Negative Correlation Learning
Presentation transcript:

Neural Computing Group Department of Computing University of Surrey In-situ Learning in Multi-net Systems 25 th August Matthew Casey and Khurshid Ahmad

2 Classification and Multi-nets Multiple classifiers: –Can give improved classification [6, 16] –Ensemble: redundancy –Modular: no redundancy, simpler classifiers Multi-net systems [12]: –Systems consisting of multiple neural networks –Parallel and sequential combinations –Mixture-of-experts (ME) [5]: modular, in-situ

3 In-situ Learning ME learns how to combine networks in parallel –Simultaneous training of networks and combination Existing techniques focus on pre-training –Ensembles Can in-situ learning in different types of multi-net system improve performance [3]? –Ensemble: Simple learning ensemble (SLE) –Modular: Sequential learning modules (SLM)

4 Simple Learning Ensemble (Pre-trained) ensembles: –Train each network individually –Parallel combination of outputs: mean/majority vote/etc –How can we assess the combined performance? –Hybrid pre- and in-situ: adaptive algorithm [14] –In-situ: negative correlation learning (NCL) [8] SLE: –Based on simple ensemble (SE): mean output –Train in-situ and assess combined performance –Generalisation loss early stopping criterion [11]

5 Sequential Learning Modules Modular systems: –Combining simpler networks –ME example of parallel combination –Sequential systems typically pre-trained –Limited examples of sequential algorithms [2, 17] SLM: –Can a sequential system improve performance? –How can in-situ learning be used sequentially? –Two networks: first unsupervised, second supervised –No backward propagation of error required

6 Experiments Compare existing pre-trained: –Single-net: MLP –Ensemble: SE with from 2 to 20 networks To in-situ trained: –Ensemble: SLE with from 2 to 20 networks –Modular: SLM 100 trials each with random initial weights Classification data sets [1] –MONK’s problems [13] –Wisconsin Breast Cancer Database (WBCD) [15]

7 Experimental Results

8 SLE: –Improves upon MLP (ES), SE (ES) for all –Improves upon MLP in MONK 1 and 2 –In-situ learning and early stopping improves on pre- training SLM: –Improves on MLP (ES), SE (ES) for MONK 1 –Improves upon all for MONK 2, 3 and WBCD –Successfully learnt and generalised: non-linear –Compares well to AdaBoost [4]: 97.6%

9 Conclusions Encouraging results for in-situ learning Ensemble in-situ learning: –How does in-situ learning affect diversity? –What is the relationship between ensembles and partially connected feedforward networks? Sequential learning: –Reinforces use of modular systems –Complexity still needs to be addressed –Can we understand the properties of such systems (parallel or sequential) to exploit them?

10 Questions?

11 References 1. Blake,C.L. & Merz,C.J.UCI Repository of Machine Learning Databases. Irvine, CA.: University of California, Irvine, Department of Information and Computer Sciences, Bottou, L. & Gallinari, P. A Framework for the Cooperation of Learning Algorithms. In Lippmann, R.P., Moody, J.E. & Touretzky, D.S. (Ed), Advances in Neural Information Processing Systems, vol. 3, pp , Casey, M.C. Integrated Learning in Multi-net Systems. Unpublished doctoral thesis. Guildford, UK: University of Surrey, Drucker, H. Boosting Using Neural Networks. In Sharkey, A. J. C. (Ed), Combining Artificial Neural Nets: Ensemble and Modular Multi-Net Systems, pp London: Springer-Verlag, Jacobs, R.A. & Tanner, M. Mixtures of X. In Sharkey, A. J. C. (Ed), Combining Artificial Neural Nets: Ensemble and Modular Multi-Net Systems, pp Berlin, Heidelberg, New York: Springer-Verlag, 1999.

12 References 6. Kittler, J., Hatef, M., Duin, R.P.W. & Matas, J. On Combining Classifiers. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 20(3), pp , Kohonen, T. Self-Organized Formation of Topologically Correct Feature Maps. Biological Cybernetics, vol. 43, pp , Liu, Y. & Yao, X. Ensemble Learning via Negative Correlation. Neural Networks, vol. 12(10), pp , Liu, Y., Yao, X., Zhao, Q. & Higuchi, T. An Experimental Comparison of Neural Network Ensemble Learning Methods on Decision Boundaries. Proceedings of the 2002 International Joint Conference on Neural Networks (IJCNN'02), vol. 1, pp Los Alamitos, CA.: IEEE Computer Society Press, Partridge, D. & Griffith, N. Multiple Classifier Systems: Software Engineered, Automatically Modular Leading to a Taxonomic Overview. Pattern Analysis and Applications, vol. 5(2), pp , 2002.

13 References 11. Prechelt, L. Early Stopping - But When? In Orr, G. B. & Müller, K-R. (Ed), Neural Networks: Tricks of the Trade, 1524, pp Berlin, Heidelberg, New York: Springer-Verlag, Sharkey, A.J.C. Multi-Net Systems. In Sharkey, A. J. C. (Ed), Combining Artificial Neural Nets: Ensemble and Modular Multi-Net Systems, pp London: Springer-Verlag, Thrun, S.B., Bala, J., Bloedorn, E., Bratko, I., Cestnik, B., Cheng, J., De Jong, K., Dzeroski, S., Fahlman, S.E., Fisher, D., Hamann, R., Kaufman, K., Keller, S., Kononenko, I., Kreuziger, J., Michalski, R.S., Mitchell, T., Pachowicz, P., Reich, Y., Vafaie, H., van de Welde, W., Wenzel, W., Wnek, J. & Zhang, J. The MONK's Problems: A Performance Comparison of Different Learning Algorithms. Technical Report CMU-CS Pittsburgh, PA.: Carnegie-Mellon University, Computer Science Department, Wanas, N.M., Hodge, L. & Kamel, M.S. Adaptive Training Algorithm for an Ensemble of Networks. Proceedings of the 2001 International Joint Conference on Neural Networks (IJCNN'01), vol. 4, pp Los Alamitos, CA.: IEEE Computer Society Press, 2001.

14 References 15. Wolberg, W.H. & Mangasarian, O.L. Multisurface Method of Pattern Separation for Medical Diagnosis Applied to Breast Cytology. Proceedings of the National Academy of Sciences, USA, vol. 87(23), pp , Xu, L., Krzyzak, A. & Suen, C.Y. Several Methods for Combining Multiple Classifiers and Their Applications in Handwritten Character Recognition. IEEE Transactions on Systems, Man, and Cybernetics, vol. 22(3), pp , Hecht-Nielsen, R. (1987). Counterpropagation Networks. Applied Optics, vol. 26(23), pp