12 March 1999Dip HI KBS Module1 Machine Learning Lucy Hederman.

Slides:



Advertisements
Similar presentations
Slides from: Doug Gray, David Poole
Advertisements

DECISION TREES. Decision trees  One possible representation for hypotheses.
Artificial Intelligence 13. Multi-Layer ANNs Course V231 Department of Computing Imperial College © Simon Colton.
Decision Tree Approach in Data Mining
Computer Science Department FMIPA IPB 2003 Neural Computing Yeni Herdiyeni Computer Science Dept. FMIPA IPB.
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
Combining Inductive and Analytical Learning Ch 12. in Machine Learning Tom M. Mitchell 고려대학교 자연어처리 연구실 한 경 수
Instance Based Learning
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Artificial Neural Networks
Machine Learning Neural Networks
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Neural Networks Basic concepts ArchitectureOperation.
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Machine Learning Group University College Dublin Nearest Neighbour Classifiers Lazy v’s Eager k-NN Condensed NN.
Induction of Decision Trees
Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers.
Data Mining with Decision Trees Lutz Hamel Dept. of Computer Science and Statistics University of Rhode Island.
Three kinds of learning
Machine Learning: Symbol-Based
Artificial Neural Networks (ANNs)
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
CS Instance Based Learning1 Instance Based Learning.
Hub Queue Size Analyzer Implementing Neural Networks in practice.
Introduction to Machine Learning Approach Lecture 5.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
Neural Networks AI – Week 21 Sub-symbolic AI One: Neural Networks Lee McCluskey, room 3/10
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Artificial Neural Network Theory and Application Ashish Venugopal Sriram Gollapalli Ulas Bardak.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Multi-Layer Perceptrons Michael J. Watts
Chapter 9 Neural Network.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
NEURAL NETWORKS FOR DATA MINING
Chapter 7 Neural Networks in Data Mining Automatic Model Building (Machine Learning) Artificial Intelligence.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
Neural Networks II By Jinhwa Kim. 2 Neural Computing is a problem solving methodology that attempts to mimic how human brain function Artificial Neural.
March 1999Dip HI KBS1 Knowledge-based Systems Alternatives to Rules.
© 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang 12-1 Chapter 12 Advanced Intelligent Systems.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
CS Inductive Bias1 Inductive Bias: How to generalize on novel data.
CS 478 – Tools for Machine Learning and Data Mining Perceptron.
CpSc 881: Machine Learning Instance Based Learning.
CHAPTER 15 Neural Computing: The Basics. n Artificial Neural Networks (ANN) n Mimics How Our Brain Works n Machine Learning.
CpSc 810: Machine Learning Instance Based Learning.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Data Mining and Decision Support
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Part II - Classification© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II - Classification Margaret H. Dunham Department of Computer.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Lecture 2 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 2/1 Dr.-Ing. Erwin Sitompul President University
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Deep Learning Amin Sobhani.
Neural Computing: The Basics
Issues in Decision-Tree Learning Avoiding overfitting through pruning
K Nearest Neighbor Classification
Chapter 12 Advanced Intelligent Systems
Biological and Artificial Neuron
Instance Based Learning
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Presentation transcript:

12 March 1999Dip HI KBS Module1 Machine Learning Lucy Hederman

12 March 1999Dip HI KBS Module2 KBS Development Stage 1: analysis of the problem that produces a representation of the problem that can be manipulated by the reasoning system - this representation is often a set of attribute values. Stage 2: developing the reasoning mechanism that manipulates the problem representation to produce a solution.

12 March 1999Dip HI KBS Module3 Stage 2 Knowledge engineering - manually –rule development for a rule-based ES Learning - similarity-based –generalise from examples (training data) Learning - explanation-based –build on prior knowledge –use small number of canonical examples –incorporate explanations, analogy,...

12 March 1999Dip HI KBS Module4 Risk Assessment Example Expert might develop rules like –if collateral is adequate and credit history is good then risk is low. Alternatively build a system which learns from existing data on loan application decisions (see attached). –Similarity-based learning

12 March 1999Dip HI KBS Module5 Classifying apples and pears To what class does this belong?

12 March 1999Dip HI KBS Module6 Supervised Learning Supervised learning –training data classified already Unsupervised learning –acquire useful(?) knowledge without correctly classified training data –category formation –scientific discovery We look at supervised learning only.

12 March 1999Dip HI KBS Module7 Learnability Induction depends on there being useful generalisations possible in the representation language used. Learnability of concepts in a representation language is the ability to express the concept concisely. Random classifications are not learnable.

12 March 1999Dip HI KBS Module8 Similarity-based learning Decision tree (rule) induction –induce a decision tree (set of rules) from the training data. k-nearest neighbour classification –classify a new problem based on the k most similar cases in the training data. Artificial Neural Networks –adjust weights in an NN to reduce errors on training data.

12 March 1999Dip HI KBS Module9 Decision Tree Induction Aim to induce a tree which –correctly classifies all training data –will correctly classify unseen cases ID3 algorithm assumes that the simplest tree that covers all the training examples is the best at unseen problems. –Leaving out extraneous tests should be good for generalising.

12 March 1999Dip HI KBS Module10 ID3 Top-down construction –add selected tests under nodes –each test further partitions the samples –continue till each partition is homogeneous Information-theoretic test selection –maximise information gain ID3 works surprisingly well. Variations and alternatives exist.

12 March 1999Dip HI KBS Module11 k-Nearest Neighbour Classification Data base of previously classified cases kept throughout. Category of target case decided by category of its k nearest neighbours. No inducing or training of a model. “Lazy” learning –work deferred to runtime –compare with neural networks - eager learners

12 March 1999Dip HI KBS Module12 “Nearest” - distance/similarity For query q and training set X (described by features F) compute d(x,q) for each x  X, where and where

12 March 1999Dip HI KBS Module13 k-NN and Noise 1-NN easy to implement –susceptible to noise a misclassification every time a noisy pattern retrieved k-NN with k  3 will overcome this Either –straight voting between the k examples or –weighted votes depending on “nearness” of each example.

12 March 1999Dip HI KBS Module14 K-NN vs. Decision Trees Decision trees test features serially. –If two cases don’t match on first feature tried they don’t match at all. K-NN considers all features in parallel. For some tasks serial testing is OK, for others it’s not.

12 March 1999Dip HI KBS Module15 Dimension reduction in k-NN Not all features required –noisy features a hindrance Some examples redundant –retrieval time depends on no. of examples p features q best features n covering examples m examples

12 March 1999Dip HI KBS Module16 Condensed NN 100 examples 2 categories Different CNN solutions

12 March 1999Dip HI KBS Module17 Feature weighting Feature weights –modify the effect of large continuous distance values –allow some features to be treated as more important than others pull cases with important features in common closer together.

12 March 1999Dip HI KBS Module18 Feature weighting Introspective learning - Test training data on itself –For a correct retrieval increase weight of matching features (pull) decrease weight of un-matching features (pull) –For an incorrect retrieval decrease weight of matching features (push) increase weight of un-matching features (push) Pull Push

12 March 1999Dip HI KBS Module19 (Artificial) Neural Networks Decision tree induction builds a symbolic “causal” model from training data. k-NN builds no model. A neural network is a sub-symbolic, non- causal, distributed, “black box”, model built from training data. ANN output is continuous whereas a k-NN classifies into discrete classes.

12 March 1999Dip HI KBS Module20 NN Prediction of Malignancy A. Tailor and co. paper describes a neural network which computes a probability of malignancy from age, morphological features, and sonographic data. Describes design and testing of the NN. Note intro to NNs in the Appendix

12 March 1999Dip HI KBS Module21 ANN Advantages Particularly suited to pattern recognition –character, speech, image Suited to domains where there is no domain theory or model. Robust - Handle noisy and incomplete data well. Potentially fast. Parallel processing. Flexible and easy to maintain.

12 March 1999Dip HI KBS Module22 ANN Problems Lack explanation Currently implemented in software mostly. Training times can be tedious. Need lots of training and test data. –True of similarity-based learning in general.

12 March 1999Dip HI KBS Module23 ANN Processing Element (PE) Summation - gives PE’s activation level Transfer function - modifies the activation level to produce a reasonable output value (e.g. 0-1).

12 March 1999Dip HI KBS Module24 Typical ANN Structure PE Input layer Hidden layer Output layer There may be –additional hidden layers. –different topologies –different connectivity Choosing ANN structure –is based on problem and –requires some expertise. PE

12 March 1999Dip HI KBS Module25 Learning/Training Aim to obtain desired outputs for each training example. Backpropagation is the most popular learning algorithm. –Initialise all weights associated with inputs to each PE. –Present sample inputs to ANN. –Compare ANN outputs with desired output. –Alter weights to reduce the mean square error, and repeat. until the error is within some tolerance.

12 March 1999Dip HI KBS Module26 Overfitting Training time Error In-sample error Generalisation error Too much training will result in a (k-NN or ANN) model that makes minimal errors on the training data (memorises), but no longer generalises well. Beware.

12 March 1999Dip HI KBS Module27 ANN Development Collect data Separate into training and test sets Define a network structure Select a learning algorithm Set parameters, values, weights Transform data to network inputs Start training, revise weights Stop and test Use the network for new cases. Get more better data Reseparate Redefine structure Select another algorithm Reset