Artificial Intelligence

Slides:



Advertisements
Similar presentations
Introduction to Machine Learning BITS C464/BITS F464
Advertisements

Perceptron Learning Rule
Data Mining Classification: Alternative Techniques
Pattern Recognition Ku-Yaw Chang
ECE 8443 – Pattern Recognition Objectives: Course Introduction Typical Applications Resources: Syllabus Internet Books and Notes D.H.S: Chapter 1 Glossary.
Bayesian Decision Theory
Introduction to Machine Learning Anjeli Singh Computer Science and Software Engineering April 28 th 2008.
Chapter 1: Introduction to Pattern Recognition
CS292 Computational Vision and Language Pattern Recognition and Classification.
OUTLINE Course description, What is pattern recognition, Cost of error, Decision boundaries, The desgin cycle.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Classification, Chapter 1 1 Basic Probability.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Introduction to machine learning
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture Notes by Neşe Yalabık Spring 2011.
Introduction to Pattern Recognition
Pattern Recognition Vidya Manian Dept. of Electrical and Computer Engineering University of Puerto Rico INEL 5046, Spring 2007
Introduction to Pattern Recognition Charles Tappert Seidenberg School of CSIS, Pace University.
Introduction to Pattern Recognition Chapter 1 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis 1.
: Chapter 1: Introduction 1 Montri Karnjanadecha ac.th/~montri Principles of Pattern Recognition.
Classification. An Example (from Pattern Classification by Duda & Hart & Stork – Second Edition, 2001)
IBS-09-SL RM 501 – Ranjit Goswami 1 Basic Probability.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Perception Introduction Pattern Recognition Image Formation
S EGMENTATION FOR H ANDWRITTEN D OCUMENTS Omar Alaql Fab. 20, 2014.
Compiled By: Raj G Tiwari.  A pattern is an object, process or event that can be given a name.  A pattern class (or category) is a set of patterns sharing.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Lecture 10: 8/6/1435 Machine Learning Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
INTRODUCTION TO Machine Learning Adapted from: ETHEM ALPAYDIN Lecture Slides for.
1 E. Fatemizadeh Statistical Pattern Recognition.
Visual Information Systems Recognition and Classification.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Introduction to Pattern Recognition (การรู้จํารูปแบบเบื้องต้น)
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
Objectives: Terminology Components The Design Cycle Resources: DHS Slides – Chapter 1 Glossary Java Applet URL:.../publications/courses/ece_8443/lectures/current/lecture_02.ppt.../publications/courses/ece_8443/lectures/current/lecture_02.ppt.
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Data Mining and Decision Support
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
Pattern recognition – basic concepts. Sample input attribute, attribute, feature, input variable, independent variable (atribut, rys, příznak, vstupní.
Introduction to Pattern Recognition Chapter 1 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis 1.
Machine Learning with Spark MLlib
Who am I? Work in Probabilistic Machine Learning Like to teach 
Machine Learning for Computer Security
Pattern Classification Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley & Sons, 2000 Dr. Ding Yuxin Pattern Recognition.
Machine Learning overview Chapter 18, 21
LECTURE 11: Advanced Discriminant Analysis
Introduction Machine Learning 14/02/2017.
IMAGE PROCESSING RECOGNITION AND CLASSIFICATION
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
School of Computer Science & Engineering
LECTURE 01: COURSE OVERVIEW
CS668: Pattern Recognition Ch 1: Introduction
Pattern Recognition Sergios Theodoridis Konstantinos Koutroumbas
Introductory Seminar on Research: Fall 2017
CH. 1: Introduction 1.1 What is Machine Learning Example:
What is Pattern Recognition?
Introduction to Pattern Recognition and Machine Learning
Introduction to Pattern Recognition
REMOTE SENSING Multispectral Image Classification
An Introduction to Supervised Learning
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Introduction to Pattern Recognition
MIS2502: Data Analytics Clustering and Segmentation
LECTURE 01: COURSE OVERVIEW
MIS2502: Data Analytics Clustering and Segmentation
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Presentation transcript:

Artificial Intelligence Lecture 7 Introduction to Machine Learning kinza c

Machine Learning Applications Optical Character Recognition (OCR) Sorting letters by postal code Reconstructing text from printed materials Handwritten character recognition Analysis and identification of human patterns (Biometric classification) Face recognition Handwriting recognition Fingerprints and DNA sequence identification Iris scan identification Speech recognition/speaker identification

Applications

Examples of ML Computer aided diagnosis Prediction systems Medical imaging, EEG, ECG signal analysis Designed to assist (not replace) physicians Prediction systems Weather forecasting (based on satellite data) Banking applications Credit cards applicants classified by income, credit worthiness, mortgage amount, # of dependents, etc. Information Retrieval Data Mining

Machine Perception and Pattern Recognition Build a machine that can recognize patterns Pattern Recognition Theory, Algorithms, Systems to Put Patterns into Categories Relate Perceived Pattern to Previously Perceived Patterns By building such systems, we gain understanding of pattern recognition systems, particularly in humans Human recognition of patterns

A Machine Learning Example A fish processing plant wants to automate the process of sorting incoming fish according to species (salmon or sea bass) The automation system consists of A conveyor belt for incoming products A vision system with an overhead camera A computer to analyze images and control the robot arm

Example “Sorting incoming fish on a conveyor belt according to species using optical sensing”.

Problem Analysis Set up a camera and take some training sample images to extract features. Length Lightness Width Number and shape of fins Position of the mouth, etc… This is the set of all suggested features to explore for use in our classifier Purpose: To classify the future samples based on the data of extracted features from the training samples

Example Models Preprocessing Feature Extraction Classification There are differences between sea bass and salmon and are viewed as having different models. Preprocessing Segmentation Isolate fish from one another and from the background. Feature Extraction Information from a single fish is sent to a feature extractor whose purpose is to reduce the data by measuring certain features. Classification Evaluates the evidence presented and makes a final decision.

Selection Criterion Suppose sea bass is generally longer than a salmon. Select only the length of the fish as a possible feature for discrimination. To choose critical value of length, we could obtain some design or training samples of the different types of fish.

Histograms for the Length Feature

Selection Criterion No matter how we choose the threshold value of length, we cannot reliably separate sea bass from salmon. The length is a poor feature alone! Select the average lightness of the fish scales as a possible feature. )

Histograms for the Lightness Feature Length or Lightness, which one is a better feature? No value of either feature will “classify” all fish correctly )

Selection Criterion and Decision Boundary Seek a different feature to separate the fish. Use more than 1 feature at a time. Adopt the lightness feature. Add the width of the fish. Feature vector x is a 2D feature space. Fish xT = [x1, x2] Lightness Width

Two Features of Lightness and Width

Generalization and Decision Boundary We might add other features that are not correlated with the ones we already have, e.g. shape parameters. A precaution should be taken not to reduce the performance by adding redundant features. Ideally, the best decision boundary should be the one which provides an optimal performance such as in the following figure:

Complex Decision Boundary

Generalization and Decision Boundary However, our satisfaction is premature because the central aim of designing a classifier is to correctly classify novel input Issue of generalization! It is unlikely that the complex decision boundary would provide good generalization.

Generalization and Decision Boundary More training samples for estimating the true characteristics of the categories. Amount of data in most problems is limited. Even with vast data, the classifier can give a complicated decision boundary. A simple classifier with non-complex decision boundary can provide good generalization.

Selected decision boundary The decision boundary can be a simple curve which might represent the optimal trade-off.

Decision Boundary Selection Which of the boundaries would you choose? Simple linear boundary – training error > 0 Nonlinear complex boundary – training error = 0 Simpler nonlinear boundary – training error > 0

Terminologies in ML Features: a set of variables believed to carry discriminating and characterizing information about the objects under consideration Feature vector: A collection of d features, ordered in some meaningful way into a d-dimensional column vector, that represents the signature of the object to be identified. Feature space: The d-dimensional space in which the feature vectors lie. A d-dimensional vector in a d-dimensional space constitutes a point in that space. 22

Terminologies in ML Class: The category to which a given object belongs, typically denoted by  Decision boundary: A boundary in the d-dimensional feature space that separates patterns of different classes from each other Training Data: Data used during training of a classifier for which the correct labels are a priori known Testing Data: Unknown data to be classified. The correct class of this data are not known a priori 23

Terminologies in ML Classifier: An algorithm which adjusts its parameters to find the correct decision boundaries –through a learning algorithm using a training dataset –such that a cost function is minimized. Error: Incorrect labelling of the data by the classifier Training Performance: The ability/performance of the classifier in correctly identifying the classes of the training data, which it has already seen. It may not be a good indicator of the generalization performance. Generalization (Test Performance): The ability/performance of the classifier in identifying the classes of previously unseen 24

Kinds of Data 25

Good Features vs. Bad Features Ideally, for a given group of patterns coming from the same class, feature values should all be similar For patterns coming from different classes, the feature values should be different 26

Components of Machine Learning System What is the cost of making a decision –what is the risk of an incorrect decision? Determine the right type of classifier, right type of training algorithm and right type of classifier parameters. How to choose a classifier that minimizes the risk and maximizes the performance? Can we add prior knowledge, context to the classifier? Identify and extract features relevant to distinguishing sea bass from salmon, but invariant to noise, occlusion, rotations, scale, shift. What if we already have irrelevant features? Dimensionality reduction? Separate & isolate fish from background image Camera on a conveyor belt Fish 27

Machine Learning Systems Sensing Use of a transducer (camera or microphone). PR system depends on the bandwidth, resolution, sensitivity, distortion, etc. of the transducer. Segmentation Patterns should be well separated and should not overlap. Feature extraction Distinguishing features Invariant features with respect to translation, rotation and scale. 28

Machine Learning Systems Classification Use a feature vector provided by a feature extractor to assign the object to a category. Not always possible to determine the values of all the features. Post Processing Post-processor uses the output of the classifier to decide on the recommended action. Error rate 29

Learning and Adaptation Learning incorporates information from training samples in classifier design. It refers to some form of algorithm for reducing the error on training data. Supervised learning – A teacher provides a category label or cost for each pattern in the training set. 30

Learning and Adaptation Unsupervised learning The system forms clusters or “natural groupings” of the input patterns. The labels of the categories are unknown. Reinforcement Learning Learning with a critic. No desired category signal is given; instead the only teaching feedback is that the tentative category is right or wrong. 31

Conclusion Overwhelmed by the number, complexity and magnitude of the sub-problems of Machine Learning. Many of these sub-problems can indeed be solved. Mathematical theories solving some of these problems have in fact been discovered. Many fascinating unsolved problems still remain. 32