Introduction to Machine Learning Dmitriy Dligach.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Computational Learning An intuitive approach. Human Learning Objects in world –Learning by exploration and who knows? Language –informal training, inputs.
Decision Trees Decision tree representation ID3 learning algorithm
Albert Gatt Corpora and Statistical Methods Lecture 13.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
Quiz 1 on Wednesday ~20 multiple choice or short answer questions
An Overview of Machine Learning
Supervised Learning Recap
Discriminative and generative methods for bags of features
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
Support Vector Machines and Kernel Methods
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Lesson 8: Machine Learning (and the Legionella as a case study) Biological Sequences Analysis, MTA.
Unsupervised Training and Clustering Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Clustering. 2 Outline  Introduction  K-means clustering  Hierarchical clustering: COBWEB.
Advanced Multimedia Text Clustering Tamara Berg. Reminder - Classification Given some labeled training documents Determine the best label for a test (query)
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Oracle Data Mining Ying Zhang. Agenda Data Mining Data Mining Algorithms Oracle DM Demo.
嵌入式視覺 Pattern Recognition for Embedded Vision Template matching Statistical / Structural Pattern Recognition Neural networks.
Introduction Mohammad Beigi Department of Biomedical Engineering Isfahan University
This week: overview on pattern recognition (related to machine learning)
Data Mining Joyeeta Dutta-Moscato July 10, Wherever we have large amounts of data, we have the need for building systems capable of learning information.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Data mining and machine learning A brief introduction.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Machine Learning Lecture 11 Summary G53MLE | Machine Learning | Dr Guoping Qiu1.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 16 Nov, 3, 2011 Slide credit: C. Conati, S.
Apache Mahout. Mahout Introduction Machine Learning Clustering K-means Canopy Clustering Fuzzy K-Means Conclusion.
Ch10 Machine Learning: Symbol-Based
Clustering Supervised vs. Unsupervised Learning Examples of clustering in Web IR Characteristics of clustering Clustering algorithms Cluster Labeling 1.
Basic Machine Learning: Clustering CS 315 – Web Search and Data Mining 1.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
CS774. Markov Random Field : Theory and Application Lecture 19 Kyomin Jung KAIST Nov
Overview of Machine Learning RPI Robotics Lab Spring 2011 Kane Hadley.
1 Pattern Recognition Pattern recognition is: 1. A research area in which patterns in data are found, recognized, discovered, …whatever. 2. A catchall.
Learning from observations
Machine Learning, Decision Trees, Overfitting Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 14,
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Unsupervised Learning. Supervised learning vs. unsupervised learning.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Information Retrieval and Organisation Chapter 16 Flat Clustering Dell Zhang Birkbeck, University of London.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Machine Learning for Natural Language Processing Robin Tibor Schirrmeister Seminar Information Extraktion Wednesday, 6th November 2013.
Clustering Instructor: Max Welling ICS 178 Machine Learning & Data Mining.
1 Unsupervised Learning and Clustering Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of.
Basic Machine Learning: Clustering CS 315 – Web Search and Data Mining 1.
Locally Linear Support Vector Machines Ľubor Ladický Philip H.S. Torr.
6.S093 Visual Recognition through Machine Learning Competition Image by kirkh.deviantart.com Joseph Lim and Aditya Khosla Acknowledgment: Many slides from.
1 Learning Bias & Clustering Louis Oliphant CS based on slides by Burr H. Settles.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Introduction to Data Mining Clustering & Classification Reference: Tan et al: Introduction to data mining. Some slides are adopted from Tan et al.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Mixture Densities Maximum Likelihood Estimates.
Clustering Machine Learning Unsupervised Learning K-means Optimization objective Random initialization Determining Number of Clusters Hierarchical Clustering.
Data Mining and Text Mining. The Standard Data Mining process.
CMPS 142/242 Review Section Fall 2011 Adapted from Lecture Slides.
Machine Learning Supervised Learning Classification and Regression
Why study NLP if it doesn’t work?
Estimating Link Signatures with Machine Learning Algorithms
Basic machine learning background with Python scikit-learn
Machine Learning Week 1.
Supervised vs. unsupervised Learning
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Decision Trees Decision tree representation ID3 learning algorithm
Text Categorization Berlin Chen 2003 Reference:
Hairong Qi, Gonzalez Family Professor
Presentation transcript:

Introduction to Machine Learning Dmitriy Dligach

Representations Objects –Real-life phenomena viewed as objects and their properties (features) Feature Vectors – Examples –Text classification –Face recognition –WSD f0f0 f1f1 fnfn

Supervised Learning Vector-value pair –(x 0, y 0 ), (x 1, y 1 ), …, (x n, y n ) Task: learn function y = f(x) Algorithms –KNN –Decision Trees –Neural Networks –SVM

Issues in Supervised Learning Training data –Why are we learning? Test data –Unseen data Overfitting –Fitting noise reduces performance

Unsupervised Learning Only feature vectors are given –x 0, x 1, …, x n Task: group feature vectors into clusters Algorithms –Clustering k-means mixture of gaussians –Principal Component Analysis –Sequence labeling HMMs

Supervised Example: Decision Trees

A Tree

Word Sense Disambiguation (WSD) bat (noun)

Another DT Example Word Sense Disambiguation Given an occurrence of a word, decide which sense, or meaning, was intended. Example, run –run1: move swiftly ( I ran to the store.) –run2: operate (I run a store.) –run3: flow (Water runs from the spring.) –run4: length of torn stitches (Her stockings had a run.)

WSD Word Sense Disambiguation Categories –Use word sense labels (run1, run2, etc.) Features – describe context of word –near(w) : is the given word near word w? –pos: word’s part of speech –left(w): is word immediately preceded by w? –etc.

Using a decision Tree pos near(race)near(stocking) near(river) run1 run3 noun yes no verb yesno yes 4pm run4 Given an event (=list of feature values): – Start at the root. – At each interior node, follow the outgoing arc for the feature value that matches our event – When we reach a leaf node, return its category. “I saw John run a race by a river.”

WSD: Sample Training Data FeaturesWord POSnear(race)near(river)near(stockings) Sense NounNo run4 VerbNo run1 VerbNoYesNorun3 NounYes run4 VerbNo Yesrun1 VerbYes Norun2 VerbNoYes run3

Unsupervised Example: K-Means Distance between two objects –Cosine distance –Euclidean distance Algorithm –Pick cluster centers at random –Assign the data points to the nearest clusters –Re-compute the cluster centers –Re-assign the data points –Continue until the clusters settle Hard clustering vs. soft clustering

Interactive Demos K-Means – SVMs –

ML Reference Tom Mitchell “Machine Learning”