Download presentation
Presentation is loading. Please wait.
1
Introduction to Machine Learning Dmitriy Dligach
2
Representations Objects –Real-life phenomena viewed as objects and their properties (features) Feature Vectors – Examples –Text classification –Face recognition –WSD f0f0 f1f1 fnfn
3
Supervised Learning Vector-value pair –(x 0, y 0 ), (x 1, y 1 ), …, (x n, y n ) Task: learn function y = f(x) Algorithms –KNN –Decision Trees –Neural Networks –SVM
4
Issues in Supervised Learning Training data –Why are we learning? Test data –Unseen data Overfitting –Fitting noise reduces performance
5
Unsupervised Learning Only feature vectors are given –x 0, x 1, …, x n Task: group feature vectors into clusters Algorithms –Clustering k-means mixture of gaussians –Principal Component Analysis –Sequence labeling HMMs
6
Supervised Example: Decision Trees
7
A Tree
8
Word Sense Disambiguation (WSD) bat (noun) http://wordnet.princeton.edu/perl/webwn http://verbs.colorado.edu/html_groupings/
9
Another DT Example Word Sense Disambiguation Given an occurrence of a word, decide which sense, or meaning, was intended. Example, run –run1: move swiftly ( I ran to the store.) –run2: operate (I run a store.) –run3: flow (Water runs from the spring.) –run4: length of torn stitches (Her stockings had a run.)
10
WSD Word Sense Disambiguation Categories –Use word sense labels (run1, run2, etc.) Features – describe context of word –near(w) : is the given word near word w? –pos: word’s part of speech –left(w): is word immediately preceded by w? –etc.
11
Using a decision Tree pos near(race)near(stocking) near(river) run1 run3 noun yes no verb yesno yes 4pm run4 Given an event (=list of feature values): – Start at the root. – At each interior node, follow the outgoing arc for the feature value that matches our event – When we reach a leaf node, return its category. “I saw John run a race by a river.”
12
WSD: Sample Training Data FeaturesWord POSnear(race)near(river)near(stockings) Sense NounNo run4 VerbNo run1 VerbNoYesNorun3 NounYes run4 VerbNo Yesrun1 VerbYes Norun2 VerbNoYes run3
13
Unsupervised Example: K-Means Distance between two objects –Cosine distance –Euclidean distance Algorithm –Pick cluster centers at random –Assign the data points to the nearest clusters –Re-compute the cluster centers –Re-assign the data points –Continue until the clusters settle Hard clustering vs. soft clustering
14
Interactive Demos K-Means –http://home.dei.polimi.it/matteucc/Clustering/tutorial_html/AppletKM.htmlhttp://home.dei.polimi.it/matteucc/Clustering/tutorial_html/AppletKM.html SVMs –http://www.csie.ntu.edu.tw/~cjlin/libsvm/#GUIhttp://www.csie.ntu.edu.tw/~cjlin/libsvm/#GUI
15
ML Reference Tom Mitchell “Machine Learning” http://home.dei.polimi.it/matteucc/Clustering/tutorial_html/AppletKM.html http://www.aaai.org/AITopics/html/machine.html
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.