Machine Learning Tutorial Amit Gruber The Hebrew University of Jerusalem.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Statistical Machine Learning- The Basic Approach and Current Research Challenges Shai Ben-David CS497 February, 2007.
Introduction to Machine Learning BITS C464/BITS F464
Godfather to the Singularity
COMPUTER AIDED DIAGNOSIS: CLASSIFICATION Prof. Yasser Mostafa Kadah –
Feature Selection as Relevant Information Encoding Naftali Tishby School of Computer Science and Engineering The Hebrew University, Jerusalem, Israel NIPS.
INTRODUCTION TO MACHINE LEARNING David Kauchak CS 451 – Fall 2013.
Machine learning continued Image source:
LECTURE 11: BAYESIAN PARAMETER ESTIMATION
An Overview of Machine Learning
Automated Anomaly Detection, Data Validation and Correction for Environmental Sensors using Statistical Machine Learning Techniques
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
The Unreasonable Effectiveness of Data Alon Halevy, Peter Norvig, and Fernando Pereira Kristine Monteith May 1, 2009 CS 652.
CIS 678 Artificial Intelligence problems deduction, reasoning knowledge representation planning learning natural language processing motion and manipulation.
Introduction to Machine Learning course fall 2007 Lecturer: Amnon Shashua Teaching Assistant: Yevgeny Seldin School of Computer Science and Engineering.
Presented by Zeehasham Rasheed
Classification and application in Remote Sensing.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Maria-Florina Balcan A Theoretical Model for Learning from Labeled and Unlabeled Data Maria-Florina Balcan & Avrim Blum Carnegie Mellon University, Computer.
Text Classification Using Stochastic Keyword Generation Cong Li, Ji-Rong Wen and Hang Li Microsoft Research Asia August 22nd, 2003.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Machine Learning Theory Maria-Florina Balcan Lecture 1, Jan. 12 th 2010.
Introduction to machine learning
Introduction to Data Mining Engineering Group in ACL.
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
1 Text Categorization  Assigning documents to a fixed set of categories  Applications:  Web pages  Recommending pages  Yahoo-like classification hierarchies.
Machine Learning Theory Maria-Florina (Nina) Balcan Lecture 1, August 23 rd 2011.
This week: overview on pattern recognition (related to machine learning)
ECSE 6610 Pattern Recognition Professor Qiang Ji Spring, 2011.
MACHINE LEARNING 張銘軒 譚恆力 1. OUTLINE OVERVIEW HOW DOSE THE MACHINE “ LEARN ” ? ADVANTAGE OF MACHINE LEARNING ALGORITHM TYPES  SUPERVISED.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Statistical Modeling with SAS/STAT Cheng Lei Department of Electrical and Computer Engineering University of Victoria April 9, 2015.
An Introduction to Machine Learning and Natural Language Processing Tools Presented by: Mark Sammons, Vivek Srikumar (Many slides courtesy of Nick Rizzolo)
Information Filtering LBSC 796/INFM 718R Douglas W. Oard Session 10, April 13, 2011.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Yazd University, Electrical and Computer Engineering Department Course Title: Advanced Software Engineering By: Mohammad Ali Zare Chahooki 1 Introduction.
Empirical Research Methods in Computer Science Lecture 6 November 16, 2005 Noah Smith.
Spam Detection Ethan Grefe December 13, 2013.
Week 1 - An Introduction to Machine Learning & Soft Computing
CHAPTER 6 Naive Bayes Models for Classification. QUESTION????
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Machine Learning for Spam Filtering 1 Sai Koushik Haddunoori.
CHAPTER 1: Introduction. 2 Why “Learn”? Machine learning is programming computers to optimize a performance criterion using example data or past experience.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
CZ5225: Modeling and Simulation in Biology Lecture 7, Microarray Class Classification by Machine learning Methods Prof. Chen Yu Zong Tel:
Supervised Machine Learning: Classification Techniques Chaleece Sandberg Chris Bradley Kyle Walsh.
Lecture 5: Statistical Methods for Classification CAP 5415: Computer Vision Fall 2006.
Learning Kernel Classifiers 1. Introduction Summarized by In-Hee Lee.
Introduction to Classification & Clustering Villanova University Machine Learning Lab Module 4.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
1 Text Categorization  Assigning documents to a fixed set of categories  Applications:  Web pages  Recommending pages  Yahoo-like classification hierarchies.
ORGANIZING . 1.Sort messages quickly. 2.Group similar messages in folders or labels. 3.Route mail efficiently to specific folders or labels. 4.Reduce.
Recommender Systems 11/04/2017
Machine Learning with Spark MLlib
Machine Learning for Computer Security
Introduction to Classification & Clustering
Information Organization: Overview
School of Computer Science & Engineering

Entity- & Topic-Based Information Ordering
Machine Learning With Python Sreejith.S Jaganadh.G.
CH. 1: Introduction 1.1 What is Machine Learning Example:
Vincent Granville, Ph.D. Co-Founder, DSC
What is Pattern Recognition?
Text Categorization Rong Jin.
Recommender Systems Copyright: Dietmar Jannah, Markus Zanker and Gerhard Friedrich (slides based on their IJCAI talk „Tutorial: Recommender Systems”)
LECTURE 07: BAYESIAN ESTIMATION
Information Organization: Overview
NAÏVE BAYES CLASSIFICATION
Speech recognition, machine learning
Presentation transcript:

Machine Learning Tutorial Amit Gruber The Hebrew University of Jerusalem

Example: Spam Filter Spam message: unwanted message –Dozens or even hundreds per day Goal: Automatically distinguish between spam and non-spam messages

Spam message 1

Spam message 2

Spam message 3

Spam message 4

How to Distinguish ? Message contents ? –Automatic semantic analysis is yet to be solved Message sender ? –What about unfamiliar senders or fake senders ? Collection of keywords ? Message Length ? Mail server ? Time of delivery ?

How to Distinguish ? It’s hard to define an explicit set of rules to distinguish between spam and non-spam Learn the concept of “spam” from examples !

Example: Gender Classification

The Power of Learning: Real Life example How much time does it take you to get to work ? –First approach: Analyze your route Distance, traffic lights, traffic, etc… Can be quite complicated… –Second Approach: how much time does it usually take ? Despite of some variance, works remarkably well! Requires “training” for different times May fail in special cases

Machine Translation

Collaborative Filtering Collaborative Filtering: Prediction of user ratings based on the ratings of other users Examples: –Movie ratings –Product recommendation Is this of merely theoretical interest ??

Netflix Prize Over 100 million ratings from 480 thousand customers over movie titles (sparsity: )

Recommendation system

Machine Learning Applications Search Engines Collaborative Filtering (Netflix, Amazon) Face, speech and pattern Recognition Machine Translation Natural language processing Medical diagnosis and treatment Bioinformatics Computer games Many more !

Generalization: Train vs. Test The central assumption we make is that the train set and the new examples are “similar” Formally, the assumption is that samples are drawn from the same distribution Is this assumption realistic ?

Train vs. Test: Might Fail to Generalize

Acquiring a good train set Have a huge train set –Train data might be available on the web –Use humans to collect data –Collect results (or aggregations thereof) of user actions Unsupervised methods – require only raw data, no need for labels !

Machine Learning Strategies Discriminative Approach –Feature selection: find the features that carry the most information for separation Generative Approach –Model the data using a generative process –Estimate the parameters of the model

Supervised vs. Unsupervised Supervised Machine Learning –Classification (learning) –Collection of large representative train set might not be simple Unsupervised Machine Learning –Clustering The number of clusters may be known or unknown –Usually plenty of train data is available

Discriminative Learning Data representation and Feature selection: What is relevant for classification ? –Gender classification: hair, ears, make up, beard, moustache, etc. Linear Separation –SVM, Fisher LDA, Perceptron and more –Different criteria for separation – what would generalize well ? –Non-linear separation

Linear Separation

Nonlinear Separation (Kernel Trick)

Generative Approach Model the observations using a generative process The generative process induces a distribution over the observations Learn a set of parameters

Statistical Approach – Real Life Example You’re stuck in traffic. Which Lane is faster? The complicated approach: –Consider the traffic, trucks, merging lanes, etc. The statistical (Bayesian) Approach: –Which lane is usually faster ? (prior) –What are you seeing ? (evidence)

Summary Machine Learning: Learn a concept from examples For good generalization, train data has to faithfully represent test data Many potential applications Already in use and works remarkably well