CS639: Data Management for Data Science

Slides:



Advertisements
Similar presentations
Statistical Machine Learning- The Basic Approach and Current Research Challenges Shai Ben-David CS497 February, 2007.
Advertisements

Learning from Observations Chapter 18 Section 1 – 3.
1 Machine Learning: Lecture 3 Decision Tree Learning (Based on Chapter 3 of Mitchell T.., Machine Learning, 1997)
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
K-NEAREST NEIGHBORS AND DECISION TREE Nonparametric Supervised Learning.
Decision Tree Rong Jin. Determine Milage Per Gallon.
Machine Learning CPSC 315 – Programming Studio Spring 2009 Project 2, Lecture 5.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Learning From Observations
Learning from Observations Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 18 Fall 2004.
Three kinds of learning
Decision trees and empirical methodology Sec 4.3,
ICS 273A Intro Machine Learning
© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
Part I: Classification and Bayesian Learning
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Data Mining Joyeeta Dutta-Moscato July 10, Wherever we have large amounts of data, we have the need for building systems capable of learning information.
Mehdi Ghayoumi Kent State University Computer Science Department Summer 2015 Exposition on Cyber Infrastructure and Big Data.
INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.
Learning CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Learning from Observations Chapter 18 Through
CHAPTER 18 SECTION 1 – 3 Learning from Observations.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Introduction to Machine Learning Supervised Learning 姓名 : 李政軒.
Learning from observations
Machine Learning, Decision Trees, Overfitting Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 14,
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Machine Learning Extract from various presentations: University of Nebraska, Scott, Freund, Domingo, Hong,
Today’s Topics Learning Decision Trees (Chapter 18) –We’ll use d-trees to introduce/motivate many general issues in ML (eg, overfitting reduction) “Forests”
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
Concept learning, Regression Adapted from slides from Alpaydin’s book and slides by Professor Doina Precup, Mcgill University.
1 Universidad de Buenos Aires Maestría en Data Mining y Knowledge Discovery Aprendizaje Automático 5-Inducción de árboles de decisión (2/2) Eduardo Poggi.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
CS Inductive Bias1 Inductive Bias: How to generalize on novel data.
Data Mining and Decision Support
ECE 5984: Introduction to Machine Learning Dhruv Batra Virginia Tech Topics: –Decision/Classification Trees Readings: Murphy ; Hastie 9.2.
DECISION TREES Asher Moody, CS 157B. Overview  Definition  Motivation  Algorithms  ID3  Example  Entropy  Information Gain  Applications  Conclusion.
Chapter 18 Section 1 – 3 Learning from Observations.
Exploration Seminar 8 Machine Learning Roy McElmurry.
Machine Learning Lecture 1: Intro + Decision Trees Moshe Koppel Slides adapted from Tom Mitchell and from Dan Roth.
Learning From Observations Inductive Learning Decision Trees Ensembles.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Who am I? Work in Probabilistic Machine Learning Like to teach 
Learning from Observations
Learning from Observations
Machine Learning Inductive Learning and Decision Trees
CS 445/545 Machine Learning Winter, 2017
DECISION TREES An internal node represents a test on an attribute.
k-Nearest neighbors and decision tree
Chapter 18 From Data to Knowledge
Instructor: Pedro Domingos
Introduce to machine learning
Machine Learning overview Chapter 18, 21
Presented By S.Yamuna AP/CSE
CS 445/545 Machine Learning Spring, 2017
Data Science Algorithms: The Basic Methods
Estimating Link Signatures with Machine Learning Algorithms
Supervised Learning Seminar Social Media Mining University UC3M
Decision Trees Greg Grudic
Overview of Machine Learning
Ensemble learning.
Learning from Observations
Learning Chapter 18 and Parts of Chapter 20
CS639: Data Management for Data Science
Learning from Observations
Decision trees One possible representation for hypotheses
Machine Learning in Business John C. Hull
Presentation transcript:

CS639: Data Management for Data Science Lecture 16: Intro to ML and Decision Trees Theodoros Rekatsinas (lecture by Ankur Goswami many slides from David Sontag)

Today’s Lecture Intro to Machine Learning Types of Machine Learning Decision Trees

Intro to Machine Learning

What is Machine Learning? “Learning is any process by which a system improves performance from experience” – Herbert Simon Definition by Tom Mitchell (1998): Machine Learning is the study of algorithms that Improve their performance P at some task T with experience E A well-defined learning task is given by <P, T, E>.

What is Machine Learning? Machine Learning is the study of algorithms that Improve their performance P at some task T with experience E A well-defined learning task is given by <P, T, E>. Experience: data-driven task, thus statistics, probability Example: use height and weight to predict gender

When do we use machine learning? ML is used when: Human expertise does not exist (navigating on Mars) Humans can’t explain their expertise (speech recognition) Models must be customized (personalized medicine) Models are based on huge amounts of data (genomics)

A task that requires machine learning What makes a hand drawing be 2?

Modern machine learning: Autonomous cars

Modern machine learning: Scene Labeling

Modern machine learning: Speech Recognition

2. Types of Machine Learning

Types of Learning Supervised (inductive) learning Given: training data + desired outputs (labels) Unsupervised learning Given: training data (without desired outputs) Semi-supervised learning Given: training data + a few desired outputs Reinforcement learning Rewards from sequence of actions

Supervised Learning: Regression Given Learn a function f(x) to predict y given x y is real-valued == regression

Supervised Learning: Classification Given Learn a function f(x) to predict y given x y is categorical == regression

Supervised Learning: Classification Given Learn a function f(x) to predict y given x y is categorical == regression

Supervised Learning Value x can be multi-dimensional. Each dimension corresponds to an attribute

We will cover later in the class Types of Learning Supervised (inductive) learning Given: training data + desired outputs (labels) Unsupervised learning Given: training data (without desired outputs) Semi-supervised learning Given: training data + a few desired outputs Reinforcement learning Rewards from sequence of actions We will cover later in the class

3. Decision Trees

A learning problem: predict fuel efficiency

Hypotheses: decision trees f: X → Y Informal A hypothesis is a certain function that we believe (or hope) is similar to the true function, the target function that we want to model.

What functions can Decision Trees represent?

Space of possible decision trees How will we choose the best one? Lets first look at how to split nodes, then consider how to find the best tree

What is the simplest tree? Always predict mpg = bad We just take the majority class Is this a good tree? We need to evaluate its performance Performance: We are correct on 22 examples and incorrect on 18 examples

A decision stump

Recursive step

Recursive step

Second level of tree

Are all decision trees equal? Many trees can represent the same concept But, not all trees will have the same size! e.g., φ = ( A ∧ B)∨(¬A ∧ C) -- ((A and B) or ( not A and C)) Which tree do we prefer?

Learning decision trees is hard Learning the simplest (smallest) decision tree is an NP-complete problem [Hyafil & Rivest ’76] Resort to a greedy heuristic: Start from empty decision tree Split on next best attribute (feature) Recurse

Splitting: choosing a good attribute

Measuring uncertainty Good split if we are more certain about classification after split Deterministic good (all true or all false) Uniform distribution bad What about distributions in between?

Entropy

High, Low Entropy

Entropy Example

Conditional Entropy

Information gain

Learning decision trees

A decision stump

Base Cases: An Idea Base Case One: If all records in current data subset have the same output then do not recurse Base Case Two: If all records have exactly the same set of input attributes then do not recurse

The problem with Base Case 3

If we omit Base Case 3

Summary: Building Decision Trees

From categorical to real-valued attributes

What you need to know about decision trees Decision trees are one of the most popular ML tools Easy to understand, implement, and use Computationally cheap (to solve heuristically) Information gain to select attributes Presented for classification but can be used for regression and density estimation too Decision trees will overfit!!! We will see the definition of overfitting and related concepts later in class.