HADOOP David Kauchak CS451 – Fall 2013. Admin Assignment 7.

Slides:



Advertisements
Similar presentations
Linear Regression.
Advertisements

Regularization David Kauchak CS 451 – Fall 2013.
Linear Classifiers/SVMs
Clustering Beyond K-means
Support Vector Machines Instructor Max Welling ICS273A UCIrvine.
PERCEPTRON LEARNING David Kauchak CS 451 – Fall 2013.
Supervised Learning Recap
Unsupervised learning
CMPUT 466/551 Principal Source: CMU
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Cos 429: Face Detection (Part 2) Viola-Jones and AdaBoost Guest Instructor: Andras Ferencz (Your Regular Instructor: Fei-Fei Li) Thanks to Fei-Fei Li,
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
x – independent variable (input)
Intro to Linear Methods Reading: DH&S, Ch 5.{1-4,8} hip to be hyperplanar...
Introduction to Linear and Logistic Regression. Basic Ideas Linear Transformation Finding the Regression Line Minimize sum of the quadratic residuals.
1 Linear Classification Problem Two approaches: -Fisher’s Linear Discriminant Analysis -Logistic regression model.
Classification and Prediction: Regression Analysis
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Section 8.3 – Systems of Linear Equations - Determinants Using Determinants to Solve Systems of Equations A determinant is a value that is obtained from.
Collaborative Filtering Matrix Factorization Approach
1 Linear Methods for Classification Lecture Notes for CMPUT 466/551 Nilanjan Ray.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
MACHINE LEARNING BASICS David Kauchak CS457 Fall 2011.
MULTICLASS CONTINUED AND RANKING David Kauchak CS 451 – Fall 2013.
GEOMETRIC VIEW OF DATA David Kauchak CS 451 – Fall 2013.
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
EM and expected complete log-likelihood Mixture of Experts
ENSEMBLE LEARNING David Kauchak CS451 – Fall 2013.
ADVANCED CLASSIFICATION TECHNIQUES David Kauchak CS 159 – Fall 2014.
Ordinary Least-Squares Emmanuel Iarussi Inria. Many graphics problems can be seen as finding the best set of parameters for a model, given some data Surface.
Jeff Howbert Introduction to Machine Learning Winter Regression Linear Regression.
LOGISTIC REGRESSION David Kauchak CS451 – Fall 2013.
BOOSTING David Kauchak CS451 – Fall Admin Final project.
ICS 178 Introduction Machine Learning & data Mining Instructor max Welling Lecture 6: Logistic Regression.
CS 478 – Tools for Machine Learning and Data Mining Linear and Logistic Regression (Adapted from various sources) (e.g., Luiz Pessoa PY 206 class at Brown.
LARGE MARGIN CLASSIFIERS David Kauchak CS 451 – Fall 2013.
ADVANCED PERCEPTRON LEARNING David Kauchak CS 451 – Fall 2013.
Text Classification 2 David Kauchak cs459 Fall 2012 adapted from:
9.2A- Linear Regression Regression Line = Line of best fit The line for which the sum of the squares of the residuals is a minimum Residuals (d) = distance.
UNSUPERVISED LEARNING David Kauchak CS 451 – Fall 2013.
Machine Learning CUNY Graduate Center Lecture 4: Logistic Regression.
Jeff Howbert Introduction to Machine Learning Winter Regression Linear Regression Regression Trees.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Math 4030 – 11b Method of Least Squares. Model: Dependent (response) Variable Independent (control) Variable Random Error Objectives: Find (estimated)
Assignments CS fall Assignment 1 due Generate the in silico data set of 2sin(1.5x)+ N (0,1) with 100 random values of x between.
Regress-itation Feb. 5, Outline Linear regression – Regression: predicting a continuous value Logistic regression – Classification: predicting a.
Logistic Regression William Cohen.
Curve Fitting Introduction Least-Squares Regression Linear Regression Polynomial Regression Multiple Linear Regression Today’s class Numerical Methods.
Engineering Analysis ENG 3420 Fall 2009 Dan C. Marinescu Office: HEC 439 B Office hours: Tu-Th 11:00-12:00.
1 Simple Linear Regression and Correlation Least Squares Method The Model Estimating the Coefficients EXAMPLE 1: USED CAR SALES.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Linear Regression. Regression Consider the following 10 data pairs comparing the yield of an experiment to the temperature at which the experiment was.
Machine Learning – Classification David Fenyő
Gradient descent David Kauchak CS 158 – Fall 2016.
Large Margin classifiers
Deep learning David Kauchak CS158 – Fall 2016.
CH 5: Multivariate Methods
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
CS 4/527: Artificial Intelligence
Classification Discriminant Analysis
Pattern Recognition PhD Course.
Collaborative Filtering Matrix Factorization Approach
Ying shen Sse, tongji university Sep. 2016
Mathematical Foundations of BME Reza Shadmehr
Linear regression Fitting a straight line to observations.
Nonlinear Fitting.
Discrete Least Squares Approximation
Mathematical Sciences
Backpropagation David Kauchak CS159 – Fall 2019.
Presentation transcript:

HADOOP David Kauchak CS451 – Fall 2013

Admin Assignment 7

logistic regression: three views linear classifier conditional model logistic linear model minimizing logistic loss

Logistic regression Why is it called logistic regression?

A digression: regression vs. classification Raw dataLabel extract features f 1, f 2, f 3, …, f n featuresLabel classification: discrete (some finite set of labels) regression: real value

linear regression Given some points, find the line that best fits/explains the data Our model is a line, i.e. we’re assuming a linear relationship between the feature and the label value How can we find this line? f1f1 response (y)

Linear regression Learn a line h that minimizes some loss/error function: feature (x) response (y) Sum of the individual errors: 0/1 loss!

Error minimization How do we find the minimum of an equation? Take the derivative, set to 0 and solve (going to be a min or a max) Any problems here? Ideas?

Linear regression feature response squared error is convex!

Linear regression Learn a line h that minimizes an error function: in the case of a 2d line: function for a line feature response

Linear regression We’d like to minimize the error Find w 1 and w 0 such that the error is minimized We can solve this in closed form

Multiple linear regression If we have m features, then we have a line in m dimensions weights

Multiple linear regression We can still calculate the squared error like before Still can solve this exactly!

Logistic function

Logistic regression Find the best fit of the data based on a logistic

Big Data What is “big data”? What are some sources of big data? What are the challenges of dealing with big data? What are some of the tools you’ve heard of? For more info:

Hadoop: guest speaker Patricia Florissi CTO of EMC PhD from Columbia University

Hadoop