Machine Learning Introduction Study on the Coursera All Right Reserved : Andrew Ng Lecturer:Much Database Lab of Xiamen University Aug 12,2014.

Slides:



Advertisements
Similar presentations
Introduction to Machine Learning BITS C464/BITS F464
Advertisements

Unsupervised Learning Clustering K-Means. Recall: Key Components of Intelligent Agents Representation Language: Graph, Bayes Nets, Linear functions Inference.
Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
Machine learning continued Image source:
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Computational Methods for Data Analysis
Indian Statistical Institute Kolkata
Logistic Regression Classification Machine Learning.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Linear Learning Machines  Simplest case: the decision function is a hyperplane in input space.  The Perceptron Algorithm: Rosenblatt, 1956  An on-line.
Linear Learning Machines  Simplest case: the decision function is a hyperplane in input space.  The Perceptron Algorithm: Rosenblatt, 1956  An on-line.
Rise of the Machines Jack Concanon
CS 4700: Foundations of Artificial Intelligence
Introduction to machine learning
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Collaborative Filtering Matrix Factorization Approach
Machine learning Image source:
More Machine Learning Linear Regression Squared Error L1 and L2 Regularization Gradient Descent.
Lecture 2: Introduction to Machine Learning
Learning with large datasets Machine Learning Large scale machine learning.
Data Mining Joyeeta Dutta-Moscato July 10, Wherever we have large amounts of data, we have the need for building systems capable of learning information.
Logistic Regression L1, L2 Norm Summary and addition to Andrew Ng’s lectures on machine learning.
1 Artificial Neural Networks Sanun Srisuk EECP0720 Expert Systems – Artificial Neural Networks.
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Model representation Linear regression with one variable
Andrew Ng Linear regression with one variable Model representation Machine Learning.
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
M Machine Learning F# and Accord.net. Alena Dzenisenka Software architect at Luxoft Poland Member of F# Software Foundation Board of Trustees Researcher.
Logistic Regression Week 3 – Soft Computing By Yosi Kristian.
Ensemble Learning Spring 2009 Ben-Gurion University of the Negev.
Week 1 - An Introduction to Machine Learning & Soft Computing
Linear Classification with Perceptrons
The problem of overfitting
Regularization (Additional)
M Machine Learning F# and Accord.net.
Introduction Welcome Machine Learning.
Machine Learning Introduction. Class Info Office Hours –Monday:11:30 – 1:00 –Wednesday:10:00 – 1:00 –Thursday:11:30 – 1:00 Course Text –Tom Mitchell:
Chapter 2-OPTIMIZATION
Introduction to Machine Learning. Learning Learning is acquiring new, or modifying existing, knowledge, behaviors, skills, values, or preferences and.
Introduction to Machine Learning. Introduce yourself Why you choose this course? Any examples of machine learning you know?
Page 1 CS 546 Machine Learning in NLP Review 1: Supervised Learning, Binary Classifiers Dan Roth Department of Computer Science University of Illinois.
語音訊號處理之初步實驗 NTU Speech Lab 指導教授: 李琳山 助教: 熊信寬
WEEK 2 SOFT COMPUTING & MACHINE LEARNING YOSI KRISTIAN Gradient Descent for Linear Regression.
Dan Roth Department of Computer and Information Science
Deep Learning Amin Sobhani.
Machine Learning & Deep Learning
Introduction Machine Learning 14/02/2017.
Lecture 3: Linear Regression (with One Variable)
Machine Learning I & II.
Estimating Link Signatures with Machine Learning Algorithms
Logistic Regression Classification Machine Learning.
Logistic Regression Classification Machine Learning.
Collaborative Filtering Matrix Factorization Approach
Logistic Regression Classification Machine Learning.
Artificial Intelligence Chapter 3 Neural Networks
Overview of Machine Learning
Artificial Intelligence Chapter 3 Neural Networks
Machine Learning Algorithms – An Overview
CS639: Data Management for Data Science
What is machine learning
Multiple features Linear Regression with multiple variables
Multiple features Linear Regression with multiple variables
Logistic Regression Classification Machine Learning.
Linear regression with one variable
Introduction Welcome Machine Learning.
Logistic Regression Classification Machine Learning.
Artificial Intelligence Chapter 3 Neural Networks
Presentation transcript:

Machine Learning Introduction Study on the Coursera All Right Reserved : Andrew Ng Lecturer:Much Database Lab of Xiamen University Aug 12,2014

Examples: - Database mining Large datasets from growth of automation/web. Web click data, medical records, biology, engineering - Applications can ’ t program by hand. Handwriting recognition, most of Natural Language Processing (NLP), Computer Vision. Machine Learning - Grew out of work in AI(Artificial Intelligence) - New capability for computers

Machine Learning Definition Tom Mitchell (1998) Well-posed Learning Problem: A computer program is said to learn from experience E with respect to some task T and some performance measure P, if its performance on T, as measured by P, improves with experience E.

Suppose your program watches which s you do or do not mark as spam, and based on that learns how to better filter spam. What is the task T in this setting? T : Classifying s as spam or not spam E : Watching you label s as spam or not spam P: The number of s correctly classified as spam/not spam “ A computer program is said to learn from experience E with respect to some task T and some performance measure P, if its performance on T, as measured by P, improves with experience E. ”

Machine Learning Algorithms -Supervised learning -Unsupervised learning -Others: - Reinforcement learning - Recommender systems.

x1x1 x2x2 Supervised Learning & Unsupervised Learning Supervised LearningUnsupervised Learning

Linear Regression with one Variable Housing Prices (Portland, OR) Price (in 1000s of dollars) Size (feet 2 ) Supervised Learning Given the “ right answer ” for each example in the data. Regression Problem Predict real-valued output

Notation: m = Number of training examples x ’ s = “ input ” variable / features y ’ s = “ output ” variable / “ target ” variable Size in feet 2 (x)Price ($) in 1000's (y) …… Training set of housing prices Training Set Learning Algorithm h Size of house Estimated price Question : How to describe h?

How to choose ‘ s ? Training Set Hypothesis: ‘ s: Parameters Size in feet 2 (x)Price ($) in 1000's (y) ……

y x Idea: Choose so that is close to for our training examples

Cost Function Hypothesis: Parameters: Cost Function: Goal: Simplified:

Price ($) in 1000 ’ s Size in feet 2 (x) Question:How to minimize J?

Gradient Descent Have some function Want Outline: Start with some Keep changing to reduce until we hopefully end up at a minimum

Gradient descent algorithm Correct: Simultaneous update Incorrect:

Gradient descent algorithm Notice : α is the learning rate.

If α is too small, gradient descent can be slow. If α is too large, gradient descent can overshoot the minimum. It may fail to converge, or even diverge.

at local optima Current value of Unchange Gradient descent can converge to a local minimum, even with the learning rate α fixed. As we approach a local minimum, gradient descent will automatically take smaller steps. So, no need to decrease α over time.

Gradient Descent for Linear Regression Gradient descent algorithm Linear Regression Model

Gradient descent algorithm update and simultaneously

  J(     )

(for fixed, this is a function of x)(function of the parameters )

(for fixed, this is a function of x)(function of the parameters )

(for fixed, this is a function of x)(function of the parameters )

(for fixed, this is a function of x)(function of the parameters )

(for fixed, this is a function of x)(function of the parameters )

(for fixed, this is a function of x)(function of the parameters )

(for fixed, this is a function of x)(function of the parameters )

(for fixed, this is a function of x)(function of the parameters )

(for fixed, this is a function of x)(function of the parameters )

Linear Regression with multiple variables Hypothesis: Cost function: Parameters: (simultaneously update for every ) Repeat Gradient descent:

(simultaneously update ) Gradient Descent Repeat Previously (n=1): New algorithm : Repeat (simultaneously update for )

Size (feet 2 )Number of bedrooms Number of floors Age of home (years) Price ($1000) Size (feet 2 )Number of bedrooms Number of floors Age of home (years) Price ($1000) Examples: simultaneously update Size (feet 2 )Number of bedrooms Number of floors Age of home (years) Price ($1000)

Summarize This is a briefly Introduction about Supervised Learning(Classification)in Machine Leaning. There is still a lot of things in this subject,such as Clustering, Support Vector Machine(SVM), Dimensionality Reduction, ETC. The Core Idea of MS is very similar,hope you will be fond of the Machine Learning ! Thanks for Listening !