机器学习 陈昱 北京大学计算机科学技术研究所 信息安全工程研究中心. Concept Learning Reference : Ch2 in Mitchell’s book 1. Concepts: Inductive learning hypothesis General-to-specific.

Slides:



Advertisements
Similar presentations
Thomas Trappenberg Autonomous Robotics: Supervised and unsupervised learning.
Advertisements

Slides from: Doug Gray, David Poole
Rutgers CS440, Fall 2003 Review session. Rutgers CS440, Fall 2003 Topics Final will cover the following topics (after midterm): 1.Uncertainty & introduction.
Intelligent Environments1 Computer Science and Engineering University of Texas at Arlington.
Supervised Learning Recap
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
Bayesian Learning Rong Jin. Outline MAP learning vs. ML learning Minimum description length principle Bayes optimal classifier Bagging.
CII504 Intelligent Engine © 2005 Irfan Subakti Department of Informatics Institute Technology of Sepuluh Nopember Surabaya - Indonesia.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Artificial Intelligence Course review AIMA. Four main themes Problem solving by search Uninformed search Informed search Constraint satisfaction Adversarial.
Cooperating Intelligent Systems Course review AIMA.
COMP 328: Midterm Review Spring 2010 Nevin L. Zhang Department of Computer Science & Engineering The Hong Kong University of Science & Technology
Announcements  Project proposal is due on 03/11  Three seminars this Friday (EB 3105) Dealing with Indefinite Representations in Pattern Recognition.
Measuring Model Complexity (Textbook, Sections ) CS 410/510 Thurs. April 27, 2007 Given two hypotheses (models) that correctly classify the training.
Computational Learning Theory
Neural Networks Marco Loog.
Programme in Statistics (Courses and Contents). Elementary Probability and Statistics (I) 3(2+1)Stat. 101 College of Science, Computer Science, Education.
Sketched Derivation of error bound using VC-dimension (1) Bound our usual PAC expression by the probability that an algorithm has 0 error on the training.
1er. Escuela Red ProTIC - Tandil, de Abril, Bayesian Learning 5.1 Introduction –Bayesian learning algorithms calculate explicit probabilities.
Bayesian Learning Rong Jin.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Online Learning Algorithms
Crash Course on Machine Learning
1 Machine Learning: Lecture 5 Experimental Evaluation of Learning Algorithms (Based on Chapter 5 of Mitchell T.., Machine Learning, 1997)
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Tuesday 15 October 2002 William.
COMP3503 Intro to Inductive Modeling
Midterm Review Rao Vemuri 16 Oct Posing a Machine Learning Problem Experience Table – Each row is an instance – Each column is an attribute/feature.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Machine Learning Lecture 11 Summary G53MLE | Machine Learning | Dr Guoping Qiu1.
ECE 8443 – Pattern Recognition Objectives: Error Bounds Complexity Theory PAC Learning PAC Bound Margin Classifiers Resources: D.M.: Simplified PAC-Bayes.
机器学习 陈昱 北京大学计算机科学技术研究所 信息安全工程研究中心. 课程基本信息  主讲教师:陈昱 Tel :  助教:程再兴, Tel :  课程网页:
机器学习 陈昱 北京大学计算机科学技术研究所 信息安全工程研究中心. 课程基本信息  主讲教师:陈昱 Tel :  助教:程再兴, Tel :  课程网页:
CpSc 881: Machine Learning Evaluating Hypotheses.
机器学习 陈昱 北京大学计算机科学技术研究所 信息安全工程研究中心. 课程基本信息  主讲教师:陈昱 Tel :  助教:程再兴, Tel :  课程网页:
机器学习 陈昱 北京大学计算机科学技术研究所 信息安全工程研究中心. 课程基本信息  主讲教师:陈昱 Tel :  助教:程再兴, Tel :  课程网页:
Concept learning, Regression Adapted from slides from Alpaydin’s book and slides by Professor Doina Precup, Mcgill University.
BCS547 Neural Decoding.
Sparse Kernel Methods 1 Sparse Kernel Methods for Classification and Regression October 17, 2007 Kyungchul Park SKKU.
机器学习 陈昱 北京大学计算机科学技术研究所 信息安全工程研究中心. 课程基本信息  主讲教师:陈昱 Tel :  助教:程再兴, Tel :  课程网页:
Chapter5: Evaluating Hypothesis. 개요 개요 Evaluating the accuracy of hypotheses is fundamental to ML. - to decide whether to use this hypothesis - integral.
Insight: Steal from Existing Supervised Learning Methods! Training = {X,Y} Error = target output – actual output.
Lecture 2: Statistical learning primer for biologists
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall DM Finals Study Guide Rodney Nielsen.
Data Mining and Decision Support
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Thursday, 08 March 2007 William.
CS 8751 ML & KDDComputational Learning Theory1 Notions of interest: efficiency, accuracy, complexity Probably, Approximately Correct (PAC) Learning Agnostic.
1 Machine Learning: Lecture 6 Bayesian Learning (Based on Chapter 6 of Mitchell T.., Machine Learning, 1997)
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
FNA/Spring CENG 562 – Machine Learning. FNA/Spring Contact information Instructor: Dr. Ferda N. Alpaslan
CMPS 142/242 Review Section Fall 2011 Adapted from Lecture Slides.
1 1)Bayes’ Theorem 2)MAP, ML Hypothesis 3)Bayes optimal & Naïve Bayes classifiers IES 511 Machine Learning Dr. Türker İnce (Lecture notes by Prof. T. M.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Usman Roshan Dept. of Computer Science NJIT
Who am I? Work in Probabilistic Machine Learning Like to teach 
ICS 280 Learning in Graphical Models
Machine Learning for dotNET Developer Bahrudin Hrnjica, MVP
Data Mining Lecture 11.
Machine Learning Today: Reading: Maria Florina Balcan
Computational Learning Theory
Computational Learning Theory
Machine Learning: Lecture 6
Machine Learning: UNIT-3 CHAPTER-1
Usman Roshan Dept. of Computer Science NJIT
A task of induction to find patterns
Chapter 14 February 26, 2004.
Presentation transcript:

机器学习 陈昱 北京大学计算机科学技术研究所 信息安全工程研究中心

Concept Learning Reference : Ch2 in Mitchell’s book 1. Concepts: Inductive learning hypothesis General-to-specific ordering of hypotheses Consistent hypothesis Version space Inductive bias, e.g. what is the inductive bias of Candidate- Elimination algorithm? 2. Algorithms: Find-S Candidate-Elimination 3. Section & 2.6.3

Concept Learning (2)  Sample problems: 2.2, 2.4, 2.7.

Decision Tree Reference: Ch3 in Mitchell’s book 1. Concepts & Facts Entropy and information gain (Approximate) inductive bias of ID3 Overfitting Ockham’s razor 2. ID3 algorithm for decision tree learning Sample problem: 3.4

Artificial Neural Network Reference: Ch 4.3—4.6 in Mitchell’s book Focus on the following:  Perceptron (Fig 4.2)  Gradient Descent (Sec 4.4.3, exercise 4.5)  Understand how a backpropagation algorithm works  Representation power of perceptron and more generally, feedforward network Sample problem: Given an error function, derive the gradient descent update rule, such as 4.10.

Evaluating Hypotheses Reference: Ch5 in Mitchell’s book 1. Concepts & Theorem Sample error, true error Confidence intervals for observed hypothesis error, in particular, section in the book Estimator, Bias, and variance (Section 5.3.4) Binomial distribution, normal distribution Sample problems: 5.3 & 5.4 (just computation), 5.6 (concept)

Bayesian Learning Reference: Ch6 in Mitchell’s book 1. Concepts and Theorem Bayes Theorem ML and MAP hypotheses Minimum Description Length Principle m-estimate of probability (6.22) 2. Algorithm Bayes Optimal Classifier Naive Bayes Classifier Sample problem: 6.1, 6.5, and the PlayTennis example in 6.9.1

Computational Learning Theory Reference: Ch7 in Mitchell’s book 1. Concepts and Theorems PAC-learnable, consistent learner, ε-exhausted Theorem 7.1 Theorem 7.2 (it provides an example of PAC-learnable) VC-dimension: concept and computation, such as exercise 7.5 Mistake bound, focus estimating mistake bound on some simple algorithms, such as those discussed in and Sample problems: 7.2, 7.5, 7.8

Kernel Methods Reference: Sec 3.1, 6.1 & 6.2 in Bishop’s book 1. Sec 6.1 Linear regression model, regularized least square Dual representation, Gram matrix, kernel function 2. Sec 6.2 Study everything except Fisher kernel Sample problem Exercise 6.10, 6.12 Verify some function is indeed a kernel, such as (6.23)

Maximum Margin Classifier Covers up to (7.37) + basic knowledge of later part of Sec 7.1 of Bishop’s book  Should have an overall picture of the main steps in the procedure of deriving the maximum margin classifier, including that for handling overlapping class distribution  Understand the following concepts Maximum margin, KKT condition, support vector  Should be able to solve problems with level of difficulty same to 7.3 & 7.4

Probabilistic Graphic Network Reference: Sec 8.1.1, 8.1.3, basic knowledge of Sec 8.3 & 8.4 of Bishop’s book, with focus on Sec 8.2  Concepts: Conditional independence, D-separation, clique, factor graph, tree, polytree.  Sample problems: Burglary example, “out-of-fuel” example, exercise 8.11

Hidden Markov Model Reference: course slides Sample problem: homework assignment of Ch11.