A Brief Survey of Machine Learning

Slides:



Advertisements
Similar presentations
Machine learning Overview
Advertisements

1 Machine Learning: Lecture 1 Overview of Machine Learning (Based on Chapter 1 of Mitchell T.., Machine Learning, 1997)
1 Machine Learning Spring 2010 Rong Jin. 2 CSE847 Machine Learning Instructor: Rong Jin Office Hour: Tuesday 4:00pm-5:00pm Thursday 4:00pm-5:00pm Textbook.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Introduction to Machine Learning Alejandro Ceccatto Instituto de Física Rosario CONICET-UNR.
Data Mining Classification: Alternative Techniques
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Tuesday, August 24, 1999 William.
CS 484 – Artificial Intelligence1 Announcements Project 1 is due Tuesday, October 16 Send me the name of your konane bot Midterm is Thursday, October 18.
1er. Escuela Red ProTIC - Tandil, de Abril, Introduction How to program computers to learn? Learning: Improving automatically with experience.
1 Machine Learning Spring 2013 Rong Jin. 2 CSE847 Machine Learning  Instructor: Rong Jin  Office Hour: Tuesday 4:00pm-5:00pm TA, Qiaozi Gao, Thursday.
CII504 Intelligent Engine © 2005 Irfan Subakti Department of Informatics Institute Technology of Sepuluh Nopember Surabaya - Indonesia.
Machine Learning Bob Durrant School of Computer Science
Machine Learning (Extended) Dr. Ata Kaban
1 Machine Learning Spring 2010 Rong Jin. 2 CSE847 Machine Learning  Instructor: Rong Jin  Office Hour: Tuesday 4:00pm-5:00pm Thursday 4:00pm-5:00pm.
Machine Learning Group University College Dublin 4.30 Machine Learning Pádraig Cunningham.
Data Mining with Decision Trees Lutz Hamel Dept. of Computer Science and Statistics University of Rhode Island.
Machine Learning CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanning.
1 Some rules  No make-up exams ! If you miss with an official excuse, you get average of your scores in the other exams – at most once.  WP only-if you.
A Brief Survey of Machine Learning Used materials from: William H. Hsu Linda Jackson Lex Lane Tom Mitchell Machine Learning, Mc Graw Hill 1997 Allan Moser.
Learning Programs Danielle and Joseph Bennett (and Lorelei) 4 December 2007.
Lecture 32 of 42 Machine Learning: Basic Concepts
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Lecture 0 Friday, January.
Short Introduction to Machine Learning Instructor: Rada Mihalcea.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Friday, 12 January 2007 William.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
CpSc 810: Machine Learning Design a learning system.
Computing & Information Sciences Kansas State University Wednesday, 13 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 9 of 42 Wednesday, 13 September.
Computing & Information Sciences Kansas State University Friday, 18 Jan 2008CIS 732 / 830: Machine Learning / Advanced Topics in AI Lecture 0 of 42 Friday,
Machine Learning Introduction. 2 교재  Machine Learning, Tom T. Mitchell, McGraw- Hill  일부  Reinforcement Learning: An Introduction, R. S. Sutton and.
Kansas State University Department of Computing and Information Sciences CIS 690: Data Mining Systems Lecture 0 Monday, May 15, 2000 William H. Hsu Department.
Computing & Information Sciences Kansas State University Lecture 10 of 42 CIS 530 / 730 Artificial Intelligence Lecture 10 of 42 William H. Hsu Department.
1 Mining in geographic data Original slides:Raymond J. Mooney University of Texas at Austin.
Introduction to Artificial Intelligence and Soft Computing
1 Machine Learning (Extended) Dr. Ata Kaban Algorithms to enable computers to learn –Learning = ability to improve performance automatically through experience.
Machine Learning (ML) and Knowledge Discovery in Databases (KDD) Instructor: Rich Maclin Texts: Machine Learning, Mitchell Notes based.
Well Posed Learning Problems Must identify the following 3 features –Learning Task: the thing you want to learn. –Performance measure: must know when you.
Computing & Information Sciences Kansas State University Friday, 10 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 33 of 42 Friday, 10 November.
1 Machine Learning 1.Where does machine learning fit in computer science? 2.What is machine learning? 3.Where can machine learning be applied? 4.Should.
Computing & Information Sciences Kansas State University Wednesday, 12 Sep 2007CIS 530 / 730: Artificial Intelligence Lecture 9 of 42 Wednesday, 12 September.
Machine Learning, Decision Trees, Overfitting Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 14,
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 9 of 42 Wednesday, 14.
Chapter 1: Introduction. 2 목 차목 차 t Definition and Applications of Machine t Designing a Learning System  Choosing the Training Experience  Choosing.
Kansas State University Department of Computing and Information Sciences CIS 690: Data Mining Systems Lab 0 Monday, May 15, 2000 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 34 of 41 Wednesday, 10.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Friday, 14 November 2003 William.
Course Overview  What is AI?  What are the Major Challenges?  What are the Main Techniques?  Where are we failing, and why?  Step back and look at.
Machine Learning Introduction. Class Info Office Hours –Monday:11:30 – 1:00 –Wednesday:10:00 – 1:00 –Thursday:11:30 – 1:00 Course Text –Tom Mitchell:
Data Mining and Decision Support
1 Introduction to Machine Learning Chapter 1. cont.
Introduction Machine Learning: Chapter 1. Contents Types of learning Applications of machine learning Disciplines related with machine learning Well-posed.
Well Posed Learning Problems Must identify the following 3 features –Learning Task: the thing you want to learn. –Performance measure: must know when you.
Computing & Information Sciences Kansas State University CIS 530 / 730: Artificial Intelligence Lecture 09 of 42 Wednesday, 17 September 2008 William H.
Machine Learning & Datamining CSE 454. © Daniel S. Weld 2 Project Part 1 Feedback Serialization Java Supplied vs. Manual.
FNA/Spring CENG 562 – Machine Learning. FNA/Spring Contact information Instructor: Dr. Ferda N. Alpaslan
1 Machine Learning Patricia J Riddle Computer Science 367 6/26/2016Machine Learning.
Machine Learning. Definition: The ability of a machine to improve its performance based on previous results.
Supervise Learning Introduction. What is Learning Problem Learning = Improving with experience at some task – Improve over task T, – With respect to performance.
Brief Intro to Machine Learning CS539
School of Computer Science & Engineering
Spring 2003 Dr. Susan Bridges
Done Done Course Overview What is AI? What are the Major Challenges?
Basic Intro Tutorial on Machine Learning and Data Mining
A Brief Survey of Machine Learning
Introduction to Artificial Intelligence and Soft Computing
Overview of Machine Learning
3.1.1 Introduction to Machine Learning
Why Machine Learning Flood of data
Machine Learning: Lecture 6
Introduction to Artificial Intelligence Instructor: Dr. Eduardo Urbina
Machine Learning (ML) and Knowledge Discovery in Databases (KDD)
Presentation transcript:

A Brief Survey of Machine Learning

Lecture Outline Why machine learning? Brief Tour of Machine Learning A case study A taxonomy of learning Intelligent systems engineering: specification of learning problems Issues in Machine Learning Design choices The performance element: intelligent systems Some Applications of Learning Database mining, reasoning (inference/decision support), acting Industrial usage of intelligent systems Robotics

Overview Learning Algorithms and Models Theory of Learning Models: decision trees, linear threshold units (winnow, weighted majority), neural networks, Bayesian networks (polytrees, belief networks, influence diagrams, HMMs), genetic algorithms, instance-based (nearest-neighbor) Algorithms (e.g., for decision trees): ID3, C4.5, CART, OC1 Methodologies: supervised, unsupervised, reinforcement; knowledge-guided Theory of Learning Computational learning theory (COLT): complexity, limitations of learning Probably Approximately Correct (PAC) learning Probabilistic, statistical, information theoretic results Multistrategy Learning: Combining Techniques, Knowledge Sources Data: Time Series, Very Large Databases (VLDB), Text Corpora Applications Performance element: classification, decision support, planning, control Database mining and knowledge discovery in databases (KDD) Computer inference: learning to reason

Why Machine Learning? New Computational Capability Database mining: converting (technical) records into knowledge Self-customizing programs: learning news filters, adaptive monitors Learning to act: robot planning, control optimization, decision support Applications that are hard to program: automated driving, speech recognition Better Understanding of Human Learning and Teaching Cognitive science: theories of knowledge acquisition (e.g., through practice) Performance elements: reasoning (inference) and recommender systems Time is Right Recent progress in algorithms and theory Rapidly growing volume of online data from various sources Available computational power Growth and interest of learning-based industries (e.g., data mining/KDD)

Rule and Decision Tree Learning Example: Rule Acquisition from Historical Data Data Patient 103 (time = 1): Age 23, First-Pregnancy: no, Anemia: no, Diabetes: no, Previous-Premature-Birth: no, Ultrasound: unknown, Elective C-Section: unknown, Emergency-C-Section: unknown Patient 103 (time = 2): Age 23, First-Pregnancy: no, Anemia: no, Diabetes: yes, Previous-Premature-Birth: no, Ultrasound: abnormal, Elective C-Section: no, Emergency-C-Section: unknown Patient 103 (time = n): Age 23, First-Pregnancy: no, Anemia: no, Diabetes: no, Previous-Premature-Birth: no, Ultrasound: unknown, Elective C-Section: no, Emergency-C-Section: YES Learned Rule IF no previous vaginal delivery, AND abnormal 2nd trimester ultrasound, AND malpresentation at admission, AND no elective C-Section THEN probability of emergency C-Section is 0.6 Training set: 26/41 = 0.634 Test set: 12/20 = 0.600

Neural Network Learning Autonomous Learning Vehicle In a Neural Net (ALVINN): Pomerleau et al http://www.cs.cmu.edu/afs/cs/project/alv/member/www/projects/ALVINN.html Drives 70mph on highways

Relevant Disciplines Machine Learning Artificial Intelligence Bayesian Methods Cognitive Science Computational Complexity Theory Control Theory Information Theory Neuroscience Philosophy Psychology Statistics Optimization Learning Predictors Meta-Learning Entropy Measures MDL Approaches Optimal Codes PAC Formalism Mistake Bounds Language Learning Learning to Reason Machine Learning Bayes’s Theorem Missing Data Estimators Symbolic Representation Planning/Problem Solving Knowledge-Guided Learning Bias/Variance Formalism Confidence Intervals Hypothesis Testing Power Law of Practice Heuristic Learning Occam’s Razor Inductive Generalization ANN Models Modular Learning

Specifying A Learning Problem Learning = Improving with Experience at Some Task Improve over task T, with respect to performance measure P, based on experience E. Example: Learning to Play Checkers T: play games of checkers P: percent of games won in world tournament E: opportunity to play against self Refining the Problem Specification: Issues What experience? What exactly should be learned? How shall it be represented? What specific algorithm to learn it? Defining the Problem Milieu Performance element: How shall the results of learning be applied? How shall the performance element be evaluated? The learning system?

Example: Learning to Play Checkers Type of Training Experience Direct or indirect? Teacher or not? Knowledge about the game (e.g., openings/endgames)? Problem: Is Training Experience Representative (of Performance Goal)? Software Design Assumptions of the learning system: legal move generator exists Software requirements: generator, evaluator(s), parametric target function Choosing a Target Function ChooseMove: Board  Move (action selection function, or policy) V: Board  R (board evaluation function) Ideal target V; approximated target Goal of learning process: operational description (approximation) of V

A Target Function for Learning to Play Checkers Possible Definition If b is a final board state that is won, then V(b) = 100 If b is a final board state that is lost, then V(b) = -100 If b is a final board state that is drawn, then V(b) = 0 If b is not a final board state in the game, then V(b) = V(b’) where b’ is the best final board state that can be achieved starting from b and playing optimally until the end of the game Correct values, but not operational Choosing a Representation for the Target Function Collection of rules? Neural network? Polynomial function (e.g., linear, quadratic combination) of board features? Other? A Representation for Learned Function bp/rp = number of black/red pieces; bk/rk = number of black/red kings; bt/rt = number of black/red pieces threatened (can be taken on next turn)

A Training Procedure for Learning to Play Checkers Obtaining Training Examples the target function the learned function the training value One Rule For Estimating Training Values: Choose Weight Tuning Rule Least Mean Square (LMS) weight update rule: REPEAT Select a training example b at random Compute the error(b) for this training example For each board feature fi, update weight wi as follows: where c is a small, constant factor to adjust the learning rate

Design Choices for Learning to Play Checkers Determine Type of Training Experience Games against experts against self Table of correct moves Determine Target Function Board  value Board  move Determine Representation of Learned Function Polynomial Linear function of six features Artificial neural network Determine Learning Algorithm Gradient descent Linear programming Completed Design

Some Issues in Machine Learning What Algorithms Can Approximate Functions Well? When? How Do Learning System Design Factors Influence Accuracy? Number of training examples Complexity of hypothesis representation How Do Learning Problem Characteristics Influence Accuracy? Noisy data Multiple data sources What Are The Theoretical Limits of Learnability? How Can Prior Knowledge of Learner Help? What Clues Can We Get From Biological Learning Systems? How Can Systems Alter Their Own Representation?

Interesting Applications Database Mining NCSA D2K - http://www.ncsa.uiuc.edu/STI/ALG Reasoning (Inference, Decision Support) Cartia ThemeScapes - http://www.cartia.com 6500 news stories from the WWW in 1997 Planning, Control Normal Ignited Engulfed Destroyed Extinguished Fire Alarm Flooding DC-ARM - http://www-kbs.ai.uiuc.edu

Department of Computing and Information Sciences, KSU Material adapted from William H. Hsu Department of Computing and Information Sciences, KSU http://www.kddresearch.org http://www.cis.ksu.edu/~bhsu Readings: Chapter 1, Mitchell