Download presentation
Presentation is loading. Please wait.
1
1 Machine Learning Spring 2013 Rong Jin
2
2 CSE847 Machine Learning Instructor: Rong Jin Office Hour: Tuesday 4:00pm-5:00pm TA, Qiaozi Gao, Thursday 4:00pm-5:00pm Textbook Machine Learning The Elements of Statistical Learning Pattern Recognition and Machine Learning Many subjects are from papers Web site: http://www.cse.msu.edu/~cse847
3
3 Requirements ~10 homework assignments Course project Topic: visual object recognition Data: over one million images with extracted visual features Objective: build a classifier that automatically identifies the class of objects in images Midterm exam & final exam
4
4 Goal Familiarize you with the state-of-art in Machine Learning Breadth: many different techniques Depth: Project Hands-on experience Develop the way of machine learning thinking Learn how to model real-world problems by machine learning techniques Learn how to deal with practical issues
5
5 Course Outline Theoretical Aspects Information Theory Optimization Theory Probability Theory Learning Theory Practical Aspects Supervised Learning Algorithms Unsupervised Learning Algorithms Important Practical Issues Applications
6
6 Today’s Topics Why is machine learning? Example: learning to play backgammon General issues in machine learning
7
7 Why Machine Learning? Past: most computer programs are mainly made by hand Future: Computers should be able to program themselves by the interaction with their environment
8
8 Recent Trends Recent progress in algorithm and theory Growing flood of online data Computational power is available Growing industry
9
Big Data Challenge 2.7 Zetabytes (10 21 ) of data exists in the digital universe today. Huge amount of data generated on the Internet every minute YouTube users upload 48 hours of video, Facebook users share 684,478 pieces of content, Instagram users share 3,600 new photos, http://www.visualnews.com/2012/06/19/how-much-data-created-every-minute/
10
Big Data Challenge High dimensional data appears in many applications of machine learning Fine grained visual classification [1] 250,000 features
11
Why Data Size Matters ? Matrix completion Classification, clustering, recommender systems
12
Why Data Size Matters ? Matrix can be perfectly recovered provided the number of observed entries O(rnlog 2 (n))
13
Why Data Size Matters ? The recovery error can be arbitrarily large if the number of observed entries < O(rnlog(n))
14
Why Data Size Matters ? error # observed entries O(rnlog (n)) O(rnlog 2 (n)) Unknown
15
Difficult to access finance for small & medium business Minimum loan Tedious loan approval procedure Low approval rate Long cycle Completely big data driven Leverage e-commerce data to financial services Alibaba Small and Micro Financial Services
16
Insurance contracts has year-on-year growth rate of 100%. Over 1 billion contracts in 2013 Over 100 million contracts one day on November 11, 2013 Shipping Insurance for Returned Products
17
Uniform 5% fixed rate Fixed rate Solely based on historical data and demographics Actuarial approach Simple Easy to explain Pricing model based on a few couple parameters Data based pricing Relatively accurate Millions of features, real time pricing Machine learned model Dynamic pricing Highly accurate Shipping Insurance for Returned Products
18
18 Three Niches for Machine Learning Data mining: using historical data to improve decisions Medical records medical knowledge Software applications that are difficult to program by hand Autonomous driving Image Classification User modeling Automatic recommender systems
19
19 Typical Data Mining Task Given: 9147 patient records, each describing pregnancy and birth Each patient contains 215 features Task: Classes of future patients at high risk for Emergency Cesarean Section
20
20 Data Mining Results One of 18 learned rules : If no previous vaginal delivery abnormal 2 nd Trimester Ultrasound Malpresentation at admission Then probability of Emergency C-Section is 0.6
21
21 Credit Risk Analysis Learned Rules : If Other-Delinquent-Account > 2 Number-Delinquent-Billing-Cycles > 1 ThenProfitable-Costumer ? = no IfOther-Delinquent-Account = 0 (Income > $30K or Years-of-Credit > 3) ThenProfitable-Costumer ? = yes
22
22 Programs too Difficult to Program By Hand ALVINN drives 70mph on highways
23
23 Programs too Difficult to Program By Hand ALVINN drives 70mph on highways
24
24 Programs too Difficult to Program By Hand Visual object recognition
25
25 Image Retrieval using Texts
26
26 Software that Models Users Description: A homicide detective and a fire marshall must stop a pair of murderers who commit videotaped crimes to become media darlings Rating: Description: Benjamin Martin is drawn into the American revolutionary war against his will when a brutal British commander kills his son. Rating: Description: A biography of sports legend, Muhammad Ali, from his early days to his days in the ring Rating: History What to Recommend? Description: A high-school boy is given the chance to write a story about an up-and-coming rock band as he accompanies it on their concert tour. Recommend: ? Description: A young adventurer named Milo Thatch joins an intrepid group of explorers to find the mysterious lost continent of Atlantis. Recommend: ? No Yes
27
27 Netflix Contest
28
28 Relevant Disciplines Artificial Intelligence Statistics (particularly Bayesian Stat.) Computational complexity theory Information theory Optimization theory Philosophy Psychology …
29
29 Today’s Topics Why is machine learning? Example: learning to play backgammon General issues in machine learning
30
30 What is the Learning Problem Learning = Improving with experience at some task Improve over task T With respect to performance measure P Based on experience E Example: Learning to Play Backgammon T: Play backgammon P: % of games won in world tournament E: opportunity to play against itself
31
31 Backgammon More than 10 20 states (boards) Best human players see only small fraction of all board during lifetime Searching is hard because of dice (branching factor > 100)
32
32 TD-Gammon by Tesauro (1995) Trained by playing with itself Now approximately equal to the best human player
33
33 Learn to Play Chess Task T: Play chess Performance P: Percent of games won in the world tournament Experience E: What experience? How shall it be represented? What exactly should be learned? What specific algorithm to learn it?
34
34 Choose a Target Function Goal: Policy: : b m Choice of value function V: b, m B = board = real values
35
35 Choose a Target Function Goal: Policy: : b m Choice of value function V: b, m V: b B = board = real values
36
36 Value Function V(b): Example Definition If b final board that is won: V(b) = 1 If b final board that is lost: V(b) = -1 If b not final boardV(b) = E[V(b*)] where b* is final board after playing optimally
37
37 Representation of Target Function V(b) Same value for each board Lookup table (one entry for each board) No Learning No Generalization Summarize experience into Polynomials Neural Networks
38
38 Example: Linear Feature Representation Features: p b (b), p w (b) = number of black (white) pieces on board b u b (b), u b (b) = number of unprotected pieces t b (b), t b (b) = number of pieces threatened by opponent Linear function: V(b) = w 0 p b (b)+ w 1 p w (b)+ w 2 u b (b)+ w 3 u w (b)+ w 4 t b (b)+ w 5 t w (b) Learning: Estimation of parameters w 0, …, w 5
39
39 Given: board b Predicted value V(b) Desired value V*(b) Calculate error(b) = (V*(b) – V(b)) 2 For each board feature f i w i w i + c error(b) f i Stochastically minimizes b (V*(b)-V(b)) 2 Tuning Weights Gradient Descent Optimization
40
40 Obtain Boards Random boards Beginner plays Professionals plays
41
41 Obtain Target Values Person provides value V(b) Play until termination. If outcome is Win: V(b) 1 for all boards Loss: V(b) -1 for all boards Draw: V(b) 0 for all boards Play one move: b b’ V(b) V(b’) Play n moves: b b’ … b (n) V(b) V(b (n) )
42
42 A General Framework MathematicalM odeling Finding Optimal Parameters StatisticsOptimization + Machine Learning
43
43 Today’s Topics Why is machine learning? Example: learning to play backgammon General issues in machine learning
44
44 Importants Issues in Machine Learning Obtaining experience How to obtain experience? Supervised learning vs. Unsupervised learning How many examples are enough? PAC learning theory Learning algorithms What algorithm can approximate function well, when? How does the complexity of learning algorithms impact the learning accuracy? Whether the target function is learnable? Representing inputs How to represent the inputs? How to remove the irrelevant information from the input representation? How to reduce the redundancy of the input representation?
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.