Patch Based Prediction Techniques University of Houston By: Paul AMALAMAN From: UH-DMML Lab Director: Dr. Eick.

Slides:



Advertisements
Similar presentations
Random Forest Predrag Radenković 3237/10
Advertisements

Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
Hierarchical Clustering. Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram – A tree-like diagram that.
CHAPTER 9: Decision Trees
1 An Adaptive GA for Multi Objective Flexible Manufacturing Systems A. Younes, H. Ghenniwa, S. Areibi uoguelph.ca.
Lectures 17,18 – Boosting and Additive Trees Rice ECE697 Farinaz Koushanfar Fall 2006.
Fast Algorithms For Hierarchical Range Histogram Constructions
Combining Classification and Model Trees for Handling Ordinal Problems D. Anyfantis, M. Karagiannopoulos S. B. Kotsiantis, P. E. Pintelas Educational Software.
Data Mining Classification: Alternative Techniques
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
SLIQ: A Fast Scalable Classifier for Data Mining Manish Mehta, Rakesh Agrawal, Jorma Rissanen Presentation by: Vladan Radosavljevic.
Evolutionary Computational Intelligence Lecture 10a: Surrogate Assisted Ferrante Neri University of Jyväskylä.
Ensemble Learning: An Introduction
Backtracking Reading Material: Chapter 13, Sections 1, 2, 4, and 5.
Parallel K-Means Clustering Based on MapReduce The Key Laboratory of Intelligent Information Processing, Chinese Academy of Sciences Weizhong Zhao, Huifang.
© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
Intelligible Models for Classification and Regression
Classification and Prediction: Regression Analysis
Ensemble Learning (2), Tree and Forest
Dependency networks Sushmita Roy BMI/CS 576 Nov 26 th, 2013.
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
Attention Deficit Hyperactivity Disorder (ADHD) Student Classification Using Genetic Algorithm and Artificial Neural Network S. Yenaeng 1, S. Saelee 2.
Module 04: Algorithms Topic 07: Instance-Based Learning
Lecture Notes 4 Pruning Zhangxi Lin ISQS
Predicting performance of applications and infrastructures Tania Lorido 27th May 2011.
COMP3503 Intro to Inductive Modeling
Linear Regression James H. Steiger. Regression – The General Setup You have a set of data on two variables, X and Y, represented in a scatter plot. You.
Engineering Statistics ENGR 592 Prepared by: Mariam El-Maghraby Date: 26/05/04 Design of Experiments Plackett-Burman Box-Behnken.
GATree: Genetically Evolved Decision Trees 전자전기컴퓨터공학과 데이터베이스 연구실 G 김태종.
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
Jeff Howbert Introduction to Machine Learning Winter Regression Linear Regression.
Comparing Univariate and Multivariate Decision Trees Olcay Taner Yıldız Ethem Alpaydın Department of Computer Engineering Bogazici University
Look-ahead Linear Regression Trees (LLRT)
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
A Novel Local Patch Framework for Fixing Supervised Learning Models Yilei Wang 1, Bingzheng Wei 2, Jun Yan 2, Yang Hu 2, Zhi-Hong Deng 1, Zheng Chen 2.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.7: Instance-Based Learning Rodney Nielsen.
Detecting Group Differences: Mining Contrast Sets Author: Stephen D. Bay Advisor: Dr. Hsu Graduate: Yan-Cheng Lin.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Copyright © 2010 SAS Institute Inc. All rights reserved. Decision Trees Using SAS Sylvain Tremblay SAS Canada – Education SAS Halifax Regional User Group.
Jeff Howbert Introduction to Machine Learning Winter Regression Linear Regression Regression Trees.
Christoph Eick: Learning Models to Predict and Classify 1 Learning from Examples Example of Learning from Examples  Classification: Is car x a family.
Apache Mahout Qiaodi Zhuang Xijing Zhang.
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
Using decision trees to build an a framework for multivariate time- series classification 1 Present By Xiayi Kuang.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
ICS 353: Design and Analysis of Algorithms Backtracking King Fahd University of Petroleum & Minerals Information & Computer Science Department.
Corresponding Clustering: An Approach to Cluster Multiple Related Spatial Datasets Vadeerat Rinsurongkawong and Christoph F. Eick Department of Computer.
Optimal Relay Placement for Indoor Sensor Networks Cuiyao Xue †, Yanmin Zhu †, Lei Ni †, Minglu Li †, Bo Li ‡ † Shanghai Jiao Tong University ‡ HK University.
Machine Learning in Practice Lecture 8 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Eco 6380 Predictive Analytics For Economists Spring 2016 Professor Tom Fomby Department of Economics SMU.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Overfitting, Bias/Variance tradeoff. 2 Content of the presentation Bias and variance definitions Parameters that influence bias and variance Bias and.
LECTURE 13: LINEAR MODEL SELECTION PT. 3 March 9, 2016 SDS 293 Machine Learning.
Christoph F. Eick Questions Review October 12, How does post decision tree post-pruning work? What is the purpose of applying post-pruning in decision.
Data Summit 2016 H104: Building Hadoop Applications Abhik Roy Database Technologies - Experian LinkedIn Profile:
Rodney Nielsen Many of these slides were adapted from: I. H. Witten, E. Frank and M. A. Hall Data Science Algorithms: The Basic Methods Clustering WFH:
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Chapter 7. Classification and Prediction
Introduction to Machine Learning and Tree Based Methods
Data Science Algorithms: The Basic Methods
Evaluating Techniques for Image Classification
Trees, bagging, boosting, and stacking
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
CSE 4705 Artificial Intelligence
Christoph Eick: Learning Models to Predict and Classify 1 Learning from Examples Example of Learning from Examples  Classification: Is car x a family.
Machine Learning: Lecture 3
GATree Genetically Evolved Decision Trees
MACHINE LEARNING IN ASIC BACKEND DESIGN
Presentation transcript:

Patch Based Prediction Techniques University of Houston By: Paul AMALAMAN From: UH-DMML Lab Director: Dr. Eick

Introduction 1. Research Goals 2. Problem Setting 3. Solutions: TPRTI-A & TPRTI-B 4. Results 5. Conclusion Future Work UH-DMML Lab 1

Research Goals To improve Machine Learning techniques for inducing predictive models based on efficient subdivisions of the input space (patches) Areas of Focus:  Linear Regression Tree Induction  Classification Tree Induction UH-DMML Lab 2

Linear regression is a global model, where there is a single predictive formula holding over the entire data-space. Y = β 0 + β T X + Linear Regression Tree When the data has lots of input attributes which interact in complicated, nonlinear ways, assembling a single global model can be very difficult. An alternative approach to nonlinear regression is to split, or partition, the space into smaller regions, where the interactions are more manageable. We then partition the sub-divisions again - this is called recursive partitioning - until finally we get to chunks of the space which can fit simple models to them. Splitting Method  selecting the pair {split variable, split value}  minimizing some error/objective function Background (Research Goals Continued) UH-DMML Lab 3

Popular approaches: 1-Variance-based  {split variable, split value} selection: try each mean value of each input attribute  objective function: variance minimization scalable, complex trees, often less accurate 2-RSS-based  {split variable, split value} selection: try each value for each input attribute ( Exhaustive search)  objective function: RSS minimization (Residual Sum of Squared Errors) Less scalable, smaller trees, better accuracy Our Research Goals: To induce smaller trees with better accuracy while improving on scalability by designing better splitting methods (patches), and objective functions UH-DMML Lab 4 Background (Research Goals Continued)

UH-DMML Lab 5 y y (A) (D) (G)(B)(C) (E) (F) A A B B G G Problem Setting x x Exhaustive search Variance-based

y (A) (B)(C) A A B B C C x Problem Setting Our Research Goals: To induce smaller trees with better accuracy while improving on scalability by designing better splitting methods (patches), and objective functions UH-DMML Lab 6

1-Variance –based approaches like M5 will miss the optimum split point 2-Exhaustive search approaches like RETIS will find optimum split point but at cost of expensive search (not scalable) Example (Problem Setting Continued) UH-DMML Lab 7

Current Proposed Solution o Detect areas in the dataset where the general trend makes sharp turns (Turning Points) o Use Turning Points as potential split points in a Linear Regression Tree induction algorithm Challenges:  Determining the turning points  Balancing accuracy, model complexity, and runtime complexity UH-DMML Lab 8 Solutions

UH-DMML Lab 9 Determining Turning Points (Solutions continued)

Two algorithms: TPRTI-A and TPRTI-B Both rely on 1.detecting potential split points in the dataset (turning points) 2.then feed a tree induction algorithm with the split points TPRTI-A and TPRTI-B differ by their objective functions o TPRTI-A RSS based node evaluation approach o TPRTI-B uses a two steps node evaluation function  Select split point based on distance  Use RSS computation to select the pair {split variable/split value} UH-DMML Lab 10 Two New Algorithms (Solutions continued)

TPRTI-A RSS based node evaluation approach. Does a look-ahead split for each turning point and select the split that best minimizes RSS UH-DMML Lab 11 TPRTI-B uses a two steps node evaluation function  Select split point based on distance  Use RSS computation to select the pair {split variable/split value} Two New Algorithms (Solutions continued)

M5TPRTI-ATPRTI-BRETISGUIDESECRET TPRTI-A(6/5/1)-(4/6/2)(4/6/0)(5/1/2)(4/2/2) TPRTI-B(4/6/2)(2/6/4)-(3/7/0)(5/1/2)(1/4/3) Table1. Comparison between TPRTI-A, TPRTI-B and state-of-the-art approaches with respect to accuracy (wins/ties/loses) Results On Accuracy UH-DMML Lab 12

Results On Complexity Table2. Number of times an approach obtained the combination (Best accuracy, fewest leaf-nodes) M5TPRTI-ATPRTI-BRETISGUIDESECRET 05 35N.A. UH-DMML Lab 13

Results On Scalability UH-DMML Lab 14

We propose a new approach for Linear Regression Tree construction called Turning Point Regression Tree Induction (TPRTI) that infuses turning points into a regression tree induction algorithm to achieve  improved scalability while maintaining high accuracy and low model complexity.  Two novel linear regression tree induction algorithms called TPRTI-A and TPRTI-B which incorporate turning points into the node evaluation were introduced and experimental results indicate that TPRTI is a scalable algorithm that is capable of obtaining a high predictive accuracy using smaller decision trees than other approaches. Conclusion UH-DMML Lab 15

FUTURE WORK We are investigating how turning point detection can also be used to induce better classification trees. UH-DMML Lab 16

Thank You