Download presentation
Presentation is loading. Please wait.
1
The Ensemble Method and Model Comparison for Predictive Modeling with Big Data
Siyan Gan, Pepperdine University Hyun Seo Lee, Azusa Pacific University Emily Brown, Azusa Pacific University Chong Ho Yu, Ph.D., D. Phil., Azusa Pacific University
2
Problem in replication
You cannot step into the same river twice! Open Science Collaboration (OSC) (2015) found that a large portion of the replicated results were not as strong as the original reports in terms of significance (p values) and magnitude (effect sizes). 97% of the original studies reported significant results (p < .05), but only 36% of the replicated studies yielded significant findings. The average effect size of the replicated studies was only half of the initial studies (es = vs. es = 0.403). Machine learning: reapeatiing analysis, each sit with a subset by invreasing the accuracy Its AI Allow the machines to make deciions (space) Machines are learning new informations, finding new erros , then make new adjustmnt to the new envorment
3
Ensemble Methods Replication within the same data
The data set are separated into many subsets and multiple analyses are run. In each run the model is refined by previous "training.” Machine learning (based on Artificial intelligence): Learning from previous analysis The Ensemble Method Compares, complements, and combines multiple methods in the analysis. Better predictive outcome than using just one analysis. Machine learning: reapeatiing analysis, each sit with a subset by invreasing the accuracy Its AI Allow the machines to make deciions (space) Machines are learning new informations, finding new erros , then make new adjustmnt to the new envorment
4
Ensemble Methods in Big Data Analytics/ Data Mining
Big data set are separated into many subsets and multiple analyses are run. In each run the model is refined by previous "training.” Machine learning (based on Artificial intelligence): Learning from previous analysis The Ensemble Method: Merging multiple analyses Compares, complements, and combines multiple methods in the analysis. Better predictive outcome than using just one analysis. Machine learning: reapeatiing analysis, each sit with a subset by invreasing the accuracy Its AI Allow the machines to make deciions (space) Machines are learning new informations, finding new erros , then make new adjustmnt to the new envorment
5
Boosting vs. Bagging Boosting Bagging (not begging)
Creates a working model from the subsets of the original data set. Adjusts weak models so they are combined to be a strong model. (Bootstrap Aggregation) Repeated multisets of additional training data from the original sample Boosting: gradual process – making improvements on the errors made throughout the process, eventually the model gets strong Bagging: run subsets in parallel without informing each other, in the end the agrithym puts them together Both good, run both in our research
6
Bagging vs. Boosting Bagging Boosting Sequent Two-step Sequential
Bagging Boosting Sequent Two-step Sequential Partitioning data into subsets Random Give misclassified cases a heavier weight Sampling method Sampling with replacement Sampling without replacement Relations between models Parallel ensemble: Each model is independent Previous models inform subsequent models Goal to achieve Minimize variance Minimize bias, improve predictive power Method to combine models Weighted average Majority vote Requirement of computing resources Highly computing intensive Less computing intensive Major difference between Bagging: each analysis is independent from the other Boosting:
7
Bagging as voting Imagine that 1,000 analysts each one randomly draw a sub-sample from the big sample and then run an analysis. The results must be diverse. Now these 1,000 analysts meet together and vote. “How many of us found Variable A as a crucial predictor? Please raise your hand.” And then move on to B, C…etc. At the end we have a list of top 10 or Top 20.
8
Boosting as gradual tuning
Imagine that you are a cook. You put spices into the dish and taste it. If it is not salty enough, you put more salt and pepper next time. But next time it is too spicy, then you put less hot sauce. You make gradual improvement every time until you have the final recipe. Similarly, boosting is a gradual tuning process.
9
Bias and variance The bias is quantified by the error which results from missing a target e.g, if an estimated mean is 3, but the actual population value is 3.5, then the bias value is 0.5. The variance is the error which results from noise or random fluctuation. When the variance of a model is high, this model is considered unstable.
10
Bias and variance A complicated model tends to have low bias but high variance. Conversely, a simple model is more likely to have a higher bias and a lower variance. We want a well-built, optimal model!
11
Bias and variance Bagging is utilized to decrease the variance
increases the size of these generated data and effectively minimizes the variance of prediction by decreasing the influence of extreme scores Boosting is used to weaken the bias in the process of building a predictive model. increase the predictive accuracy
12
Debate on boosting and bagging
In theory, an ensemble method should suppress both bias and variance by merging overfitted and underfitted models. Kotsiantis (2013) found that bagging tends to generate less heterogeneous models than boosting. Fumera, Roli, and Serrau (2005) found that the misclassification rate of bagging has the same bias as a single bootstrap. When these overfitted models are averaged, the same bias is retained while the variance is cancelled out.
13
PIAAC Study Programme for the International Assessment of Adult Competencies (PIAAC). Developed by Organization for Economic and Cooperation and Development (OECD). In 2014, PIAAC collected data from 33 participating nations (OECD, 2016). U.S. adults were behind in all three test categories: Literacy, numeracy, and problem solving in technology-rich environments. Survey items included factors related to learning: Readiness to learn, cultural engagement, political efficacy, and social trust.
14
Method
15
Variables The scores of literacy, numeracy, and technology-based problem-solving strongly correlated. # of variables Independent varaibles and dependent PCA tells us that 3 Vs can be combnined into 1 Component DATA REDUCTION TECHNIQUE There are debates on these two methods. So we run both! Correlation matrix of literacy, numeracy, and problem-solving.
16
Variables All three skills were combined into one component.
Composite score of literacy, numeracy, and problem-solving was treated as the dependent variable. # of variables Independent varaibles and dependent PCA tells us that 3 Vs can be combnined into 1 Component DATA REDUCTION TECHNIQUE There are debates on these two methods. So we run both! Screen plot of PCA of literacy, numeracy, and problem solving.
17
Boosting and bagging in JMP Pro
# of variables Independent varaibles and dependent PCA tells us that 3 Vs can be combnined into 1 Component DATA REDUCTION TECHNIQUE There are debates on these two methods. So we run both!
18
Boosting and bagging in SAS
proc hpforest data= <data>; target <dependent variable> / level= <binary, nominal, interval>; input <variable 1, variable 2, variable 3…etc.>/ level= <binary, nominal, ordinal, interval> ; run; PROC TREEBOOST DATA= <data>; INPUT <predictor 1, predictor 2…>/ LEVEL=; TARGET <dependent variable> /; ASSESS VALIDATA= <training/validation>; RUN; # of variables Independent varaibles and dependent PCA tells us that 3 Vs can be combnined into 1 Component DATA REDUCTION TECHNIQUE There are debates on these two methods. So we run both!
19
Boosting and bagging in Enterprise Miner (EM)
The analyst can go one step further by creating an ensemble of models yielded from different modeling techniques. The control point does not perform any computation; it simply stores the results for the ensemble node. # of variables Independent varaibles and dependent PCA tells us that 3 Vs can be combnined into 1 Component DATA REDUCTION TECHNIQUE There are debates on these two methods. So we run both!
20
Results
21
Model Comparison Subset type Method R2 RASE AAE No subset OLS regression 0.1647 43.692 34.603 Training Boosting 0.2058 42.708 34.031 Bagging 0.4813 34.515 26.979 Validation 0.1791 43.488 34.597 0.1685 43.768 34.689 Bagging and boosting outperformed than OLS regression in variance explained and error rate. In training the bootstrap method yielded overfitted models because the R2 is unreasonably high. Training: to training the agrthym to propose a model Validation: to confirm In terms of R square aND THE AAE (error rate) RASE (), boosting showed The boosted tree model outperformed the bagging approach (higher variance explained and lower error). Using R-square, RASE, and AAE
22
OLS Regression Result Predictor Estimate Std. Error t Ratio p
Relate new ideas into real-life 13.07 0.85 15.32 <.0001* Like learning new things 1.93 1.02 1.89 0.0595 Attribute something new 1.54 0.98 1.56 0.1180 Get to the bottom of difficult things 1.80 0.91 1.96 0.0497* Figure out how different ideas fit together -3.46 0.96 -3.61 0.0003* Looking for additional info 0.56 0.95 0.59 0.5576 Voluntary work for non-profit organizations 4.50 7.97 No influence on the government -3.08 0.53 -5.85 Trust only few people -3.57 0.61 -5.84 Other people take advantage of you -3.28 0.73 -4.50 Problem with regression EVERYTHING IS SIGNIFICANT
23
Final Boosted Tree Model for the USA sample
PIAAC also includes many survey items that are believed to be related to those learning outcomes We choose 8 as independent variable, after Independent variables Top predictors: cultural engagement (voluntary work for non-profit organizations), social trust (other people take advantage of you), and readiness to learn (like learning new things).
24
Median smoothing plots
Learning outcomes and social trust in the US sample. Visualization Non-linear 8 CHANGE THE LABLES LINEAR REALTIONSHIP
25
Median smoothing plots
Learning outcomes and readiness to learn in the US sample. Visualization Non-linear 8 CHANGE THE LABLES LINEAR REALTIONSHIP
26
Median smoothing plots
Learning outcomes and cultural engagement in the US sample. Visualization Non-linear 8 CHANGE THE LABLES LINEAR REALTIONSHIP
27
Discussion Method choice and model goodness should be assessed on a case-by-case basis Run both bagging and boosting, then choose the best result according to the model comparison. Big data analytics fixes the problem of hypothesis testing by using model building and data visualization When the ensemble method, model comparison, and data visualization are employed side by side, interesting patterns and meaningful conclusions can be found from a big data set.
28
Contact Information Name: Chong Ho Alex Yu Company: Azusa Pacific University City/State: Azusa CA Phone:
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.