L. Isella, A. Karvounaraki (JRC) D. Karlis (AUEB)

Slides:



Advertisements
Similar presentations
Innovation Surveys: Advice from the Oslo Manual National training workshop Amman, Jordan October 2010.
Advertisements

Statistics for Improving the Efficiency of Public Administration Daniel Peña Universidad Carlos III Madrid, Spain NTTS 2009 Brussels.
Neural networks Introduction Fitting neural networks
CountrySTAT Team-I November 2014, ECO Secretariat,Teheran.
Kin 304 Regression Linear Regression Least Sum of Squares
Brief introduction on Logistic Regression
CHAPTER 2: Supervised Learning. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Learning a Class from Examples.
Model generalization Test error Bias, variance and complexity
Model Assessment, Selection and Averaging
Linear Methods for Regression Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
2008 Chingchun 1 Bootstrap Chingchun Huang ( 黃敬群 ) Vision Lab, NCTU.
Comparison of Regularization Penalties Pt.2 NCSU Statistical Learning Group Will Burton Oct
1 Chapter 17: Introduction to Regression. 2 Introduction to Linear Regression The Pearson correlation measures the degree to which a set of data points.
Simple Linear Regression Analysis
Classification and Prediction: Regression Analysis
Incorporating Temporal Effect into Crash Safety Performance Functions Wen Cheng, Ph.D., P.E., PTOE Civil Engineering Department Cal Poly Pomona.
Relationships Among Variables
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition
Measuring the quality of regional estimates from the ABS Jennie Davies and Daniel Ayoubkhani.
2015 AprilUNIVERSITY OF HAIFA, DEPARTMENT OF STATISTICS, SEMINAR FOR M.A 1 Hastie, Tibshirani and Friedman.The Elements of Statistical Learning (2nd edition,
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Hydrologic Modeling: Verification, Validation, Calibration, and Sensitivity Analysis Fritz R. Fiedler, P.E., Ph.D.
Model Building III – Remedial Measures KNNL – Chapter 11.
Shrinkage Estimation of Vector Autoregressive Models Pawin Siriprapanukul 11 January 2010.
Chapter 12: Linear Regression 1. Introduction Regression analysis and Analysis of variance are the two most widely used statistical procedures. Regression.
Jeff Howbert Introduction to Machine Learning Winter Regression Linear Regression.
Statistical Tools for Solar Resource Forecasting Vivek Vijay IIT Jodhpur Date: 16/12/2013.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Chapter 4: Introduction to Predictive Modeling: Regressions
CROSS-VALIDATION AND MODEL SELECTION Many Slides are from: Dr. Thomas Jensen -Expedia.com and Prof. Olga Veksler - CS Learning and Computer Vision.
Stat 112 Notes 9 Today: –Multicollinearity (Chapter 4.6) –Multiple regression and causal inference.
Multivariate Data Analysis Chapter 1 - Introduction.
CpSc 881: Machine Learning
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
1 Chapter 4: Introduction to Predictive Modeling: Regressions 4.1 Introduction 4.2 Selecting Regression Inputs 4.3 Optimizing Regression Complexity 4.4.
Machine Learning 5. Parametric Methods.
From OLS to Generalized Regression Chong Ho Yu (I am regressing)
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Model based approach for estimating and forecasting crop statistics: Update, consolidation and improvement of AGROMET model “AGROMET Project” Working Group.
LECTURE 13: LINEAR MODEL SELECTION PT. 3 March 9, 2016 SDS 293 Machine Learning.
Data Summit 2016 H104: Building Hadoop Applications Abhik Roy Database Technologies - Experian LinkedIn Profile:
Data Science Credibility: Evaluating What’s Been Learned
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Multiple Regression.
Computer aided teaching of statistics: advantages and disadvantages
Chapter 7. Classification and Prediction
Inference for Least Squares Lines
B&A ; and REGRESSION - ANCOVA B&A ; and
Generalized regression techniques
Kaniz Rashid Lubana Mamun MS Student: CSU Hayward Dr. Eric A. Suess
Econometric methods of analysis and forecasting of financial markets
Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Section 11.1: Least squares estimation CIS Computational.
Multiple Regression.
Dr. Morgan C. Wang Department of Statistics
CHAPTER- 17 CORRELATION AND REGRESSION
Predict Failures with Developer Networks and Social Network Analysis
What is Regression Analysis?
Chap. 7 Regularization for Deep Learning (7.8~7.12 )
Linear Model Selection and regularization
Bootstrapping Jackknifing
Product moment correlation
Cross-validation Brenda Thomson/ Peter Fox Data Analytics
Ensemble learning Reminder - Bagging of Trees Random Forest
15.1 The Role of Statistics in the Research Process
Predicting Complex Models for Industrial Processes
Neural networks (1) Traditional multi-layer perceptrons
New Techniques and Technologies for Statistics 2017  Estimation of Response Propensities and Indicators of Representative Response Using Population-Level.
Introduction to Machine learning
Presentation transcript:

L. Isella, A. Karvounaraki (JRC) D. Karlis (AUEB) Nowcasting BERD at MS Level: a Penalised Linear Model Approach L. Isella, A. Karvounaraki (JRC) D. Karlis (AUEB)

Definition Nowcasting: "Exploitation of information published early and/or at higher frequencies than the target variable of interest to obtain an early estimate before the official figure" (M. Bańbura et al, Now-casting and the real-time data flow. ECB, NO 1564 / July 2013) BERD: Business enterprise expenditure on R&D covers R&D activities carried out in the business sector by performing firms and institutes, regardless of the origin of funding and is arguably most closely linked to the creation of new products and production techniques. Used to improve the timeliness of the data

Nowcasting BERD Rational Timeliness matters: official data are two years old BERD accounts for more than 60% of the total R&D expenditure at EU level BERD expenditure is used in a number of reports and indicators including the EIS. Time series analysis and clustering methods have not performed sufficiently well. Previous attempts Merit: Timeliness of Innovation Union Scoreboard data. Options for forecasting data Brussels October 2015 M.Moucharta, J. Romboutsb Clustered panel data models: An efficient approach for nowcasting from poor data International Journal of Forecasting 21 (2005) 577– 594 Statistics Austria: Nowcasting procedures in Austria for estimating Research & Development. OECD Statistics Newsletter July 2012 Reference to Eurostat's update of R&D statistics Merit:Expert Workshop on Opportunities and Challenges of Improving the Timeliness of Innovation Data.

Model Requirements Open sustainable data Easy to update/maintain hence No manual selection of the input data for each MS No hand tuning the parameters for each MS No computational intensive Model as a prediction tool: validation as a part of model development. Keep things as simple as possible, but not any simpler need to go beyond plain OLS

Methodology overview Selection of the best predictors in terms of relevance and data availability. We consider 27 predictors Selection of an appropriate model based on its prediction capability. GLMNET in R Test the predictions. Cross validation. Leave the last available year out for testing Calculate prediction errors and their CI. Bootstrap

Model Input Same predictors as model input for each MS Feature selection performed algorithmically

OLS approach OLS fits the model by solving But it cannot handle p>Noverfitting!

Ridge and Lasso Regressions When p>N or many correlated predictors, best to use Ridge regression: it shrinks the coefficients of correlated predictors towards each other Lasso regression: it picks one of the correlated predictors and ignores the rest

GLMNET Elastic net models combine ridge and lasso ideas 𝛼=0ridge; 𝛼=1lasso penalty Biased {𝛽}, but reduced model variance Handles well many correlated predictors How to tune the model? Choice of 𝛼 and λ (hyperparameters)?cross validation

Cross Validation Split randomly the data into K=3 groups and take a couple of values {λ,𝛼} Each time 1 group as test and the other 2 as train set. Evaluate model performance on test set Repeat many times with different {λ,𝛼} and select { λ 𝑜𝑝𝑡 , 𝛼 𝑜𝑝𝑡 } Test Train Prediction error 1 Prediction error 2 Prediction error 3

CI for prediction We used a Bootstrap approach Example on a simple linear regression By resampling the residuals, create many realization of the independent variable to retrain your model (fixed-x bootstrap) and get new predictions It is a non-parametric technique, so no assumptions are made on the distribution of the residuals.

After GLMNET and cross validation Train model only on data up to 2013 (withhold 2014) Nowcast BERD 2014 (not used in model training)  Overfitting and poor nowcast GLMNET+CV+CI

EU28 nowcast Prediction error less than 1%.

Overall Assessment 11 MS have prediction error<5%. Their BERD accounts for 70% of the total EU28 BERD 10 MS have prediction error>10%. Their BERD accounts for less than 5% of the total EU28 BERD This is the reason why the method gives so go predictions for the EU28 At MS level the problem is with small economies where even small changes have visible impacts

Way forward Improvement of the methodology (e.g. introduction of weights) Investigate poor performance of the model in certain countries Apply the method to other R&D indicators (e.g. public R&D expenditure)

Thank you for your attention! Last but not least…… Thank you for your attention! 21 November 2018