U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Predicting Solar Generation from Weather Forecasts Using Machine Learning Navin.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

(SubLoc) Support vector machine approach for protein subcelluar localization prediction (SubLoc) Kim Hye Jin Intelligent Multimedia Lab
Generative Models Thus far we have essentially considered techniques that perform classification indirectly by modeling the training data, optimizing.
ECG Signal processing (2)
INTRODUCTION TO Machine Learning 2nd Edition
An Introduction of Support Vector Machine
P ROJECT P ROPOSALS F OR CS620 Rajdeep Sardar Mtech 2 nd year student, Department Of Computer Science & Engineering, IIT Bombay.
An Introduction of Support Vector Machine
Support Vector Machines
SVM—Support Vector Machines
Pattern Recognition and Machine Learning: Kernel Methods.
Computer vision: models, learning and inference Chapter 8 Regression.
CS Statistical Machine learning Lecture 13 Yuan (Alan) Qi Purdue CS Oct
SunCast: Fine-grained Prediction of Natural Sunlight Levels for Improved Daylight Harvesting Jiakang Lu and Kamin Whitehouse Department of Computer Science,
A Comprehensive Study on Third Order Statistical Features for Image Splicing Detection Xudong Zhao, Shilin Wang, Shenghong Li and Jianhua Li Shanghai Jiao.
1 Multivariate Statistics ESM 206, 5/17/05. 2 WHAT IS MULTIVARIATE STATISTICS? A collection of techniques to help us understand patterns in and make predictions.
Support Vector Machines (and Kernel Methods in general)
Dr. Yukun Bao School of Management, HUST Business Forecasting: Experiments and Case Studies.
Speaker Adaptation for Vowel Classification
Load Forecasting Eugene Feinberg Applied Math & Statistics Stony Brook University NSF workshop, November 3-4, 2003.
U NIVERSITY OF M ASSACHUSETTS, A MHERST – Department of Computer Science Dynamic Resource Allocation for Shared Data Centers Using Online Measurements.
CSE 300: Software Reliability Engineering Topics covered: Software metrics and software reliability Software complexity and software quality.
Multivariate Analysis A Unified Perspective
Data mining and statistical learning - lecture 13 Separating hyperplane.
1 Ensembles of Nearest Neighbor Forecasts Dragomir Yankov, Eamonn Keogh Dept. of Computer Science & Eng. University of California Riverside Dennis DeCoste.
Oregon State University – Intelligent Systems Group 8/22/2003ICML Giorgio Valentini Dipartimento di Scienze dell Informazione Università degli Studi.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
An Introduction to Support Vector Machines Martin Law.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Efficient Model Selection for Support Vector Machines
Geostatistical approach to Estimating Rainfall over Mauritius Mphil/PhD Student: Mr.Dhurmea K. Ram Supervisors: Prof. SDDV Rughooputh Dr. R Boojhawon Estimating.
Support Vector Machines Mei-Chen Yeh 04/20/2010. The Classification Problem Label instances, usually represented by feature vectors, into one of the predefined.
Guidance on Intensity Guidance Kieran Bhatia, David Nolan, Mark DeMaria, Andrea Schumacher IHC Presentation This project is supported by the.
A hybrid SOFM-SVR with a filter-based feature selection for stock market forecasting Huang, C. L. & Tsai, C. Y. Expert Systems with Applications 2008.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
An Introduction to Support Vector Machines (M. Law)
Dimensionality Reduction Motivation I: Data Compression Machine Learning.
Christopher M. Bishop, Pattern Recognition and Machine Learning.
Principal Component Analysis Machine Learning. Last Time Expectation Maximization in Graphical Models – Baum Welch.
Support Vector Machines in Marketing Georgi Nalbantov MICC, Maastricht University.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Text Classification using Support Vector Machine Debapriyo Majumdar Information Retrieval – Spring 2015 Indian Statistical Institute Kolkata.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
Feature Selection and Extraction Michael J. Watts
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
Chapter 15: Classification of Time- Embedded EEG Using Short-Time Principal Component Analysis by Nguyen Duc Thang 5/2009.
LOAD FORECASTING. - ELECTRICAL LOAD FORECASTING IS THE ESTIMATION FOR FUTURE LOAD BY AN INDUSTRY OR UTILITY COMPANY - IT HAS MANY APPLICATIONS INCLUDING.
Multi-Area Load Forecasting for System with Large Geographical Area S. Fan, K. Methaprayoon, W. J. Lee Industrial and Commercial Power Systems Technical.
Forecasting. Model with indicator variables The choice of a forecasting technique depends on the components identified in the time series. The techniques.
CSE 4705 Artificial Intelligence
PREDICT 422: Practical Machine Learning
Sparse Kernel Machines
Recap Finds the boundary with “maximum margin”
Boosting and Additive Trees (2)
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Computer vision: models, learning and inference
Outline Peter N. Belhumeur, Joao P. Hespanha, and David J. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,”
Blind Signal Separation using Principal Components Analysis
Hyperparameters, bias-variance tradeoff, validation
CSSE463: Image Recognition Day 14
Recap Finds the boundary with “maximum margin”
Recitation 6: Kernel SVM
Generally Discriminant Analysis
Support Vector Machine _ 2 (SVM)
Support Vector Machines and Kernels
Recursively Adapted Radial Basis Function Networks and its Relationship to Resource Allocating Networks and Online Kernel Learning Weifeng Liu, Puskal.
Multivariate Methods Berlin Chen
Feature Selection Methods
Support Vector Machines 2
Presentation transcript:

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Predicting Solar Generation from Weather Forecasts Using Machine Learning Navin Sharma, Pranshu Sharma, David Irwin, and Prashant Shenoy

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Harvesting Examples Perpetual Sensor Networks Run forever off harvested energy[EWSN 2009] Off-the-grid infrastructure Power cellular towers & ATM Smart homes and smart cities Use on-site solar & wind power[BuildSys 2011]

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Renewables are Intermittent Example: Solar shows significant variation Nearly no energy How much energy will we harvest today?

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Predictions are Important Better predictions == Better performance Examples: Smart homes[BuildSys 2011] Reduce utility bill by 2.7X Eliminate peak power demands Sensor Network[SECON 2010] Lexicographical sensor network: increases sensing rate by 60% Sensor testbeds: serve 70% more requests

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Prediction Methods Existing Prediction Methods Past Predicts Future (PPF) Variants of PPF EWMA[TECS 2007] WCMA[VITAE 2009] Past Predicts Future Accurate for short time scales (seconds to minutes) Hard to predict at medium time scales (hours to days)

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Problem Statement How can we statistically predict solar harvesting ? Approach: Leverage weather forecast to predict solar energy Use statistical power of machine learning

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Outline Motivation Intuition & Methodology Prediction Model Evaluation Conclusion

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Forecast-based Predictions Idea for using weather forecasts PPF accurate for constant weather Forecasts also predict significant weather changes

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Methodology Analyze Weather Data Forecast data from National Weather Service Formulate Forecast  Solar Intensity Model Use machine learning regression techniques Solar Intensity = F (time, multiple weather parameters) Derive Solar Intensity  Solar Energy Model Empirically from our solar panel deployment

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Data Analysis Solar intensity exhibits strong (but not perfect) correlation with sky cover, humidity, and precipitation

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Data Analysis Solar intensity exhibits no correlation with wind speed, but weak correlation with temperature

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Prediction Technique ML Regression Techniques Training data set to find regression coefficients Testing data set to verify the model’s accuracy Our data set Training data set: First 8 months of 2010 Testing data set: Next 2 months of 2010 What to predict? Solar intensity at noon Based on 3-hr weather forecast at 9 AM

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Support Vector Machines Support Vector Machine (SVM) Used for classification & regression Independent of input space dimensionality Resistant to overfitting Kernel Function Maps data from low-dimensional input space to high- dimensional feature space Common Kernels Linear kernel Polynomial kernel Radial Basis Function (RBF)

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 SVM Regression: Steps Step 1:Data Preparation Normalize to zero mean and unit variance Step 2:Kernel Selection RBF performs better than linear & polynomial Grid search to find optimal parameters Optimal parameters: cost (soft margin parameter) = 256 γ (Gaussian function parameter) = ε (loss function parameter) =

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 SVM with RBF Kernel Average prediction error: 22 %

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Dimensionality Reduction Redundant Information Reduces prediction accuracy Principal Component Analysis (PCA) Correlated variables  uncorrelated variables Uncorrelated variables called principal components Choose first 4 PCs with first 4 (highest) Eigen values

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 SVM with RBF Kernel Reducing dimensions from 7 to 4 reduces prediction error from 22 % to 2 % 4-dimensions7-dimensions

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Comparison with Cloudy Model SVM-RBF with 4 dimensions predicts 27 % better than cloudy-forecast SVM-RBFCloudy-forecast Cloudy-forecast: Sky cover based empirical model for solar prediction [SECON 2010]

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Intensity  Energy Model Solar power from solar intensity Depends on solar panel characteristics Panel orientation & surrounding environments Empirically derived for a particular setup Our solar panel deployment Kyocera KC65T Solar Panel Power = * Intensity Accurate to within 2.5 % of actual harvesting

U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science 2011 Conclusions Weather forecasts can improve prediction accuracy See dramatic weather changes before they occur Facilitates better planning ML statistical models work well Future Work Design a better kernel function Hybrid Prediction: use a combination of past & forecast Apply to wind and wind gust