Wind power scenario generation Geoffrey Pritchard University of Auckland by regression clustering.

Slides:



Advertisements
Similar presentations
3.6 Support Vector Machines
Advertisements

Subspace Embeddings for the L1 norm with Applications Christian Sohler David Woodruff TU Dortmund IBM Almaden.
STATISTICS Joint and Conditional Distributions
STATISTICS Univariate Distributions
STATISTICS Random Variables and Distribution Functions
Detection Chia-Hsin Cheng. Wireless Access Tech. Lab. CCU Wireless Access Tech. Lab. 2 Outlines Detection Theory Simple Binary Hypothesis Tests Bayes.
An Application of Linear Programming Lesson 12 The Transportation Model.
Discrimination and Classification. Discrimination Situation: We have two or more populations  1,  2, etc (possibly p-variate normal). The populations.
Local Search Jim Little UBC CS 322 – CSP October 3, 2014 Textbook §4.8
CPSC 322, Lecture 14Slide 1 Local Search Computer Science cpsc322, Lecture 14 (Textbook Chpt 4.8) Oct, 5, 2012.
Unsupervised Learning Clustering K-Means. Recall: Key Components of Intelligent Agents Representation Language: Graph, Bayes Nets, Linear functions Inference.
Networks Prim’s Algorithm
Maximum Likelihood And Expectation Maximization Lecture Notes for CMPUT 466/551 Nilanjan Ray.
Supervised Learning Recap
Chapter 4: Linear Models for Classification
Second order cone programming approaches for handing missing and uncertain data P. K. Shivaswamy, C. Bhattacharyya and A. J. Smola Discussion led by Qi.
Visual Recognition Tutorial
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
Non-Linear Problems General approach. Non-linear Optimization Many objective functions, tend to be non-linear. Design problems for which the objective.
4. Ad-hoc I: Hierarchical clustering
Basic Data Mining Techniques Chapter Decision Trees.
Clustering.
Visual Recognition Tutorial
Today Logistic Regression Decision Trees Redux Graphical Models
Modelling inflows for SDDP Dr. Geoffrey Pritchard University of Auckland / EPOC.
Lecture 10: Robust fitting CS4670: Computer Vision Noah Snavely.
On the convergence of SDDP and related algorithms Speaker: Ziming Guan Supervisor: A. B. Philpott Sponsor: Fonterra New Zealand.
Radial Basis Function Networks
Collaborative Filtering Matrix Factorization Approach
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Inductive learning Simplest form: learn a function from examples
ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 1 Chapter 1: The DP Algorithm To do:  sequential decision-making  state.
Outline Classification Linear classifiers Perceptron Multi-class classification Generative approach Naïve Bayes classifier 2.
CSC321: Neural Networks Lecture 12: Clustering Geoffrey Hinton.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 16 Nov, 3, 2011 Slide credit: C. Conati, S.
Stochastic Linear Programming by Series of Monte-Carlo Estimators Leonidas SAKALAUSKAS Institute of Mathematics&Informatics Vilnius, Lithuania
Markov Decision Processes1 Definitions; Stationary policies; Value improvement algorithm, Policy improvement algorithm, and linear programming for discounted.
Learning from observations
Learning a Small Mixture of Trees M. Pawan Kumar Daphne Koller Aim: To efficiently learn a.
Overview of Supervised Learning Overview of Supervised Learning2 Outline Linear Regression and Nearest Neighbors method Statistical Decision.
1 Statistical Techniques Chapter Linear Regression Analysis Simple Linear Regression.
© Department of Statistics 2012 STATS 330 Lecture 20: Slide 1 Stats 330: Lecture 20.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Single-Factor Studies KNNL – Chapter 16. Single-Factor Models Independent Variable can be qualitative or quantitative If Quantitative, we typically assume.
CHAPTER 17 O PTIMAL D ESIGN FOR E XPERIMENTAL I NPUTS Organization of chapter in ISSO –Background Motivation Finite sample and asymptotic (continuous)
CSSE463: Image Recognition Day 23 Midterm behind us… Midterm behind us… Foundations of Image Recognition completed! Foundations of Image Recognition completed!
Flat clustering approaches
Stochastic processes for hydrological optimization problems
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Monte-Carlo based Expertise A powerful Tool for System Evaluation & Optimization  Introduction  Features  System Performance.
Ence 627 Decision Analysis for Engineering Project Portfolio Selection: “Optimal Budgeting of Projects Under Uncertainty” Javier Ordóñez.
6.S093 Visual Recognition through Machine Learning Competition Image by kirkh.deviantart.com Joseph Lim and Aditya Khosla Acknowledgment: Many slides from.
Central Limit Theorem Let X 1, X 2, …, X n be n independent, identically distributed random variables with mean  and standard deviation . For large n:
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Given a set of data points as input Randomly assign each point to one of the k clusters Repeat until convergence – Calculate model of each of the k clusters.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Wind Power & Grid Operation Dr. Geoffrey Pritchard University of Auckland.
Outline Time series prediction Find k-nearest neighbors Lag selection Weighted LS-SVM.
6.5 Stochastic Prog. and Benders’ decomposition
Machine Learning Basics
Overview of Supervised Learning
CSSE463: Image Recognition Day 23
Collaborative Filtering Matrix Factorization Approach
by quantile regression
Clustering 77B Recommender Systems
Addressing uncertainty in a transshipment problem
Networks Prim’s Algorithm
Recap: Naïve Bayes classifier
6.5 Stochastic Prog. and Benders’ decomposition
Presentation transcript:

Wind power scenario generation Geoffrey Pritchard University of Auckland by regression clustering

Scenarios for stochastic optimization Uncertain problem data represented by a probability distribution. For computational tractability, need a finite discrete distribution, i.e. a collection of scenarios. Make decision here ?

Power system applications Wind power generation, 2 hours from now. Inflow to hydroelectric reservoir, over the next week. Typical problems solved repeatedly: –Need a procedure to generate scenarios for many problem instances, not just one.

Situation-dependent uncertainty Scenarios represent the conditional distribution of the variable(s) of interest, given some known information x. Different problem instances have different x.

Change in wind power over next 2hr Tararua/Te Apiti 28/5/ /3/2010

Change in wind power over next 2hr Tararua/Te Apiti 28/5/ /3/2010 Change in wind power: 7 discrete scenarios Each scenario is a function of the present wind power x.

Change in wind power over next 2hr Tararua/Te Apiti 28/5/ /3/2010 Change in wind power: 7 discrete scenarios Each scenario is a function of the present wind power x.

Have data x i and y i for i=1,…n x y Scenarios by quantile regression

Have data x i and y i for i=1,…n Want scenarios for y, given x. x y Scenarios by quantile regression

Have data x i and y i for i=1,…n Want scenarios for y, given x. Quantile regression: choose scenario s k () to minimize  i  k ( y i – s k (x i ) ) for a suitable loss function  k (). x y

Quantile regression fitting For a scenario at quantile ,   is the loss function  

Scenarios as functions Choose each scenario to be linear on a feature space: s k (x) =  j  jk b j (x) Typically b j () are basis functions (e.g. cubic splines). The quantile regression problem is then a linear program.

Change in wind power over next 2hr Tararua/Te Apiti 28/5/ /3/2010 Change in wind power: 7 discrete scenarios Equally likely scenarios, modelled by quantiles 1/14, 3/14, … 13/14.

Quantile regression: pros and cons Each scenario has its own model. Scenario models are fitted separately. Fitting is computationally easy. Scenarios have fixed probabilities. Events with low probability but high importance may be left out.

Another way to choose scenarios … choose scenarios to minimize expected distance of a random point to the nearest scenario. (Wasserstein approximation.) Robust to general stochastic optimization problems. Given one probability distribution …

Scenarios for conditional distributions Have data x i and y i for i=1,…n Want scenarios for y, given x. x y

Scenarios for conditional distributions Have data x i and y i for i=1,…n Want scenarios for y, given x. Wasserstein: minimize  i min k | y i – s k (x i ) | over scenarios s k () chosen from some function space. x y

Scenarios as functions Choose each scenario to be linear on a feature space: s k (x) =  j  jk b j (x) Typically b j () are basis functions (e.g. cubic splines). The Wasserstein approximation problem is then a MILP with SOS1 constraints (not that that helps).

Algorithm: clustering regression Let each observation (x i,y i ) be assigned to a scenario k(i). Choose alternately the functions s k the assignments k(i) to minimize  i | y i – s k(i) (x i ) |, until convergence (cf. k-means clustering algorithm).

Clustering regression Let each observation (x i,y i ) be assigned to a scenario k(i). Choose alternately the functions s k the assignments k(i) to minimize  i | y i – s k(i) (x i ) |, until convergence (cf. k-means clustering algorithm). For univariate y, a median regression problem

Example: wind power Example: wind power, next 2 hours

Scenario probabilities Each scenario gets a probability: that of the part of the distribution closest to it. Given one probability distribution …

Probability p k (x) of scenario k must reflect the local density of observations (x i, y i ) near (x, s k (x)). Multinomial logistic regression: probabilities proportional to exp( j  jk b j (x)) where  jk are to be found. Conditional scenario probabilities

Wind: scenarios and probabilities

9% 7% 3% 90% 33% 70% 41% 26% 21% Wind: scenarios and probabilities

The End

Wind power 2hr from now: lowest scenario, conditional on present power/wind direction

Wind power 2hr from now: lowest scenario, conditional on present power/wind direction