Machine learning methods with applications to precipitation and streamflow William W. Hsieh Dept. of Earth, Ocean & Atmospheric Sciences, The University.

Slides:



Advertisements
Similar presentations
The Nonlinear Patterns of North American Winter Climate associated with ENSO Aiming Wu, William Hsieh University of British Columbia Amir Shabbar Environment.
Advertisements

Data Mining Classification: Alternative Techniques
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Mechanistic crop modelling and climate reanalysis Tom Osborne Crops and Climate Group Depts. of Meteorology & Agriculture University of Reading.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Radial Basis Functions
Experimental Real-time Seasonal Hydrologic Forecasting Andrew Wood Dennis Lettenmaier University of Washington Arun Kumar NCEP/EMC/CMB presented: JISAO.
Large-scale atmospheric circulation characteristics and their relations to local daily precipitation extremes in Hesse, central Germany Anahita Amiri Department.
Artificial Neural Network using for climate extreme in La-Plata Basin: Preliminary results and objectives David Mendes * José Antonio Marengo * Chou Sin.
MOS Developed by and Run at the NWS Meteorological Development Lab (MDL) Full range of products available at:
Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers.
The Canadian Climate Impacts Scenarios (CCIS) Project is funded by the Climate Change Action Fund and provides climate change scenarios and related information.
Chapter 13 – Weather Analysis and Forecasting. The National Weather Service The National Weather Service (NWS) is responsible for forecasts several times.
Downscaling in time. Aim is to make a probabilistic description of weather for next season –How often is it likely to rain, when is the rainy season likely.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Zhang Mingwei 1, Deng Hui 2,3, Ren Jianqiang 2,3, Fan Jinlong 1, Li Guicai 1, Chen Zhongxin 2,3 1. National satellite Meteorological Center, Beijing, China.
ABSTRACTS General Structure Background and Objective Downscaling CGCM climate change output scenario using the Artificial Neural Network model Kang Boosik.
Machine Learning Chapter 4. Artificial Neural Networks
Development of a downscaling prediction system Liqiang Sun International Research Institute for Climate and Society (IRI)
Climate data sets: introduction two perspectives: A. What varieties of data are available? B. What data helps you to identify...
Exploring sample size issues for 6-10 day forecasts using ECMWF’s reforecast data set Model: 2005 version of ECMWF model; T255 resolution. Initial Conditions:
2 nd ACRE Workshop 1st - 3rd April, 2009 O’Reilly’s Rainforest Retreat, Lamington National Park, Queensland, Australia Downscaling.
Alan Robock Department of Environmental Sciences Rutgers University, New Brunswick, New Jersey USA
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Regional climate prediction comparisons via statistical upscaling and downscaling Peter Guttorp University of Washington Norwegian Computing Center
Downscaling and its limitation on climate change impact assessments Sepo Hachigonta University of Cape Town South Africa “Building Food Security in the.
Comparison of Bayesian Neural Networks with TMVA classifiers Richa Sharma, Vipin Bhatnagar Panjab University, Chandigarh India-CMS March, 2009 Meeting,
June 16th, 2009 Christian Pagé, CERFACS Laurent Terray, CERFACS - URA 1875 Julien Boé, U California Christophe Cassou, CERFACS - URA 1875 Weather typing.
OUTLINE Current state of Ensemble MOS
Ensemble Methods: Bagging and Boosting
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss Accounting for Change: Local wind forecasts from the high-
Model Post Processing. Model Output Can Usually Be Improved with Post Processing Can remove systematic bias Can produce probabilistic information from.
Climate Scenario and Uncertainties in the Caribbean Chen,Cassandra Rhoden,Albert Owino Anthony Chen,Cassandra Rhoden,Albert Owino Climate Studies Group.
Potential for medium range global flood prediction Nathalie Voisin 1, Andrew W. Wood 1, Dennis P. Lettenmaier 1 1 Department of Civil and Environmental.
Mike Dettinger USGS, La Jolla, CA DOWNSCALING to local climate.
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Pang-Ning Tan Associate Professor Dept of Computer Science & Engineering Michigan State University
Robust Simulation of Future Hydrologic Extremes in BC under Climate Change Arelia T. Werner Markus A. Schnorbus and Rajesh R. Shrestha.
Alan F. Hamlet Andy Wood Dennis P. Lettenmaier JISAO Center for Science in the Earth System Climate Impacts Group and the Department.
The ENSEMBLES high- resolution gridded daily observed dataset Malcolm Haylock, Phil Jones, Climatic Research Unit, UK WP5.1 team: KNMI, MeteoSwiss, Oxford.
COMPARISONS OF NOWCASTING TECHNIQUES FOR OCEANIC CONVECTION Huaqing Cai, Cathy Kessinger, Nancy Rehak, Daniel Megenhardt and Matthias Steiner National.
Downscaling Global Climate Model Forecasts by Using Neural Networks Mark Bailey, Becca Latto, Dr. Nabin Malakar, Dr. Barry Gross, Pedro Placido The City.
DOWNSCALING GLOBAL MEDIUM RANGE METEOROLOGICAL PREDICTIONS FOR FLOOD PREDICTION Nathalie Voisin, Andy W. Wood, Dennis P. Lettenmaier University of Washington,
VERIFICATION OF A DOWNSCALING SEQUENCE APPLIED TO MEDIUM RANGE METEOROLOGICAL PREDICTIONS FOR GLOBAL FLOOD PREDICTION Nathalie Voisin, Andy W. Wood and.
Incremental Reduced Support Vector Machines Yuh-Jye Lee, Hung-Yi Lo and Su-Yun Huang National Taiwan University of Science and Technology and Institute.
NOAA Northeast Regional Climate Center Dr. Lee Tryhorn NOAA Climate Literacy Workshop April 2010 NOAA Northeast Regional Climate.
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
March Outline: Introduction What is the Heat Wave? Objectives Identifying and comparing the current and future status of heat wave events over.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
of Temperature in the San Francisco Bay Area
Learning with Perceptrons and Neural Networks
Tushar Sinha Assistant Professor
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
CSE 4705 Artificial Intelligence
Pfizer HTS Machine Learning Algorithms: November 2002
Professor A.K.M. Saiful Islam
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
COMP61011 : Machine Learning Ensemble Models
Downscaling sea level rise in the Mediterranean Sea under different future climate change scenarios ( ) Kareem M. Tonbol (Ph.D.) Assistant Professor.
of Temperature in the San Francisco Bay Area
Model Post Processing.
ECE 5424: Introduction to Machine Learning
Machine Learning Today: Reading: Maria Florina Balcan
Post Processing.
Kreshna GOPAL C. Prakash KHEDUN Anoop SOHUN
Seasonal Forecasting Using the Climate Predictability Tool
Facultad de Ingeniería, Centro de Cálculo
Aiming Wu, William W. Hsieh
Presentation transcript:

Machine learning methods with applications to precipitation and streamflow William W. Hsieh Dept. of Earth, Ocean & Atmospheric Sciences, The University of British Columbia Collaborators: Alex Cannon, Carlos Gaitan & Aranildo Lima

Nonlinear regression 2 Linear regression (LR): Neural networks (NN/ANN): Adaptive basis fns h j

Cost function J minimized to solve for the weights Here J is the mean squared error. Underfit and overfit: 3

Why is climate more linear than weather? [Yuval & Hsieh, Quart. J. Roy. Met. Soc.] 4

Curse of the central limit theorem Central limit theorem says averaging weather data makes climate data more Gaussian and linear => no advantage using NN! Nowadays, climate is not just the mean of the weather data, but can be other types of statistics from weather data,  E.g. climate of extreme weather Use NN on daily data, then compile climate statistics of extreme weather => Can we escape from the curse of the central limit theorem? 5

Statistical downscaling Global climate models (GCM) have poor spatial resolution (a) Dynamical downscaling: imbed regional climate model (RCM) in the GCM (b) Statistical downscaling (SD): Use statistical/machine learning methods to downscale GCM output. Statistical downscaling at 10 stations in Southern Ontario & Quebec. [Gaitan, Hsieh & Cannon, Clim.Dynam. 2014] Predictors ( ) from the NCEP/NCAR Reanalysis interpolated to the grid (approx. 3.75° lat. by 3.75° lon.) used by the Canadian CGCM3.1. 6

How to validate statistical downscaling in future climate?  Following Vrac et al. (2007), use regional climate model (RCM) output as pseudo-observations.  CRCM 4.2 provides 10 “pseudo-observational” sites  For each site, downscale from 9 surrounding CGCM 3.1 grid cells:  6 predictors/cell : T max,T min, surface u, v, SLP, precipitation.  6 predictors/cell x 9 cells = 54 predictors  2 Periods: (20 th century climate: 20C3M run) 2041 – 2070 (future climate: SRES A2 scenario) 7

10 meteorological stations 8

Precipitation Occurrence Models Using the 54 available predictors and a binary predictand (precip./ no precip), we implemented the following models:  Linear Discriminant classifier  Naive Bayes classifier  kNN (k nearest neighbours) classifier (45 nearest neighbours)  Classification Tree  TreeEnsemble: Ensemble of classification trees.  ANN-C: Artificial Neural Network Classifier. 9

Persistence Discriminant naïve-Bayes kNN ClassTree TreeEnsem. ANN-C Peirce skill score (PSS) for downscaled precipitation occurrence: 20 th Century (20C3M) and future (A2) periods. 10

Climdex Climate indices 11 Compute indices from downscaled daily data

ANN-F ARES-F SWLR-F IOA 20C3MA2 Index of agreement (IOA) of climate indices ANN-F ARES-F SWLR-F 12

ANN-F ARES-F SWLR-F Differences between the IOA of future (A2) and 20 th Century (20C3M) climates 13

Conclusion Use NN on daily data, then compile climate statistics of extreme weather => beat linear method => escaped from the curse of the central limit theorem 14

Extreme learning machine (ELM): [G.-B. Huang] ANN: randomly ELM: Choose the weights ( w ij and w 0j ) of the hidden neurons randomly. Only need to solve for a j and a 0 by linear least squares. ELM turns nonlinear regression by NN into a linear regression problem! 15

Tested ELM on 9 environmental datasets [Lima, Cannon and Hsieh, Environmental Modelling & Software, under revision ] Goal is to develop ELM into nonlinear updateable model output statistics (UMOS). 16

17 Deep learning

Spare slides 18

Compare 5 models over 9 environmental datasets [Lima, Cannon & Hsieh. Environmental Modelling & Software (under revision)] MLR = multiple linear regression ANN = Artificial neural network SVR-ES = Support vector regression (with Evolutionary Strategy) RF = random forest ELM-S = Extreme Learning Machine (with scaled weights)  Optimal number of hidden neurons in ELM chosen over validation data by a simple hill climbing algorithm. Compare models in terms of: RMSE skill score = 1 – RMSE/RMSE MLR t = cpu time 19

RMSE skill score (relative to MLR) 20

Cpu time 21

Conclusions ELM turns nonlinear ANN into a multiple linear regression problem, but with same skills as ANN. ELM-S is faster than ANN and SVR-ES in 8 of the 9 datasets and faster than RF in 5 of the 9 datasets.  When dataset has both large no. of predictors and large sample size, ELM loses its advantage over ANN. RF is fast but could not outperform MLR in 3 of 9 datasets (ELM-S outperformed MLR in all 9 datasets). 22

Online sequential learning Previously, we used ELM for batch learning. When new data arrive, need to retrain model using the entire data record => very expensive. Now use ELM for “online sequential learning” (i.e. as new data arrive, update the model with only the new data) For multiple linear regression (MLR), online sequential MLR (OS-MLR) is straightforward. [Envir.Cda’s updateable MOS (model output statistics), used to post- process NWP model output, is based on OS-MLR] Online sequential ELM (OS-ELM) (Liang et al. 2006, IEEE Trans. Neural Networks) is easily derived from OS- MLR. 23

Predict streamflow at Stave, BC at 1-day lead. 23 potential predictors (local observations, GFS reforecast data, climate indices) [same as Rasouli et al. 2012]. Data during used to find the optimal number of hidden neurons (3 for ANN and 27 for ELM), and to train the first model. New data arrive (a) weekly, (b) monthly or (c) seasonally. Validate forecasts for models: MLR, OS-MLR, ANN, ELM, OS-ELM Compare correlation (CC), mean absolute error (MAE), root mean squared error (RMSE) and cpu time 24

25

26

Conclusion With new data arriving frequently, OS-ELM provides a much cheaper way to update & maintain a nonlinear regression model. Future research OS-ELM retains the information from all the data it has seen. If data are non-stationary, need to forget the older data. Adaptive OS-ELM has an adaptive forgetting factor. 27