Incorporating Multi-model Ensemble Techniques into a Probabilistic Hydrologic Forecasting System: Relative Merits of Ensemble vs. Bias-Corrected Models.

Slides:



Advertisements
Similar presentations
Introduction The agricultural practice of field tillage has dramatic effects on surface hydrologic properties, significantly altering the processes of.
Advertisements

QbQb W2W2 T IPIP Redistribute W 0 W 1 and W 2 to Crop layers Q W1W1 ET 0, W 0, W 1, W 2 I T from 0, 1 & 2, I P A Coupled Hydrologic and Process-Based Crop.
Alan F. Hamlet Andy Wood Dennis P. Lettenmaier JISAO Center for Science in the Earth System Climate Impacts Group and Department of Civil and Environmental.
Alan F. Hamlet Andy Wood Dennis P. Lettenmaier JISAO Center for Science in the Earth System Climate Impacts Group and Department of Civil and Environmental.
A Macroscale Glacier Model to Evaluate Climate Change Impacts in the Columbia River Basin Joseph Hamman, Bart Nijssen, Dennis P. Lettenmaier, Bibi Naz,
The Calibration Process
Alan F. Hamlet Andy Wood Seethu Babu Marketa McGuire Dennis P. Lettenmaier JISAO Climate Impacts Group and the Department of Civil Engineering University.
Ensemble Post-Processing and it’s Potential Benefits for the Operational Forecaster Michael Erickson and Brian A. Colle School of Marine and Atmospheric.
Colorado Basin River Forecast Center Water Supply Forecasting Method Michelle Stokes Hydrologist in Charge Colorado Basin River Forecast Center April 28,
Current Website: An Experimental Surface Water Monitoring System for Continental US Andy W. Wood, Ali.
Andy Wood, Ted Bohn, George Thomas, Ali Akanda, Dennis P. Lettenmaier University of Washington west-wide experimental hydrologic forecast system OBJECTIVE.
Effect of Model Calibration on Streamflow Forecast Results Ali Akanda, Andrew Wood, and Dennis Lettenmaier Civil and Environmental Engineering University.
Figure 1: Schematic representation of the VIC model. 2. Model description Hydrologic model The VIC macroscale hydrologic model [Liang et al., 1994] solves.
Hydrologic Modeling: Verification, Validation, Calibration, and Sensitivity Analysis Fritz R. Fiedler, P.E., Ph.D.
Evaluation of the Surface Water Balance of Southeast Asia from a Land Surface Model and ERA40 Reanalysis Mergia Y. Sonessa 1, Jeffrey E. Richey 2 and Dennis.
Experimental seasonal hydrologic forecasting for the Western U.S. Dennis P. Lettenmaier Andrew W. Wood, Alan F. Hamlet Climate Impacts Group University.
Assessment of Hydrology of Bhutan What would be the impacts of changes in agriculture (including irrigation) and forestry practices on local and regional.
Multi-Model Estimates of Arctic Land Surface Conditions Theodore J. Bohn 1, Andrew G. Slater 2, Dennis P. Lettenmaier 1, and Mark C. Serreze 2 1 Department.
Global Flood and Drought Prediction GEWEX 2005 Meeting, June Role of Modeling in Predictability and Prediction Studies Nathalie Voisin, Dennis P.
Aihui Wang, Kaiyuan Li, and Dennis P. Lettenmaier Department of Civil and Environmental Engineering, University of Washington Integration of the VIC model.
Understanding hydrologic changes: application of the VIC model Vimal Mishra Assistant Professor Indian Institute of Technology (IIT), Gandhinagar
Andy Wood, Alan Hamlet, Seethu Babu, Marketa McGuire and Dennis P. Lettenmaier A West-wide Seasonal to Interannual Hydrologic Forecast System OVERVIEW.
Efficient Methods for Producing Temporally and Topographically Corrected Daily Climatological Data Sets for the Continental US JISAO/SMA Climate Impacts.
A Multi-Model Hydrologic Ensemble for Seasonal Streamflow Forecasting in the Western U.S. Theodore J. Bohn, Andrew W. Wood, Ali Akanda, and Dennis P. Lettenmaier.
Drought Prediction (In progress) Besides real-time drought monitoring, it is essential to provide an utlook of what future might look like given the current.
Potential for medium range global flood prediction Nathalie Voisin 1, Andrew W. Wood 1, Dennis P. Lettenmaier 1 1 Department of Civil and Environmental.
MSRD FA Continuous overlapping period: Comparison spatial extention: Northern Emisphere 2. METHODS GLOBAL SNOW COVER: COMPARISON OF MODELING.
Sources of Skill and Error in Long Range Columbia River Streamflow Forecasts: A Comparison of the Role of Hydrologic State Variables and Winter Climate.
Assessing the Influence of Decadal Climate Variability and Climate Change on Snowpacks in the Pacific Northwest JISAO/SMA Climate Impacts Group and the.
ABSTRACT Since the 1930's, combined streamflow from the six largest Eurasian rivers discharging to the Arctic Ocean has been increasing. For many of these.
Alan F. Hamlet Andy Wood Dennis P. Lettenmaier JISAO Center for Science in the Earth System Climate Impacts Group and the Department.
Performance Comparison of an Energy- Budget and the Temperature Index-Based (Snow-17) Snow Models at SNOTEL Stations Fan Lei, Victor Koren 2, Fekadu Moreda.
Implementing Probabilistic Climate Outlooks within a Seasonal Hydrologic Forecast System Andy Wood and Dennis P. Lettenmaier Department of Civil and Environmental.
Drought and Model Consensus: Reconstructing and Monitoring Drought in the US with Multiple Models Theodore J. Bohn 1, Aihui Wang 2, and Dennis P. Lettenmaier.
DOWNSCALING GLOBAL MEDIUM RANGE METEOROLOGICAL PREDICTIONS FOR FLOOD PREDICTION Nathalie Voisin, Andy W. Wood, Dennis P. Lettenmaier University of Washington,
VERIFICATION OF A DOWNSCALING SEQUENCE APPLIED TO MEDIUM RANGE METEOROLOGICAL PREDICTIONS FOR GLOBAL FLOOD PREDICTION Nathalie Voisin, Andy W. Wood and.
EVALUATION OF A GLOBAL PREDICTION SYSTEM: THE MISSISSIPPI RIVER BASIN AS A TEST CASE Nathalie Voisin, Andy W. Wood and Dennis P. Lettenmaier Civil and.
From catchment to continental scale: Issues in dealing with hydrological modeling across spatial and temporal scales Dennis P. Lettenmaier Department of.
Long-lead streamflow forecasts: 2. An approach based on ensemble climate forecasts Andrew W. Wood, Dennis P. Lettenmaier, Alan.F. Hamlet University of.
Machiavellian Forecasting: do the ends justify the means? Ted Bohn UW CEE UBC/UW 2005 Hydrology and Water Resources Symposium.
Current WEBSITE: Experimental Surface Water Monitor for the Continental US Ali S. Akanda, Andy W. Wood,
Evaluation of TRMM satellite precipitation product in hydrologic simulations of La Plata Basin Fengge Su 1, Yang Hong 2, and Dennis P. Lettenmaier 1 1.
LSM Hind Cast for the Terrestrial Arctic Drainage System Theodore J. Bohn 1, Dennis P. Lettenmaier 1, Mark C. Serreze 2, and Andrew G. Slater 2 1 Department.
Upper Rio Grande R Basin
RESULTS and DISCUSSION
Use of Extended Daily Hydroclimatalogical Records to Assess Hydrologic Variability in the Pacific Northwest Department of Civil and Environmental Engineering.
American Geophysical Union Fall Meeting (December 2004)
Model-Based Estimation of River Flows
Streamflow Simulations of the Terrestrial Arctic Regime
Dennis P. Lettenmaier, Andrew W. Wood, Ted Bohn, George Thomas
Nathalie Voisin, Andy W. Wood and Dennis P. Lettenmaier
Arctic CHAMP Freshwater Initiative Conference (June )
Hydrologic ensemble prediction - applications to streamflow and drought Dennis P. Lettenmaier Department of Civil and Environmental Engineering And University.
Multimodel Ensemble Reconstruction of Drought over the Continental U.S
Kostas M. Andreadis1, Dennis P. Lettenmaier1
Hydrology and Water Management Applications of GCIP Research
Andy Wood and Dennis Lettenmaier
Long-Lead Streamflow Forecast for the Columbia River Basin for
Effects of Temperature and Precipitation Variability on Snowpack Trends in the Western U.S. JISAO/SMA Climate Impacts Group and the Department of Civil.
A. Wood, A.F. Hamlet, M. McGuire, S. Babu and Dennis P. Lettenmaier
Model-Based Estimation of River Flows
Andy Wood and Dennis P. Lettenmaier
A Multimodel Streamflow Forecasting System for the Western U.S.
Results for Basin Averages of Hydrologic Variables
Andrew W. Wood Dennis P. Lettenmaier
A Multimodel Drought Nowcast and Forecast Approach for the Continental U.S.  Dennis P. Lettenmaier Department of Civil and Environmental Engineering University.
Evaluation of the TRMM Multi-satellite Precipitation Analysis (TMPA) and its utility in hydrologic prediction in La Plata Basin Dennis P. Lettenmaier and.
Dennis P. Lettenmaier Andrew W. Wood, and Kostas Andreadis
Multimodel Ensemble Reconstruction of Drought over the Continental U.S
Results for Basin Averages of Hydrologic Variables
Presentation transcript:

Incorporating Multi-model Ensemble Techniques into a Probabilistic Hydrologic Forecasting System: Relative Merits of Ensemble vs. Bias-Corrected Models Mergia Y. Sonessa, Theodore J. Bohn and Dennis P. Lettenmaier Department of Civil and Environmental Engineering, Box , University of Washington, Seattle, WA American Geophysical Union Fall Meeting, December 2008 ABSTRACT Multi-model ensemble techniques have been shown to reduce bias and to aid in quantification of the effects of model uncertainty in hydrologic modeling. These techniques are only beginning to be applied in operational hydrologic forecast systems. Much of the analyses (e.g. Ajami et al., 2006) that have been performed to date have focused on daily data over short durations, with constant ensemble model weights and constant bias correction parameters (if bias correction is applied at all). However, typical hydrologic forecasts can involve monthly flow volumes, lead times of several months, various probabilistic forecast techniques, and monthly bias corrections. Under these conditions the question arises as to whether a multi-model ensemble is as effective in improving forecast skill as under the conditions that have been investigated so far. To investigate the performance of a multi-model ensemble in the context of probabilistic hydrologic forecasting, we have extended the University of Washington's West-wide Seasonal Hydrologic Forecasting System to use an ensemble of three models: the Variable Infiltration Capacity (VIC) model version 4.0.6, the NCEP NOAH model version 2.7.1, and the NWS grid-based Sacramento/Snow-17 model (SAC). The objective of this presentation is to assess the performance of the ensemble of the three models as compared to the performance of the models individually, with and without various forms of bias correction, and in both retrospective and probabilistic forecast modes. Three forecast points within the West-wide forecast system domain were used for this research: the Feather River at Oroville, CA, the Salmon River at White horse, ID, and the Colorado River at Grand Junction. The forcing and observed streamflow data are for years for the Feather and Salmon Rivers; and for the Colorado. The models were first run for the retrospective period, then bias-corrected, and model weights were then determined using unconstrained multiple linear regression as a best-case scenario for the ensemble. We assessed the performance of the ensemble in comparison with the individual models in terms of correlation with observed flows (R), Root Mean Square Error (RMSE), and Coefficient of Prediction (Cp). To test forecast skill, we performed Ensemble Streamflow Prediction (ESP) forecasts for each year of the retrospective period, using forcings from all other years, for individual models and for the multi-model ensemble. To form the ensemble for the ESP runs, we used the model weights from the retrospective simulations. In both the retrospective and the probabilistic forecast cases, we found that a monthly bias correction applied to an individual model generally makes the individual model competitive with the multi-model ensemble. For ensemble methods other than unconstrained multiple least squares regression, ensemble performance tends to be worse, with individual bias-corrected models sometimes outperforming the ensemble. It should be noted that the entire time series was used for determining bias correction parameters, ensemble model weights, and assessing forecasts, i.e. the calibration and validation periods were the same. The relative benefits of monthly bias correction vs. multimodel ensemble in a validation period outside the training period remain to be investigated. 4 CONCLUDING REMARKS  Performance improvement due to a monthly bias correction can be much larger than for a multi-model ensemble, when dealing with monthly data from hydrologic models.  Forming a multi-model ensemble from models to which a monthly bias correction has been applied yields little improvement over the individual models.  The reason for this is that a monthly bias correction, performed on monthly data, tends to increase model collinearity.  Those months in which the multi-model ensemble yielded the largest improvements over individual models were those for which model errors were least correlated.  These relationships held for both retrospective simulations and probabilistic forecasts. 1 Models The models in our ensemble all share the same basic structure, consisting of grid cells containing a multi-layer soil column overlain by one or more “tiles” of different land covers, including vegetation with and without canopy and bare soil. Water and/or energy fluxes are tracked vertically throughout the column from the atmosphere through the land cover to the bottom soil layer. The figure below illustrates these features as implemented in the VIC (Variable Infiltration Capacity) macroscale land surface model (Liang et al., 1994). Basins used in this Study Model Forcings and Parameters Land Surface Parameters VIC LDAS parameters (Maurer et al, 2002) followed by iterative calibration of runoff and baseflow parameters by the UA-SCEM method NOAH NLDAS parameters (Mitchell et al, 2004) but with maximum snow albedo replaced by 0.85, to better match snow melt signatures SAC NLDAS (Mitchell et al, 2004) followed by iterative calibration of LZTWM, LZFPM, PFREE, LZSK, ASIMP, and UZTWM by the UA-SCEM method Forcing degree LDAS forcings (Maurer et al, 2002) Disaggregated to 3-hourly time step 2 ESP Run – Monthly Forecast R and Cp for all start months Models at a glance VIC Physically-based horizontal soil layers Energy balance 2-layer snow pack Elevation bands NOAH Physically-based horizontal soil layers Energy balance Single-layer snow pack No elevation bands Sacramento/SNOW17 (SAC) Conceptually-based soil storages No energy balance Elevation bands Degree-day snow melt scheme No explicit vegetation Potential Evapotranspiration computed by NOAH is an input 3 Retrospective Run – Whole-Timeseries Metrics Retrospective Run – Monthly Metrics References  Ajami, N. K., Q. Duan, X. Gao, and S. Sorooshian (2006), Multi-model combination techniques for hydrological forecasting: Application to Distributed Model Intercomparison Project results, J. Hydrometeorol., 7(4), 755– 768.  Liang, X., D. P. Lettenmaier, E. F. Wood, and S. J. Burges, A Simple hydrologically Based Model of Land Surface Water and Energy Fluxes for GSMs, J. Geophys. Res., 99(D7), 14,415-14,428,  Mitchell, et al The Multi-institution North American Land Data Assimilation system (NLDAS): utilization of multiple GCIP products and partners in a continental distributed hydrological modeling system. Journal of Geophysical Research 109, doi: /2003JD Salmon River Colorado River at Grand Junction Feather River Scenarios Considered for Comparing Multimodel Ensemble (ENS) with Individual Models We first examine performance of the models and ensemble over the entire timeseries, as is typical of other studies such as Ajami et al. (2006). Previous studies have employed model weights and bias correction parameters (if any) that were constant in time. Here we investigate performance of both constant and monthly-varying model weights and model parameters, in the following four cases: Case 1: Raw model output Constant ensemble model weights Case 2: Constant bias correction Constant ensemble model weights Case 3: Raw model output Monthly ensemble model weights Case 4: Monthly bias correction Monthly ensemble model weights Assessment: Bias Correction v. Ensemble Constant Parameters 1. No Bias Correction The multimodel ensemble showed small improvement over the best individual model, in terms of RMSE and correlation with observed flow 2. Bias Correction The constant bias correction gave moderate reduction in RMSE for individual models and for the ensemble, but little improvement in correlation with observed flow Ensemble performance only slightly better than the best model Monthly Parameters 3. No Bias Correction Allowing model weights to vary monthly improved the ensemble performance substantially over the best individual model, in terms of both RMSE and correlation with observed flow, especially in the Salmon River 4. Bias Correction Applying a monthly bias correction to the individual models resulted in dramatic reductions in the models’ RMSE and substantial improvements in their correlations with observed flows Despite the improvements in individual model performance, the ensemble average improved only slightly Improvements in individual model performance made even the worst individual model competitive with the ensemble Comparison of hydrographs Fig shows a sample of the hydrographs for the cases of monthly- varying model weights, with (b) and without (a) monthly bias correction. It can be seen that the monthly bias correction removes differences in model timing of peak flows (e.g. circle “1”) arising from differences in the model snow pack formulations. In addition, the monthly bias correction removes systematic bias shared by all models (e.g. circle “2”). Ensemble Performance, Bias, and Model Collinearity These results can be explained by examining model collinearity and its sensitivity to bias correction. A multi-model ensemble benefits from having some degree of model independence. In particular, the more independent model errors are, the more likely they are to cancel out in the ensemble. Fig 4.3 shows the norms of the matrices of model error cross-correlation, model error covariance (scaled by observed flow variance) and model cross-correlation, for all 4 cases outlined above. Applying a constant bias correction has little effect on the correlation and covariance of model errors or the total cross correlation of the models. However, applying the monthly bias correction tends to increase error and model correlations, making them more collinear. This leaves less random error for the ensemble to operate on. 4.1 RMSE of Models and Ensemble vs Observations 4.2 Correlation of Models and Ensemble with Observations COLORADO FEATHR SALMON Raw Bias Corrected Bias Corrected - Raw Colorado RawBias Corrected 4.3 Error Correlation and Covariance, and Model Cross Correlation Constant Parameters COLORADO FEATHER SALMON 4.4 Comparison of Hydrographs 6 5 Ensemble Performance vs. Collinearity We can gain further insight if we examine the monthly statistics of the models and ensemble with and without monthly bias correction. Fig. 5 shows mean flow, correlation with observed flow, RMSE, and the norms of the matrices of error correlation, error covariance, and model cross-correlation with (b) and without (a) monthly bias correction for the Colorado basin, as well as the difference between the two cases (c). Bias correction in this case was obtained by fitting 3-parameter log normal distributions to the observed and simulated flows, tranforming them all to log space, adjusting the means and variances of the simulations to match the observed flows, and transforming them all back to flow space.. Ensemble performance varies from month to month, sometimes being worse than the best individual model. It can be seen that those months for which the ensemble performance is better than the best models are those months having relatively low error correlation and model cross correlation. From the plots in the difference column, we can see that the monthly bias correction tends to increase the cross correlation of model errors and model flows, particularly in the winter months in the Colorado basin. Results are similar in the other two basins. Flow, cms Error Cor Model XCor Error Cov RMSE R Colorado VICSACNOAHENSOBS Month In all cases, the bias correction is derived from quantile mapping between simulated and observed flows, and the ensemble is formed from an unconstrained multiple linear regression between the observed flows and the simluated flows. a b Assessment: Bias Correction v. Ensemble Here we investigate bias correction and multi-model ensembles in probabilistic forecasts, using the ensemble streamflow prediction (ESP) method. For each year of the same 52-year retrospective period ( ) analyzed in sections 4 and 5, we created an ensemble of forcings (the ESP ensemble) composed of the forcings from the other 51 years. Models were started from each of the 12 possible starting months within the forecast year, and ran for 1 year. Thus, each month of the 52-year period was forecast with lead times ranging from 1 to 12 months. For each combination of forecast month and lead time, the forecast skill of the ESP ensemble mean was assessed across all 52 years in terms of coefficient of prediction (CP) (also called Nash-Sutcliffe Efficiency) and correlation with observations (R). This was performed for the individual models’ ESP means and the multi-model ensemble of the individual model ESP means. The bias correction and multi-model ensemble parameters used were the same ones derived in the retrospective simulations discussed in section 5. We show the results for the Colorado basin above; the other basins exhibited similar behavior. No Bias Correction, Monthly Ensemble Weights The multi-model ensemble generally performed better than the individual models, in terms of CP. This resulted in turn from reduction of RMSE. The multi-model ensemble generally did not perform any better than the best model in terms of correlation with observations (R) Monthly Bias Correction & Ensemble Weights Applying a monthly bias correction to the individual models yielded substantial improvements in CP, due to reduction of RMSE. Little change was seen in R. Individual models with a monthly bias correction outperformed both uncorrected models and the ensemble of uncorrected models. Forming a multi-model ensemble from bias corrected models yielded little or no improvement in CP or R in most forecast month/lead time combinations. Months for which error correlation among models was low (circles 1-3) corresponded to large improvements in multi-model ensemble over individual models. An exception was for forecasts in summer months (circle 4)