Download presentation
Presentation is loading. Please wait.
1
01-12-20001 Deriving observational constraints on climate model predictions Gabriele Hegerl, GeoSciences, University of Edinburgh
2
01-12-20002 The Problem Climate model predictions are uncertain, and quantifying these uncertainties is essential for useful predictions Only observations can really constrain predictions – so attempts to arrive at probabilistic predictions make use of observations in some form There are a number of ways to do that, depending on the problem, timescale, information available, climate variable….
3
01-12-20003 Prediction uncertainty 1. Internal variability uncertainty: weather/climate variability not/not entirely predictable beyond days => uncertainty, can be estimated 2. Forcing uncertainty: Future emissions unknown – scenarios; volcanoes? Sun? 3. Model uncertainty: uncertainty due to unknown physics and unknown parameters in models, structural errors, missing processes… 3 is mainly what we try to estimate, although some recent work also tries to predict 1
4
01-12-20004 Forcing uncertainty, model uncertainty and internal climate variability also vary with timescale From Hawkins and Sutton, 2009 BAMS: fraction of uncertainty due to climate variability, model uncertainty, forcing uncertainty and model uncertainty
5
01-12-20005 How predictons are constrained depends on timescale Nearterm: initial conditions MAY matter Intermediate transient warming Longterm: Equilibrium climate sensitivity From IPCC AR4, CH10 (Meehl et al.)
6
01-12-20006 Why? A very simple Energy Balance Model: Held et al., 2010 Total radiative forcing G Surface Temperature change T Change in outgoing radiation T Exchange of heat into deep ocean H Equation for surface ocean with heat capacity C F Equation for deep ocean with heat capacity C D H cFcF cDcD
7
01-12-20007 There are two distinct timescales Fast response: deep ocean has not yet significantly taken up heat (T D )=0 Response time: Dominated by transient climate response timescale for the deep ocean is much slower. Equilibrium climate sensitivity reached once ocean takes up no more heat This works well for the GFDL model, with transient near term warming almost completely dominated by the first case H cFcF cDcD
8
01-12-20008 Lets start with the transient climate change in the 21rst century ‘naïve approach’: models that do well over the 20 th century will do well over the future But: external forcings driving this are uncertain… Greenhouse gas forcing is quite well known, but sulfate aerosol forcing, other anthropogenic forcing (BC etc) poorly, solar not well know either models can agree with data because they are correct, but they could also agree because of cancelling errors We may also wind up rejecting models that are correct but their forcing was wrong…(CMIP5)
9
01-12-20009 Evaluation of climate model ability to simulate an observed change models can agree with data because they are correct, but they could also agree because of cancelling errors We may also wind up rejecting models that are correct but their forcing was wrong… This is happening for some CMIP5 simulations IPCC SPM
10
01-12-200010 What we need to do is: - Identify what is the response to individual forcings that influenced 20 th century climate - Project forward the components that are predictable or projectable (eg greenhouse gas increases) - Fingerprints can separate the contribution by different external drivers because of different physics of forcing -Eg: solar warms entire atmosphere -Aerosols have different spatial pattern and temporal evolution than greenhouse gases -Volcanoes have pronounced shortterm impact Pinatubo, 1991 El Chichón, 1982
11
01-12-200011 Ingredients for detection and attribution o Observation y o Climate change signals (“fingerprints”) X=(x i ),i=1..n typically from model simulation: one for each fingerprint, or for a small number of combinations o Noise: data for internal climate variability, usually from a long model control simulation o If X contains realizations of climate variability, a total least square fit can be used (Allen and Stott)
12
01-12-200012 Observed amplitude estimate Signal amplitude scalar product can reduce noise if using inverse noise covariance Observational uncertainty: -Use model data only where observations exist – like with like -Use samples of observations Uncertainty in determined by superimposing samples of climate variability onto fingerprint (bit more complicated for tls)
13
01-12-200013 Observed amplitude estimate Signal amplitude scalar product can reduce noise if: using inverse noise covariance Test residual consistent with properties of model noise. This alarms to model problems and sampling problems in C -1 ! Uncertainty in determined by superimposing samples of climate variability onto fingerprint (bit more complicated for tls)
14
01-12-200014 Attribution results yields range of scaling factors that are consistent with observed change Scaling factors that show which range of up-or downscaling of model response is consistent with observations Warming due to greenhouse gases Fig. 9.9c, Hegerl et al., 2007 Greenhouse gases; other anthropogenic, natural (solar+volcanic)
15
01-12-200015 Fig 9.21 after Stott et al. This translates directly to transient climate response: Estimated warming at the time of CO 2 doubling in response to a 1% per year increase in CO 2 Constrained by observed 20th century warming via the estimated greenhouse gas signal Observational constraints suggest … very likelyvery likely >1°C very unlikelyvery unlikely >3.5°C very unlikelysupports the overall assessment that TCR very unlikely >3°C
16
01-12-200016 Probabilistic prediction of global mean change: one of the pdfs is based on TCR, others on other obs. constraints Scalability: the pattern does not change much with signal strength
17
01-12-200017 Other approaches: select models based on observational feedbacks Hall et al., 2006: spring albedo against temperature change vs seasonal cycle Using information from climate model ‘quality’ ie ability to simulate mean climate and short term variations used particularly for regional scales – needs to be demonstrated that it is relevant for predictions Use of observations: compare like with like; sample as observations do, not models do; bring models to observations
18
01-12-200018 Can we get closer with initial conditions for the near term: Can climate change be predicted based on initial condition just like weather? => initial value problem Initialization is not easy Large ensemble of such runs done right now for nearterm Problem: evidence for useful predictability beyond a year or few is weak, particularly for things that matter (regional climate) Smith et al., 2006, science Top: 1yr, middle 9 yr, bottom ave 1-9 yrs, 5- 95% ranges
19
01-12-200019 EQUIP: End to End Quantification of Impacts Prediction Heatwaves: Summer maximum daily temperature – predictions capture not only trend, but some of the structure (is this just plain lucky?) Hanlon and Hegerl, in prep. EQUIP lead by Andy Challinor, Leeds Other work: Crop predictions, Water deficit, fisheries
20
01-12-200020 How to First: Bias correct the model Needs to be done differently for hot extremes than mean Is there any added value from initial conditions? – not clear based on correlation
21
01-12-200021 Use skill score based on Murphy 1988 Forecast system Y vs W (eg climatology, noninitialized) MSSS can be composed into correlation, conditional bias and mean bias Updated to compare against NoAssim (no ICs) rather than persistence (Goddard et al., in prep)
22
01-12-200022 Decadal predictions raise many questions Is there skill that is statistically significant How can the prediction be quantified – arrive at uncertainty ranges How long does skill last So far: predictability largely for fish…
23
01-12-200023 Approach for long term: Estimating climate system properties Equilibrium climate sensitivity First: example single line of evidence
24
01-12-200024 1. What can we learn about climate sensitivity from the last millennium? Decadal NH 30-90N land temperature; Hegerl et al., J Climate, 2007
25
01-12-200025 CH-blend reconstruction Weighted average of decadal records, many treering data (RSS) Calibration: Total least square scales communality between instrumental and proxy data to same size Method tested with climate model data to assess if uncertainties estimated properly
26
01-12-200026 Climate forcing over the last millennium Northern Hemispheric 30-90N mean radiative forcing (decadally smoothed) from Crowley Uncertainties: ~ 40% in amplitude of volcanic forcing Large in amplitude and shape of solar forcing And in aerosol forcing
27
01-12-200027 Simulate observed climate change not with a single best fit, but a large ensemble of model simulations with different sensitivities Determine probability of models in agreement with data, given: internal variability, uncertainty in data, uncertainty in model ECS [K] p Miss uncertainties: too narrow Use information incompletely: too wide 1.Estimating equilibrium climate sensitivity
28
01-12-200028 Estimating ECS Run EBM with > 1000 model simulations, varying ECS, effective ocean diffusivity, and aerosol forcing Residuals between reconstruction and range of EBM simulations with different climate sensitivity Var(Res-res min ) ~ F(k,l) (after Forest et al., 2001) Uncertainties included: Calibration uncertainty of reconstruction Noise and internal variability Uncertainty in magnitude of past solar and volcanic forcing Uncertainties: Simple representation of efficacy Systematic biases in reconstructions
29
01-12-200029 Estimated PDF for climate sensitivity Result for different reconstructions, 13 th century to 1850 Larger amplitude Smaller forcing Nonlinear relationship sensitivity – volcanic cooling Response small ~ climate variability
30
01-12-200030 Results vary between reconstructions
31
01-12-200031 2. Multiple lines of evidence Hegerl et al. 06, nature Bayesian update, using a prior pdf based on late 20 th century (Frame et al) Multiple lines of evidence reduce probability of high sensitivity, and of very small sensitivity Independence? Proper error estimate?
32
01-12-200032 Schneider von Deimling et al. 2006 Problem: other models do not show such a linear relationship
33
01-12-200033 Estimates from many different sources Difficult question: can these estimates strengthen each other? How? Knutti and Hegerl, 2008
34
01-12-200034 Towards improving use of observations for constraints on predictions Make sure that the model is brought as closely to the observational product as possible: synthetic satellite data; synthetic palaeo data Use uncertainty estimates; models can be used to test processing uncertainty Be wary of spatial and temporal autococrrelation
35
01-12-200035 If using observations to provide constraints on predictions Relevance of observed evidence for prediction needs to be established Uncertainty in observations needs to be included in estimate Approach will vary with timescale Many questions remain: Decadal predictions with reasonable uncertainty How to predict regional changes? How to combine information from different sources into an overall estimate of uncertainty of ECS
36
01-12-200036
37
01-12-200037 “Perfect model study” “observation” from simulation, sens 5, with noise added Effect of noise on forcing
38
01-12-200038 Fingerprints for external forcing Simulated with 2-D UD-EBM, 30-90N annual, sensitivity 2.5 o C
39
01-12-200039 Future change uncertainty ranges reflect uncertainty in transient response Are based, among other things, on observational constraints This is a significant advance SPM Fig. 5 Likely (>66%) range
40
01-12-200040 Example Make predictions with many models by systematically varying uncertain parameters (climateprediction.net) But simply accepting all of them will surely overestimate uncertainty The sample will also underestimate it – all versions of this model share a common structure Stainforth et al., 2005
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.