People often want to know how accurate today’s ENSO (El NiñoSouthern Oscillation) predictions are. To provide a flavor for how they have been performing lately, we look at results since 2012. For an assessment of ENSO predictions over 20022012, check out Barnston et al. (2012).
One of my responsibilities as the lead ENSO forecaster at IRI is to judge how well the forecasts have matched reality. One way I do this is I go back through the archived forecasts and make graphics that compare the forecasts with actual sea surface temperature observations. I look for places where they agree and where they don’t, and try to understand what went wrong where they don’t. But eyeballing the graphs is subjective. I’m also interested in mathematically based tests, like correlation and mean error, because they tell me in a more objective fashion just how the forecast fared. The numbers don’t lie!
Each month, predictions for the Niño3.4 index (an SSTbased measure of ENSO) from more than 20 dynamical and statistical (1) models are collected and posted (2), the most recent of which is illustrated in Figure 1.
In this post, I’ll talk about what the prediction models on the plume had to say over the last 2 to 3 years, compared to what actually happened. We will see that the forecasts often provided useful information for the coming few months, but have more modest accuracy and value in forecasting farther into the future. It is particularly hard to predict the timing of ENSO transitions and the correct strength. Clearly, more research and better understanding are still needed.
How Have the Forecasts Compared with Observations in 20122014?
Because models have different errors, using an average of the forecasts of several models is considered better than using any one individual model (Kirtman et al. 2014). To get a flavor for how good or bad the ENSO forecasts have been since 2012, we compare the average forecasts of the dynamical models (Figure 2) and statistical models (Figure 3) to the observed SST anomalies in the Niño3.4 region of the tropical Pacific.
In each figure, the orange line shows the observed Niño3.4 index for overlapping threemonth seasons from JanuaryFebruaryMarch 2012 to JuneJulyAugust 2014. The three other lines show the seasonal predictions made at three different “lead times:” 1 month in advance, 4 months in advance and 7 months in advance (see footnote 3 for more explanation of lead time).
Figure 2, showing the dynamical model performance since 2012, reveals several features. There was a brief period of time around late summer 2012 where the Niño3.4 region was rather warm, and reached thresholds required for El Niño for two to three consecutive seasons—but, it was never considered a fullblown El Niño because it did not persist for very long, and there was no associated atmospheric response (a situation similar to earlier this year). Yet, the dynamical models predicted an El Niño, and only stopped doing so after SSTs in the initial conditions (4) starting cooling. The 4month lead forecasts (bluegreen line) showed a moderatestrength El Niño. However, neutralENSO (neither El Niño or La Niña) prevailed.
During summer to fall 2013, the model predictions were approximately on target. However, the models missed the brief period of cooler Niño3.4 index values during winter 201314 (never reaching a fullblown La Niña).
Beginning in spring 2014 more warming took place, and the forecasts at both moderate (4month) and long (7month) leads reproduced this warming nicely: note the forecasts for MJJ 2014, where the forecasts from all three lead times and the observations coincide at index values of about 0.4°C. Rather unexpectedly though, during summer 2014 the observations cooled, while the forecasts continued to be warm. It remains to be seen whether the model forecasts will be correct in predicting a weak El Niño for the last quarter of 2014 and into 2015. It has been discussed in a previous post how several features this year were dissimilar this year to mid2012, so we should not consider 2012 as a good description of the current ENSO predictions.
Figure 2 also shows that the dynamical forecasts often lag the observations by several months, and this lagging increases as the lead time increases. Such a lag implies a sluggishness of the models in “catching on” to new developments—they may not be predicted until they begin to show up in the observations. This lagging tendency, called “target period slippage” (Barnston et al. 2012), is also seen clearly in the brief warming in summer 2012 (5). Another feature seen in Figure 2 is that the forecasts average slightly warmer conditions than the observations during 20122014.
Figure 3 shows the statistical model performance since 2012. The pattern is similar to that of the dynamical models, with a few key differences:
 the statistical models predicted weaker warming during both the 2012 and the 2014 event,
 the lag in the forecasts relative to the observations is slightly larger in the statistical models (for example, the behavior in late 2012), and
 the statistical forecasts, overall, averaged cooler than the dynamical model forecasts (6).
The slightly larger lag is likely a result of the fact that statistical models are often based on seasonal mean predictor data (i.e., data that is averaged over 3 months), which prevents them from detecting very recent changes in the observed conditions; dynamical models are run with the most recent observed data, and are generally run more than once each month.
Objectively, How Good were the Forecasts?
Here we consider two measurements (among many) that summarize how well the forecasts match with the observations. One is the correlation coefficient, which shows how well the pattern of the forecasts (i.e. up and downs of the time series) follows that of the observations. The correlation coefficient ranges from 1 to +1; +1 means that the observations follow perfectly what was forecast, while 1 means the observations behave exactly the opposite of what was forecast. A coefficient of 0 means that the forecasts and observations show no relationship with one another. Coefficients of 0.5 or more are often considered to show useful forecast information.
The other measure is the mean absolute error, which is the average difference between the forecast and observation (and which one is higher does not matter). Here are the results since 2012 for the two measures at 1, 4 and 7 month lead times, respectively, for the two model types:
Model Type 
Correlation Coefficient 
Mean Absolute Error 


Lead 1 
Lead 4 
Lead 7 
Lead 1 
Lead 4 
Lead 7 
Dynamical 
0.89 
0.60 
0.14 
0.17 
0.32 
0.44 
Statistical 
0.79 
0.46 
0.12 
0.22 
0.29 
0.31 
As one would expect, forecasts made from farther in the past (longer lead times) are less skillful than more recent (shortlead) forecasts, and the 7 month lead forecasts were of little use over this particular period. The dynamical models showed somewhat higher (i.e., better) correlations than the statistical models. The mean absolute error is generally larger for the dynamical models, partly because they averaged too warm during the period, especially when they predicted the warmest SST levels (7). The better correlations of dynamical models were also found in the 11year period of 20022012 (Barnston et al. 2012). Based on the objective performance measures, it is clear that while our ENSO forecasts can be helpful for the coming few months, we have a long way to go in improving their performance and utility beyond that. It is especially hard to predict the timing of ENSO transitions and the correct strength.
Footnotes:
(1) Dynamical prediction models use the physical equations of the ocean and atmosphere to forecast the climate. Statistical models, by contrast, do not use physical equations, but rather statistical formulations that produce forecasts based on a long history of past observations of (a) what is being predicted (e.g., the Niño3.4 SST anomaly), and (b) relevant predictors (e.g., sea level pressure patterns, tropical Pacific subsurface temperatures, or even the Niño3.4 SST itself). It detects and uses systematic relationships between those two sets of data. Statistical models came into existence earlier than dynamical models, because the dynamical ones require high performance computers that have only been around during the last decade or two. Dynamical models, in theory, should deliver the higher accuracy between the two types, but they are still plagued with specific problems that are beyond the scope of this post.
(2) This plume of forecasts is posted on the IRI site on the third Thursday of the month, and is also shown, along with additional ENSO information, in CPC/IRI ENSO Diagnostic Discussion early in the following month. The discussion, in addition to providing a narrative summary of the current and predicted ENSO state, contains an official probability forecast of the ENSO condition for the 9 forthcoming overlapping 3month periods. This forecast is based partially on the forecasts of the plume among several other key inputs, including the expert judgment of a group of forecasters.
(3) An example of a 1month lead forecast is a forecast for JFM 2012 that was developed during the first part of January 2012, based on observed data running through the end of December 2011. Due to the time required for data collection among many different national and international agencies who finish running their models at different times, the 1month lead forecast for JFM 2012 does not appear on the Web until the middle of January even though it does not use data from that month. The 4month lead forecast for JFM 2012 had been issued in the middle of October 2011, and the 7month lead forecast had been made in midJuly 2011. Describing the lead time from the point of view of the time the forecasts are made, the 1, 4, and 7month lead forecasts made in January 2012 were for JFM, AMJ and JAS 2012, respectively. The month from which forecasts are made is often called the start month, and the season that the forecast is for is often called the target season.
(4) Dynamical models begin their forecast computations using the latest observations. These observations include many fields (e.g., SST, sea temperature below the surface, atmospheric pressure at the surface and extending into the upper atmosphere, winds, humidity); all of these together are called the initial conditions.
(5) The lack of a lag in the predictions of the 2014 warming may be due to the subsurface sea temperatures having warmed greatly early in the year, serving to predict surface warming at a later time (Meinen and McPhaden 2000; McPhaden 2003).
(6) The somewhat muted warming in the statistical predictions is due to the typically conservative behavior of statistical models, as they often seek to minimize squared errors. This would also show up as somewhat muted cooling during episodes of La Niña.
(7) Note that the mean absolute error may be degraded without the correlation being degraded. For example, if the forecasts are always far too high (but by about the same amount), or if the forecasts are always far too strong, both for warm and for cold episodes (but by about the same factor), the correlation can still be very high but the mean absolute error may be terrible (very high). But if the correlation is poor, it is difficult for the mean absolute error to look good either.
References
Barnston, A. G., M. K. Tippett, M. L. L’Heureux, S. Li, and D. G. DeWitt, 2012: Skill of realtime seasonal ENSO model predictions during 200211: Is our capability increasing? Bull. Amer. Meteor. Soc., 93, 631651 (free access).
Kirtman et al., 2014: The North American Multimodel Ensemble: Phase1 seasonal to interannual prediction; Phase 2 toward developing intraseasonal prediction. Bull. Amer. Meteor. Soc., 95, 585601 (free access).
McPhaden, M. J., 2003: Tropical Pacific Ocean heat content variations and ENSO persistence barriers. Geophys. Res. Lett., 30, 2003. DOI: 10.1029/2003GL016872.
Meinen, C. S., and M. J. McPhaden, 2000: Observations of warm water volume changes in the equatorial Pacific and their relationship to El Niño and La Niña. J. Climate, 13, 35513559.
ENSO Blog
A blog about monitoring and forecasting El Niño, La Niña, and their impacts.
Disclaimer:
The ENSO blog is written, edited, and moderated by Michelle L’Heureux (NOAA Climate Prediction Center), Emily Becker (University of Miami/CIMAS), Nat Johnson (NOAA Geophysical Fluid Dynamics Laboratory), and Tom DiLiberto and Rebecca Lindsey (contractors to NOAA Climate Program Office), with periodic guest contributors.
Ideas and explanations found in these posts should be attributed to the ENSO blog team, and not to NOAA (the agency) itself. These are blog posts, not official agency communications; if you quote from these posts or from the comments section, you should attribute the quoted material to the blogger or commenter, not to NOAA, CPC, or Climate.gov.
Comments
Autocorrelation
autocorrelation for lead time =1, r = 0.81, therefore, the dynamical model is only improving the forecast by 1%.
autocorrelation for lead time =4, r = 0.14, therefore, the dynamical model is improving the forecast by 55%.
if lead time autocorrelation coefficients 1 4 are averaged,however, r = 0.28, therefore, the dynamical model is improving the forecast by only 10%
is it true that autocorrelation at lead times =1, 2 is potentially overstating the dynamical models ability to predict ENSO?
reference
Alessandri and Navarra, 2008
RE: Autocorrelation
It is true that considerably predictability exists by virtue of persistence alone. The models attempt to improve upon this baseline reference. If the persistence (or autocorrelation) score is regarded as a zeroskill baseline, then the skill of the models does not appear very impressive. But one might challenge use of persistence as a zeroskill baseline, when it can also be considered as part of the forecast, and therefore have value in its own right.
enso forecasts
is there a month when the models show more skill in their forecasts?
RE: enso forecasts
Model skill for El Nino forecasts are typically lowest during the spring, with maximum skill during the Northern Hemisphere fall and early winter.
Historical Forecast data
Hello Sir  If we were to find historical Forecast data ( suppose for 4 month lead time , forecast made on JAX,20XX) , how could we find it !! Is there any data library either in NOAA or other org. !! We are trying to do the same exercise for accuracy of their predictions for last 10 years.
RE: Historical Forecast data
Hi, there are a number of options for scoring the ENSO forecasts. One possibility is to use the forecasts issued by various centers and shown in the IRI/CPC ENSO prediction plume since 2002. Forecasts for 3month SST anomaly in the Nino3.4 region (often used as the SST aspect of ENSO) are shown in the file found at: http://iri.columbia.edu/~forecast/ensofcst/Data/ and then click on the second link in the list. These show departures from average (anomalies), not actual SST. A list of many models is shown for each foreast start time, from 2002 until April 2017. The numbers are given in hundredths of degree C. The lead time increases as you move to the right in each row. The 3month season in the first and the last column is given in the heading text for each set of forecasts. The models change over time  some quit, some change version number, and some new ones come along. Now, for the observations, you can use NOAA's SST observations, and there's more than one SST dataset. The one I often use is the one given in the table on the site: http://www.cpc.ncep.noaa.gov/data/indices/sstoi.indices which gives 1month SST and SST anomaly. The data for the Nino3.4 region is farthest to the right. Another possible SST dataset from NOAA is found at http://www.cpc.ncep.noaa.gov/data/indices/oni.ascii.txt and this one is already averaged into 3month periods for you, saving you the trouble of doing that, and it is only for Nino3.4 region. If you do not want to verify these model forecasts, but want to verify the CPC/IRI monthly ENSO probabilistic forecasts, you only have several years of these. If you do want those, please email Michelle L'Heureux at Michelle.LHeureux@noaa.gov about obtaining those. At the web site: http://www.cpc.ncep.noaa.gov/products/analysis_monitoring/enso_advisory/... I only see the current Discussion and forecast, and don't know how to get older ones going back to 2011 or 2012 or so. Good luck with your scoring project!
Add new comment