 Original article
 Open Access
 Published:
Model confidence sets and forecast combination: an application to agespecific mortality
Genus volume 74, Article number: 19 (2018)
Abstract
Background
Model averaging combines forecasts obtained from a range of models, and it often produces more accurate forecasts than a forecast from a single model.
Objective
The crucial part of forecast accuracy improvement in using the model averaging lies in the determination of optimal weights from a finite sample. If the weights are selected suboptimally, this can affect the accuracy of the modelaveraged forecasts. Instead of choosing the optimal weights, we consider trimming a set of models before equally averaging forecasts from the selected superior models. Motivated by Hansen et al. (Econometrica 79(2):453–497, 2011), we apply and evaluate the model confidence set procedure when combining mortality forecasts.
Data and methods
The proposed model averaging procedure is motivated by Samuels and Sekkel (International Journal of Forecasting 33(1):48–60, 2017) based on the concept of model confidence sets as proposed by Hansen et al. (Econometrica 79(2):453–497, 2011) that incorporates the statistical significance of the forecasting performance. As the model confidence level increases, the set of superior models generally decreases. The proposed model averaging procedure is demonstrated via national and subnational Japanese mortality for retirement ages between 60 and 100+.
Results
Illustrated by national and subnational Japanese mortality for ages between 60 and 100+, the proposed modelaveraged procedure gives the smallest interval forecast errors, especially for males.
Conclusion
We find that robust outofsample point and interval forecasts may be obtained from the trimming method. By robust, we mean robustness against model misspecification.
Introduction
Because of declining mortality rates in mainly developed countries, the improvement in human survival probability contributes significantly to an aging population. As a consequence, pension funds and insurance companies face longevity risk. The longevity risk is a potential risk attached to the increasing life expectancy of policyholders, which can eventually result in a higher payout ratio than expected (Crawford et al. 2008). The concerns about longevity risk have led to a surge in interest among pension funds and insurance companies in accurately modeling and forecasting agespecific mortality rates (or death counts or survival probabilities). Any improvement in the forecast accuracy of mortality rates will be beneficial for determining the allocation of current and future resources at the national and subnational levels (see, e.g., Koissia (2006); Denuit et al. (2007); Hanewald et al. (2011)).
Many different models for forecasting agespecific mortality rates have been proposed in the literature (see Booth (2006); Booth and Tickle (2008); Currie et al. (2004); Girosi and King (2008); Shang et al. (2011); Tickle and Booth (2014), for reviews). Of these, a significant milestone in demographic modeling and forecasting was the work by Lee and Carter (1992). They implemented a principal component method to model agespecific mortality and extracted a single timevarying index representing the trend in the level of mortality rates, from which the forecasts are obtained by a random walk with drift. While the LeeCarter method is simple and robust in situations where agespecific log mortality rates have linear trends (Booth 2006), it has the limitation of attempting to capture the patterns of mortality rates using only one principal component and its associated scores. To rectify this deficiency, the LeeCarter method has been extended and modified. For example, from a discrete data matrix perspective, Booth et al. (2002), Renshaw and Haberman (2003a), and Cairns et al. (2006); Cairns et al. (2009) proposed the use of more than one component in the LeeCarter method to model agespecific mortality rates; Renshaw and Haberman (2006) proposed an ageperiodcohort extension to the LeeCarter model under a Poisson error structure, while Plat (2009b) extended this model by incorporating the dependence between ages. Cairns et al. (2006) used a logistic transformation to model the relationship between the death probability and age observed over time, while Cairns et al. (2009) extended this model by incorporating the cohort effect. Girosi and King (2008) and Wiśniowski et al. (2015) considered a Bayesian paradigm for the LeeCarter model estimation and forecasting. Hatzopoulos and Haberman (2009) followed a generalized linear model approach which leads to models that have a similar structure to the LeeCarter model but with a generalized error structure. Hunt and Blake (2014) presented a general structural form of mortality models. From a continuous function perspective, Hyndman and Ullah (2007) proposed a functional data model that utilizes nonparametric smoothing and higherorder principal components, while Shang et al. (2011) and Shang (2016) considered a multilevel functional data model to model mortality rates jointly for multiple populations.
There exist many papers on comparing the forecast accuracy among several mortality forecasting methods. However, the most accurate forecasting method has been determined based on an aggregate loss function. Instead of identifying the most accurate method, Shang (2012) considered a model averaging approach to combine forecasts from a range of methods, such as the LeeCarter and functional timeseries methods. In this paper, we propose a new model averaging approach motivated by Samuels and Sekkel (2017). The proposed model averaging method uses the model confidence set procedure to select a set of superior models and combines the forecasts by assigning equal weights to the set of superior models. In contrast with Shang (2012), the problem is centered not on the selection of optimal weights, but on the selection of superior models. In the “A competing model averaging method” section, we compare the forecast accuracy between the existing and proposed model average methods.
With the aim of evaluating and comparing the forecast accuracy of different forecasting methods, forecast competition has a long history. The “M” competition originated from Makridakis et al. (1979) was the first attempt at a large empirical comparison of forecasting methods. In that “M” competition, there were 1001 time series for which participants were invited to submit their forecasts. Later, the results were published in Makridakis et al. (1982). The “M” competition has progressed slowly over the years, with the most recent M4 competition taken place in 2018. The topperforming teams all combine forecasts from a range of statistical and machine learning methods via some model averaging to improve the point and interval forecast accuracies.
Despite the popularity of model averaging in statistical and forecasting literature (see, e.g., Bates and Granger (1969); Dickinson (1975); Clemen (1989)), model averaging has not received increasing attention in the demographic literature with the noticeable exceptions of Shang (2012); Shang (2015) in the context of mortality forecasting, ((Bijak 2011), Chapter 5) in the context of migration forecasting, and Abel et al. (2013) and Shang et al. (2014) for the overall population growth rate. Shang (2012) revisited many statistical methods and combined their forecasts based on two weighting schemes, one of which has been adapted for comparison in the “A competing model averaging method” section. Both weighting schemes determine the weights by using either insample forecast accuracy or insample goodnessoffit. Because of the finite sample, the weight assigned to the worst model is often small but not zero. In turn, this may lead to inferior forecast accuracy than the one based on oracle weights. This motivates us to consider an alternative model averaging idea. Instead of assigning weights to the forecasts from all models, we trim out the worse performing models based on a statistical significance test, such as the model confidence set procedure of Hansen et al. (2011) (see also Samuels and Sekkel (2017)).
While most attention has been paid to selecting model combination weights see, (e.g., Fischer and Harvey (1999); Genre et al. (2013)), Aiolfi et al. (2010) pointed out that there has been little research focusing on which models to include in the model pool. Graefe (2015) asserted that a simple average of the forecasts produced by individual models is a benchmark and commonly outperforms more complicated weighting schemes that rely on the estimation of theoretically optimal weights. The simple average of the forecasts performs well when the model fits the data poorly: when the sample number per predictor is low and when the predictors are highly correlated (Graefe 2015).
This paper focuses on a statistical significance approach to select the models to be included in the forecast combination. Via the model confidence set procedure of Hansen et al. (2011), we determine a set of statistically superior models, conditional on the model’s insample performance for forecasting agespecific mortality rates. By equally averaging the forecasts from the superior models, we evaluate and compare point and interval forecast accuracies, as measured by the root mean square forecast error and mean interval score, respectively.
The outline of this paper is described as follows: From the Japanese national and subnational agespecific mortality data in the “Japanese national and subnational agespecific mortality data” section, we first visualize the heterogeneity in agespecific mortality rates among 47 prefectures (Additional file 1). Then, we revisit some commonly used multivariate and functional timeseries extrapolation methods for forecasting agespecific mortality rates in the “Timeseries extrapolation models” section. Using the model confidence set procedure of (Hansen et al. 2011) described in the “Model confidence set” section, we select a set of superior models based on their point or interval forecast accuracy and demonstrate the robust accuracy of the proposed model averaging method in the “Forecast results” section. In the “A competing model averaging method” section, we present an adaption of an existing model averaging method where optimal weights are estimated based on insample forecast accuracy and assigned to the forecasts from all models. In the “Discussion” section, we conclude and outline how the methodology presented here can be further extended.
Japanese national and subnational agespecific mortality data
We study the Japanese agespecific mortality rates from 1975 to 2015, obtained from the Japanese Mortality Database (2017). We consider ages from 60 to 99 in single years of age, while the last age group contains all ages at and beyond 100 (abbreviated as 100+). We consider modeling mortality at older ages, as the mortality forecasts are an important input for calculating annuity prices for retirees and the corresponding reserves held by insurance companies and pension funds. Some of the models considered were designed for modeling mortality at older ages, such as the CairnsBlakeDowd models.
We split the Japanese mortality rates by sex and prefecture. We are also interested in the mortality data at the subnational (i.e., prefecture) level. The mortality forecasts at the prefecture level are more useful than the mortality forecasts at the national level for local policy making and planning.
In the supplement, we have plotted the geographic locations (from North to South) of the 47 prefectures within eight regions of Japan (Additional file 2). Also, we present the names of prefecture within each of the eight regions of Japan. Shang and Haberman (2017) and Shang and Hyndman (2017) present plots of the ratio of mortality between each prefecture and Japan by age or year.
Timeseries extrapolation models
We study some timeseries extrapolation methods for modeling and forecasting agespecific mortality rates. The models that we have considered are subjective and far from extensive, but they suffice to serve as a test bed for demonstrating the performance of forecast combination. From actuarial science, we consider a family of RenshawHaberman (RH) models (see, e.g., Renshaw and Haberman (2003a, b, 2006, 2008); Haberman and Renshaw (2008, 2009) and a family of CairnsBlakeDowd (CBD) models (Cairns et al. 2006). These two models perform well for mortality at higher ages, such as between 60 and 100+. From demography, we consider a family of LeeCarter (LC) models (see, e.g., Lee and Carter (1992); Booth et al. (2006); Zhao (2012); Zhao et al. (2013)). From statistics, we consider a family of functional timeseries models (see, e.g., Hyndman and Booth (2008); Hyndman and Shang (2009); Hyndman et al. (2013)). For implementation, we use the StMoMo package of Villegas et al. (2018) for the RH and CBD models; we use the demography package of Hyndman (2017) for the LC models for the Gaussian error setting and functional timeseries models.
Notations
Let the random variable D_{x,t} be the number of death counts in a population at age x and year t. A rectangular data array (d_{x,t},e_{x,t}) is available for data analysis where d_{x,t} is the observed number of deaths and e_{x,t} is the corresponding exposure to risk (Hatzopoulos and Haberman 2009). The force of mortality and central mortality rates are given by μ_{x,t} and m_{x,t}=d_{x,t}/e_{x,t}, respectively. Crossclassification is by individual calendar year t∈[t_{1},t_{n}] (range n) and by age x∈[x_{1},x_{k}], either grouped into k (ordered) categories or by individual year (range k), in which case yearofbirth or cohort year z=t−x∈[t_{1}−x_{k},t_{n}−x_{1}] (range n+k−1) is defined (see also Hatzopoulos and Haberman (2009)).
LeeCarter model under a Gaussian error setting
With the central mortality rates m_{x,t}, the LC model structure is
subject to the identification constraints
Note that α_{x} is the age pattern of the log mortality rates averaged across years; \(\beta _{x}^{(1)}\) is the first principal component capturing relative change in the log mortality rate at each age x; \(\kappa _{t}^{(1)}\) is the first set of principal component scores measuring general level of the log mortality rate at year t; bilinear terms \(\beta _{x}^{(1)}\kappa _{t}^{(1)}\) incorporating the agespecific period trends (Pitacco et al. (2009), Section 6.2); and ε_{x,t} is the model residual at age x and year t.
In the demographic forecasting literature, the LC model adjusts κ_{t} by refitting to the total number of deaths (see Lee and Carter (1992)). In Lee and Miller’s (2001) method, the adjustment of κ_{t} involves fitting life expectancy at birth in the year t. In Booth et al.’s (2006) method, the adjustment of κ_{t} involves fitting to the age distribution of deaths rather than to the total number of deaths.
The adjusted principal component scores {κ_{1},…,κ_{n}} are then extrapolated by a random walk with drift method, from which forecasts are obtained by (1) with the estimated mean function α_{x} and principal component β_{x}. That is,
where \(\widehat {\kappa }_{n+hn}\) denotes forecasts of principal component scores obtained from a univariate timeseries forecasting method, such as the random walk with drift.
Two sources of uncertainty ought be considered: estimation errors in the parameters of the LC model and forecast errors in the forecast principal component scores. Because of orthogonality between the first principal component and the error term in (1), the overall forecast variance can be approximated by the sum of the two variances (see Lee and Carter (1992)). Conditioning on the past data \(\boldsymbol {\mathcal {J}}=(\boldsymbol {m}_{1},\dots,\boldsymbol {m}_{n})\) and the first principal component b_{x}, we obtained the overall forecast variance of ln(m_{x,n+h}),
where \(b_{x}^{2}\) is the variance of the first principal component, calculated as the square of the β_{x} in (1); u_{n+hn}=var(κ_{n+h}κ_{1},…,κ_{n}) can be obtained from the univariate timeseries model; and the model residual variance v_{x} is estimated by averaging the residual squares \(\left \{\epsilon _{x,1}^{2},\dots,\epsilon _{x,n}^{2}\right \}\) for each x in (1).
RenshawHaberman model under a Poisson error setting
Renshaw and Haberman (2006) generalizes the LC model structure to include ageperiodcohort modeling by formulating the mortality reduction factor as
where α_{x} is an age function capturing the general shape of mortality by age; a time index \(\kappa _{t}^{(1)}\) specifies the mortality trend, and \(\beta ^{(1)}_{x}\) modulates its effect across ages; and γ_{t−x} denotes a random cohort effect as a function of the birthyear (t−x) (see also Villegas et al. (2018)). To estimate the parameters in (3), Renshaw and Haberman (2006) assume a Poisson distribution of deaths and use a loglink function targeting the force of mortality.
To facilitate the model identifiability, a set of parameter constraints are imposed by setting
Ageperiodcohort (APC) model
The APC model studied by Clayton and Schifflers (1987a, b) can be derived from the Renshaw and Haberman (2006) model. The APC model corresponds to \(\beta _{x}^{(1)} = \beta _{x}^{(0)} = 1\) in (3), that is
To ensure the model identifiability, a set of parameters are constrained by setting
CairnsBlakeDowd (CBD) model
While the LC model is a datadriven method, the CBD model attempts to find factors that may affect agespecific log mortality rates. The former approach is nonparametric, while the latter one is parametric. Note that in the original CBD model, the authors proposed the modeling of agespecific death probability q_{x,t}. Here, for the sake of comparison, we use the CBD model to model and forecast agespecific log mortality rates. Let the prespecific agemodulating parameters be \(\beta _{x}^{(1)} = 1\) and \(\beta _{x}^{(2)} = x\bar {x}\), the CBD model can be expressed as
where \(\bar {x}\) is the average age in the sample range. While \(\kappa _{t}^{(1)}\) can be viewed as a timevarying interpret, \(\kappa _{t}^{(2)}\) can be viewed as a timevarying slope. Cairns et al. (2006) produce mortality forecasts by projecting \(\kappa _{t}^{(1)}\) and \(\kappa _{t}^{(2)}\) jointly using a bivariate random walk with drift.
M7: quadratic CBD model with cohort effects
Cairns et al. (2009) extend the original CBD model in (4) by adding a cohort effect and a quadratic age effect to form
where \(\widehat {\sigma }_{x}^{2}\) is the average value of \((x\bar {x})^{2}\). To ensure the model identifiability, Cairns et al. (2009) impose a set of constraints:
In addition to M7 model, Cairns et al. (2009) also consider two simpler predictors given by
where x_{c} is a constant parameter to be estimated. Equations (5) and (6) are referred to as M6 and M8, respectively.
Plat model
By combining the features of the LC and CBD models, Plat (2009a) proposed the following model:
where \((\bar {x}x)^{+}=\max (\bar {x}x, 0)\). To ensure the model identifiability, the following set of parameter constraints have been imposed
In the families of the RH and CBD models, Cairns et al. (2006, 2011), Haberman and Renshaw (2011), and Villegas et al. (2018) assume that the period indexes follow a multivariate random walk with drift. For the cohort index, they assume it follows a univariate autoregressive integrated moving average model.
The functional timeseries models for one population
Functional principal component analysis
Hyndman and Ullah (2007) consider a functional timeseries model for forecasting agespecific mortality rate, where age is treated as a continuum. The functional timeseries model allows one to smooth the observed data points, in order to reduce or eliminate measurement error. To smooth data, Hyndman and Ullah (2007) suggest the use of a penalized regression spline with monotonic constraints applied to agespecific log mortality rates denoted by lnm_{t}(x) (see Hyndman and Ullah (2007), for detail). Here, we propose to smooth agespecific mortality rates from ages 0 to 100+ and then truncate the smoothed mortality rates from 60 to 100+.
With smoothed agespecific log mortality curves (lnm_{t}(x)), we obtain a mean function denoted by μ(x). With the decentered smoothed data, we apply a functional principal component analysis to reduce dimensionality to some functional principal components (i.e, ϕ_{k}(x)) and their associated scores (β_{k}=(β_{1,k},…,β_{n,k})). The functional principal component is constructed by sample variance of discretized functional data. Conditioning on the observed data and the estimated principal components, the point forecast of future mortality curves can be obtained by forecasting estimated principal component scores via a univariate timeseries method. The prediction interval can be constructed similarly to the way of constructing prediction intervals for the LeeCarter method in the “LeeCarter model under a Gaussian error setting” section.
Robust functional principal component analysis
Because the presence of outliers can seriously affect the performance of modeling and forecasting, it is important to eliminate the effect of outliers where possible. As considered in Hyndman and Ullah (2007), the robust functional timeseries method calculates the integrated squared error for each year, that is
The integrated squared error provides a measure of estimation accuracy for the functional principal component approximation of the functional data. Outliers are those years that have a larger integrated squared error than a critical value calculated from a χ^{2} distribution (see Hyndman and Ullah (2007), for details). By assigning zero weight to outliers, we can again apply the functional timeseries method to model and forecast agespecific mortality rates.
The functional timeseries models for multiple subpopulations
When forecasting agespecific mortality for multiple subpopulations, it is advantageous to use a model that can capture correlation among subpopulations as the covariance of the multiple subpopulations often exhibits crosscorrelation. By modeling the crosscorrelation, it may improve forecast accuracy. Here, we consider the problem of jointly modeling and forecasting female and male mortality in order to produce coherent forecasts. Note that coherent forecasts can also be achieved by jointly modeling female or male mortality for all 47 prefectures.
Productratio method of Hyndman et al. (2013)
In Hyndman et al. (2013), they define the square roots of the product and ratio functions of the smoothed mortality rates for female and male data:
Instead of modeling female and male mortality data, we model the product and ratio functions. The advantage of this approach is that the product and ratio functions tend to behave roughly independently of each other, provided that the multiple subpopulations have approximately equal variances. On the logarithmic scale, these are sums and differences that are nearly uncorrelated. The functional timeseries method in “The functional timeseries models for one population” section can be applied to forecast the product and ratio functions (see Hyndman et al. (2013), fore details).
Multivariate functional timeseries method
We consider data where each observation consists of w≥2 functions, [ lnm^{(1)}(x),…, lnm^{(w)}(x)]^{⊤}∈R^{w}. These multivariate functions are defined over the same domain \(\mathcal {I}\) (e.g., Jacques and Preda (2014); Chiou et al. (2014); Shang and Yang (2017)).
We follow Shang and Yang (2017) and consider the stacking of multiple subpopulations into a long vector of functions, i.e., we stack the discretized data points of each subpopulation together for the same year. Then, we perform a multivariate functional principal component analysis to reduce dimensionality and summarize the main mode of information. With the extracted principal components and their scores, a functional timeseries method in “The functional timeseries models for one population” section can be applied again.
Multilevel functional timeseries method
The multilevel functional data model has a strong resemblance to a twoway functional analysis of variance model studied by Morris et al. (2003), CuestaAlbertos and FebreroBande (2010), and Zhang (2014, Section 5.4). It is a special case of the general “functional mixed model” proposed in Morris and Carroll (2006). In the case of two subpopulations, the basic idea is to decompose agespecific log mortality curves among different subpopulations into a sexspecific average μ^{j}(x), a common trend across subpopulations R_{t}(x), a sexspecific residual trend \(U_{t}^{j}(x)\), and measurement error \(e_{t}^{j}(x)\) with finite variance (σ^{2})^{j} (see, e.g., Hatzopoulos and Haberman (2013); Shang (2016)). The common and sexspecific residual trends are modeled by projecting them onto the eigenvectors of covariance operators of the aggregate and populationspecific centered stochastic processes, respectively.
Model confidence set
The model confidence set procedure proposed by Hansen et al. (2011) consists of a sequence of tests permitting the construction of a set of “superior” models, where the null hypothesis of equal predictive ability (EPA) is not rejected at a specified confidence level. The EPA test statistic can be evaluated for any arbitrary loss function, such as the square or absolute loss function.
Let M be some subset of original models denoted by M^{0} and let m be the number of models in M, and let d_{ρξ,ℓ} denote the loss differential between two models ρ and ξ, that is
and calculate
as the loss of model ρ relative to any other model ξ at time point ℓ. Let c_{ρξ}=E(d_{ρξ}) and c_{ρ.}=E(d_{ρ.}) be finite and not time dependent. The EPA hypothesis for a set of M candidate models can be formulated in two ways:
or
Based on c_{ρξ} or c_{ρ.}, we construct two hypothesis tests as follows:
where \(\overline {d}_{\rho.} = \frac {1}{m}\sum _{\xi \in M}\overline {d}_{\rho \xi }\) is the sample loss of the ρth model compared to the averaged loss across models, and \(\overline {d}_{\rho \xi } =\frac {1}{m}\sum ^{m}_{\ell =1}d_{\rho \xi,\ell }\) measures the relative sample loss between the ρth and ξth models. Note that \(\widehat {\text {Var}}\left (\overline {d}_{\rho.}\right)\) and \(\widehat {\text {Var}}\left (\overline {d}_{\rho \xi }\right)\) are the bootstrapped estimates of \(\text {Var}\left (\overline {d}_{\rho.}\right)\) and \(\text {Var}\left (\overline {d}_{\rho \xi }\right)\), respectively. Bernardi and Catania (2014) perform a block bootstrap procedure with 5,000 bootstrap samples by default, where the block length is given by the maximum number of significant parameters obtained by fitting an autoregressive process on all the d_{ρξ} terms. For both hypotheses in (8) and (9), there exist two test statistics:
where t_{ρξ} and t_{ρ.} are defined in (10) and (11), respectively. While T_{R,M} uses the loss differential between models ρ and ξ, T_{max,M} uses the aggregated loss differential between models ρ and ξ over ξ. Oftentimes but not always, the models selected on the basis of T_{R,M} form a subset of the models selected on the basis of T_{max,M}.
The model confidence set (MCS) procedure is a sequential testing procedure, which eliminates the worst model at each step until the hypothesis of equal predictive ability is accepted for all the models belonging to a set of superior models. The selection of the worst model is determined by an elimination rule that is consistent with the test statistic,
Forecast results
Point forecast evaluation
An expanding window analysis of a timeseries model is commonly used to assess model and parameter stabilities over time. It assesses the constancy of a model’s parameter by computing parameter estimates and their corresponding forecasts over an expanding window of a fixed size through the sample size (see Zivot and Wang 2006, Chapter 9 for details). Using the first 21 observations from 1975 to 1995 in the Japanese agespecific mortality rates, we produce onestepahead point forecasts. Through an expanding window approach, we reestimate the parameters in the timeseries forecasting models using the first 22 observations from 1975 to 1996. Forecasts from the estimated models are then produced for onestepahead. We iterate this process by increasing the sample size by one year until reaching the end of the training data period in 2005. This process produces ten onestepahead forecasts in the validation data period from 1996 to 2005. We compare these forecasts with the holdout samples to determine the point and interval forecast accuracies. By using the MCS procedure, we identify a superior set of models for averaging. Through the expanding window approach, we evaluate the outofsample point and interval forecast accuracies for the testing data from 2006 to 2015.
To evaluate the point forecast accuracy, we use the root mean squared forecast error (RMSFE). The RMSFE measures how close the forecasts are in comparison to the actual values of the variable being forecast. For each series, they can be written as
where \(\mathcal {Y}_{n+\xi }(x_{j})\) represents the actual holdout sample for the jth age and ξth curve of the forecasting period, while \(\widehat {\mathcal {Y}}_{n+\xi }(x_{j})\) represents the point forecasts for the holdout sample.
Interval forecast evaluation
In addition to point forecasts, we also evaluate the pointwise interval forecast accuracy using the interval score of Gneiting and Raftery (2007) (see also Gneiting and Katzfuss (2014)). For each year in the forecasting period, the onestepahead prediction intervals were calculated at the 100(1−α)% nominal coverage probability. We consider the common case of symmetric 100(1−α)% prediction interval, with lower and upper bounds that are predictive quantiles at α/2 and 1−α/2, denotes by \(\widehat {\mathcal {Y}}_{n+\xi }^{l}(x_{i})\) and \(\widehat {\mathcal {Y}}_{n+\xi }^{u}(x_{i})\). As defined by Gneiting and Raftery (2007), a scoring rule for the pointwise interval forecast at time point x_{i} is
where α denotes the level of significance, customarily α=0.2, and \(\mathbbm {1}\{\cdot \}\) denotes a binary indicator. The optimal interval score is achieved when \(\mathcal {Y}_{n+\xi }(x_{i})\) lies between \(\widehat {\mathcal {Y}}_{n+\xi }^{l}(x_{i})\) and \(\widehat {\mathcal {Y}}_{n+\xi }^{u}(x_{i})\), with the distance between the upper bound and lower bound being minimal.
We define the mean interval score for different points in a curve and different lengths in the forecasting period as
where \(S_{\alpha, \xi }\left [\widehat {\mathcal {Y}}_{n+\xi }^{l}(x_{j}), \widehat {\mathcal {Y}}_{n+\xi }^{u}(x_{j}); \mathcal {Y}_{n+\xi }(x_{j})\right ]\) denotes the interval score at the ξth curve of the forecasting period.
Determining a superior set of models
Based on the RMSFE error measure in the training data, we examine statistical significance in point forecast accuracy among the 17 timeseries extrapolation methods. The 17 models considered are listed in Table 1.
With the 90% confidence level of the MCS tests, we identify the set of superior models regarding point forecast accuracy. In Tables 2 and 3, we determine the set of superior models among the 17 models considered using the T_{max,M} and T_{R,M} tests.
Based on the mean interval scores in the validation period, we examine statistical significance in interval forecast accuracy among the timeseries extrapolation methods. With the 90% confidence level of the MCS tests, we identify the set of superior models regarding interval forecast accuracy. In Tables 4 and 5, we determine the set of superior models using both the T_{max,M} and T_{R,M} tests among the 17 models considered.
Point and interval forecast comparison
Based on the selected set of superior models, we produce modelaveraged point and interval forecasts using equal weights. In Table 6, we compute the point and interval forecast accuracies for female and male data.
As measured by the mean RMSFE over 10 years in the forecasting period, the Plat model gives the most accurate point forecasts for the whole of Japan. For the average of 47 prefectures in Japan, the Plat model gives the most accurate point forecasts for females, while the multilevel functional timeseries method performs the best for males. Although the model averaging methods are not the best model, in this case, they rank among the topperforming methods. Between the two statistical significance tests, there is a marginal difference in the point forecast accuracy.
As measured by the mean interval scores over ten different horizons, the Plat model produces the most accurate interval forecasts for the whole of Japan and its prefectures for females. For males, the model averaging methods perform the best for the whole of Japan and rank as the second best after the productratio method for the average of 47 prefectures. Again, there is a marginal difference between the two statistical significance tests regarding interval forecast accuracy.
A competing model averaging method
An existing model averaging method combines forecasts from the top two methods (see Shang (2012)). Given the top two methods are arbitrary from sample to sample, we have decided to combine forecasts from all methods and assign weights differently. Among all methods, we determine point or interval forecast accuracy as measured by the corresponding forecast errors in the validation set and assign the weights to be the inverse of their forecast errors. We then standardize all weights so that the weights sum to 1. Conceptually, the method that performs better in the validation set receives a higher weight in the combined forecasts. The point and interval forecast accuracies of this model averaging method are presented in Table 7.
Compared to our proposal model averaging method, the existing model averaging method assigns different weights to different models. From the insample forecast errors, the existing model averaging method assigns higher weights for those more accurate models and lower weights for those less accurate models. By contrast, our proposed model averaging method selects a superior subset of models and assign equal weights. From the results, we find that the proposed model averaging method gives a smaller forecast error than the existing model averaging method.
Discussion
Using the national and subnational Japanese mortality rates, we evaluate and compare point and interval forecast accuracies, as measured by the root mean squared error and mean interval score, among the 17 timeseries extrapolation methods and two model averaging methods considered. From the viewpoint of the point forecast accuracy, the Plat model gives the smallest point forecast errors, followed by the LeeCarter and modelaveraged method for Japanese females and males. In this case, because the superior set of the models was determined from insample forecast errors, it is the case where the best model for the insample forecasts may not be the best model for outofsample forecasts. This affects the forecast accuracy of the modelaveraged method. Also, the LeeCarter method with the Poisson error structure is more accurate than the version with the Gaussian error structure. For modeling subnational Japanese females, the Plat model also performs the best, while the multilevel functional timeseries model performs the best for males.
From the viewpoint of the interval forecast accuracy, the Plat model gives the smallest interval forecast errors, followed by the modelaveraged methods for Japanese females. For Japanese males, the modelaveraged methods produce the smallest interval forecast errors for Japan and are on a par with the productratio method for Japanese subnational data. To our surprise, the RenshawHaberman methods produce relatively worse interval forecast accuracy than that produced by the LeeCarter and functional timeseries methods. The result could be due to the instability of parameter estimation.
From Tables 2, 3, 4, and 5, the best model for producing point forecasts does not necessary the same as the best model for producing interval forecasts, and the best model for producing the national series does not necessarily perform the best for the subnational series, as the features of the data may be different.
Because different models have their advantages and disadvantages, we apply a modelaveraged method to select a set of superior model based on the model confidence set. The model confidence set is a procedure that determines a set of superior models based on the insample forecast errors.
For producing point forecasts, the selected superior sets of models are more diverse. For producing interval forecasts, the selected superior set of models often includes the productratio method, especially for male series. Given the productratio method is a joint modeling and coherent forecasting method, it achieves better forecast accuracy for males with a small sacrifice in terms of the female results. By coherent, we believe within each prefecture, female and male subpopulations share the similar characteristics, such as health facilities. Generally, the joint modeling methods, which also include the multivariate and multilevel functional timeseries methods, often but not always outperform the model without incorporating correlation among subpopulations. The modelaveraged forecasts may not perform the best for females and males, but they tend to give an aggregate best performance.
We show that potential gains in forecast accuracy can be achieved by discarding the worse performing models before combining the forecasts equally. By contrast, an existing model averaging method assigning different weights for all models performs worse than the proposed method. We find that the proposed model averaging method offers a more robust procedure for selecting the forecasting models based on their insample performances. By robustness, the modelaveraged methods are protected against model misspecification. The advantage of the modelaveraged methods is more apparent for males than females.
The accurate forecasting of mortality at retirement ages is essential to determine life, fixedterm, and delayed annuity prices for various maturities and starting ages (see, e.g., Shang and Haberman (2017)). In the online supplement, we present a study on calculating singlepremium fixedterm immediate annuity. To forecast mortality rates, we suggest considering the notion of model averaging.
In our modeling and analysis, we have made several choices. Below, we set out the ways in which the different choices could potentially have affected our results/overall conclusions.

1)
We considered ages from 60 to 100+ to study the mortality pattern of retirees. We could apply the model averaging idea to other age groups, such as ages from 0 to 100+. With different age groups, the point and interval forecast results may be different.

2)
Other agespecific mortality forecasting models could be incorporated into the model averaging. We consider only timeseries extrapolation methods and did not consider the expectation or explanation methods. For longterm forecasts and the expectation, expectation method has been used by the The Institute and Faculty of Actuaries and the CMI (2018). The inclusion of the expectation or explanation methods may alter the selection of superior models.

3)
We modeled agespecific mortality rates, but the focus could also be on death counts or survival probabilities. For example, when we model agespecific death counts, there are other models, such as compositional data analysis, that could be included in the initial model pool.

4)
After selecting the superior set of models, one could assign different weights instead of equal weights considered. With different weights, our forecast results may be improved as considered in Shang (2012).

5)
In applying the model confidence set procedure of Hansen et al. (2011), we consider the 90% confidence level. Considering other levels of confidence is possible. In general, as the confidence level increases, the number of superior models decreases.

6)
We evaluated and compared onestepahead, fivestepahead, and tenstepahead point and interval forecast accuracies. Considering other forecast horizons is also possible. For the longer term, the extrapolation methods may not perform well.

7)
We evaluated point forecast accuracy by the root mean square error and interval forecast accuracy by the mean interval score and coverage probability deviance, respectively. Considering other forecast error criteria, such as mean absolute percentage error or mean absolute scaled error, is also possible. With different error measures, the point and interval forecast results may be different.
We believe the present work paves the way for the above possible future research directions, and the proposed model averaging method should be a welcome addition to the demographic modeling and forecasting.
References
Abel, G.J., Bijak, J., Forster, J.J., Raymer, J., Smith, P.W.F., Wong, J.S.T. (2013). Integrating uncertainty in time series population forecasts: an illustration using a simple projection model. Demographic Research, 29, 1187–1226.
Aiolfi, M., Capistrán, C., Timmermann, A.G. (2010). Forecast combinations, Working paper 201021, CREATES. https://papers.ssrn.com/sol3/papers.cfm?abstract_id=1609530. Accessed 20 June 2018.
Bates, J.M., & Granger, C.W.J. (1969). The combination of forecasts. Operational Research Quarterly, 20(4), 451–468.
Bernardi, M., & Catania, L. (2014). The model confidence set package for R. Technical report, Sapienza University of Rome. https://arxiv.org/abs/1410.8504. Accessed 20 June 2018.
Bijak, J. (2011). Forecasting international migration in Europe: a Bayesian view. New York: Springer.
Booth, H. (2006). Demographic forecasting 1980 to 2005 in review. International Journal of Forecasting, 22(3), 547–581.
Booth, H., Hyndman, R.J., Tickle, L., De Jong, P. (2006). LeeCarter mortality forecasting: a multicountry comparison of variants and extensions. Demographic Research, 15, 289–310.
Booth, H., Maindonald, J., Smith, L. (2002). Applying LeeCarter under conditions of variable mortality decline. Population Studies, 56(3), 325–336.
Booth, H., & Tickle, L. (2008). Mortality modelling and forecasting: a review of methods. Annals of Actuarial Science, 3(12), 3–43.
Cairns, A.J.G., Blake, D., Dowd, K. (2006). A twofactor model for stochastic mortality with parameter uncertainty: theory and calibration. The Journal of Risk and Insurance, 73(4), 687–718.
Cairns, A.J.G., Blake, D., Dowd, K., Coughlan, G.D., Epstein, D., KhalafAllah, M. (2011). Mortality density forecasts: an analysis of six stochastic mortality models. Insurance: Mathematics & Economics, 48(3), 355–367.
Cairns, A.J.G., Blake, D., Dowd, K., Coughlan, G.D., Epstein, D., Ong, A., Balevich, I. (2009). A quantitative comparison of stochastic mortality models using data from England and Wales and the United States. North American Actuarial Journal, 13(1), 1–35.
Chiou, J.M., Chen, Y.T., Yang, Y.F. (2014). Multivariate functional principal component analysis: a normalization approach. Statistica Sinica, 24(4), 1571–1596.
Clayton, D., & Schifflers, E. (1987a). Models for temporal variation in cancer rates. I: ageperiod and agecohort models. Statistics in Medicine, 6(4), 449–467.
Clayton, D., & Schifflers, E. (1987b). Models for temporal variation in cancer rates. II: ageperiodcohort models. Statistics in Medicine, 6(4), 469–481.
Clemen, R.T. (1989). Combining forecasts: a review and annotated bibliography. International Journal of Forecasting, 5(4), 559–583.
Crawford, T., de Haan, R., Runchey, C. (2008). Longevity risk quantification and management: a review of relevant literature. Technical report, Society of Actuaries. https://www.soa.org/researchreports/2009/researchlongriskquant/. Accessed 20 June 2018.
CuestaAlbertos, J.A., & FebreroBande, M. (2010). A simple multiway ANOVA for functional data. Test, 19(3), 537–557.
Currie, I.D., Durban, M., Eilers, P.H.C. (2004). Smoothing and forecasting mortality rates. Statistical Modelling, 4(4), 279–298.
Denuit, M., Devolder, P., Goderniaux, A.C. (2007). Securitization of longevity risk: pricing survivor bonds with Wang transform in the LeeCarter framework. The Journal of Risk and Insurance, 74(1), 87–113.
Dickinson, J.P (1975). Some statistical results in the combination of forecasts. Operational Research Quarterly, 24(2), 253–260.
Fischer, I., & Harvey, N. (1999). Combining forecasts: what information do judges need to outperform the simple average?. International Journal of Forecasting, 15(3), 227–246.
Genre, V., Kenny, G., Meyler, A., Timmermann, A. (2013). Combining expert forecasts: can anything beat the simple average?. International Journal of Forecasting, 29(1), 108–121.
Girosi, F., & King, G. (2008). Demographic forecasting. Princeton: Princeton University Press.
Gneiting, T., & Katzfuss, M. (2014). Probabilistic forecasting. Annual Review of Statistics and Its Application, 1, 125–151.
Gneiting, T., & Raftery, A.E. (2007). Strictly proper scoring rules, prediction and estimation. Journal of the American Statistical Association, 102(477), 359–378.
Graefe, A. (2015). Improving forecasts using equally weighted predictors. Journal of Business Research, 68(8), 1792–1799.
Haberman, S., & Renshaw, A. (2008). Mortality, longevity and experiments with the LeeCarter model. Lifetime Data Analysis, 14(3), 286–315.
Haberman, S., & Renshaw, A. (2009). On ageperiodcohort parametric mortality rate projections. Insurance: Mathematics & Economics, 45(2), 255–270.
Haberman, S., & Renshaw, A. (2011). A comparative study of parametric mortality projection models. Insurance: Mathematics & Economics, 48(1), 35–55.
Hanewald, K., Post, T., Gründl, H. (2011). Stochastic mortality, macroeconomic risks and life insurer solvency. The Geneva Papers on Risk and InsuranceIssues and Practice, 36(3), 458–475.
Hansen, P.R., Lunde, A., Nason, J.M. (2011). The model confidence set. Econometrica, 79(2), 453–497.
Hatzopoulos, P., & Haberman, S. (2009). A parameterized approach to modeling and forecasting mortality. Insurance: Mathematics and Economics, 44(1), 103–123.
Hatzopoulos, P., & Haberman, S. (2013). Common mortality modeling and coherent forecasts. An empirical analysis of worldwide mortality data. Insurance: Mathematics & Economics, 52(2), 320–337.
Hunt, A., & Blake, D. (2014). A general procedure for constructing mortality models. North American Actuarial Journal, 18(1), 116–138.
Hyndman, R., & Booth, H. (2008). Stochastic population forecasts using functional data models for mortality, fertility and migration. International Journal of Forecasting, 24(3), 323–342.
Hyndman, R.J. (2017). Demography: forecasting mortality, fertility, migration and population data. R package version 1.20. https://CRAN.Rproject.org/package=demography. Accessed 20 June 2018.
Hyndman, R.J., Booth, H., Yasmeen, F. (2013). Coherent mortality forecasting: the productratio method with functional time series models. Demography, 50(1), 261–283.
Hyndman, R.J., & Shang, H.L. (2009). Forecasting functional time series (with discussions). Journal of the Korean Statistical Society, 38(3), 199–211.
Hyndman, R., & Ullah, M. (2007). Robust forecasting of mortality and fertility rates: a functional data approach. Computational Statistics & Data Analysis, 51(10), 4942–4956.
Jacques, J., & Preda, C. (2014). Modelbased clustering for multivariate functional data. Computational Statistics & Data Analysis, 71, 92–106.
Japanese Mortality Database. (2017). National Institute of Population and Social Security Research. Available at http://www.ipss.go.jp/ptoukei/JMD/indexen.html. Accessed 20 June 2018.
Koissia, M.C. (2006). Longevity and adjustment in pension annuities, with application to Finland. Scandinavian Actuarial Journal, 2006(4), 226–242.
Lee, R.D., & Carter, L.R. (1992). Modeling and forecasting U.S. mortality. Journal of the American Statistical Association, 87(419), 659–671.
Lee, R.D., & Miller, T. (2001). Evaluating the performance of the LeeCarter method for forecasting mortality. Demography, 38(4), 537–549.
Makridakis, S., Andersen, A., Carbone, R., Fildes, R., Hibon, M., Lewandowski, R., Newton, J., Parzen, E., Winkler, R. (1982). The accuracy of extrapolation (time series) methods: results of a forecasting competition. Journal of Forecasting, 1(2), 111–153.
Makridakis, S., Hibon, M., Moser, C. (1979). Accuracy of forecasting: an empirical investigation. Journal of the Royal Statistical Society: Series A, 142(2), 97–145.
Morris, J.S., & Carroll, R.J. (2006). Waveletbased functional mixed models. Journal of the Royal Statistical Society: Series B, 68(2), 179–199.
Morris, J.S., Vannucci, M., Brown, P.J., Carroll, R.J. (2003). Waveletbased nonparametric modeling of hierarchical functions in colon carcinogenesis. Journal of the American Statistical Association, 98(463), 573–583.
Pitacco, E., Denuit, M., Haberman, S., Olivieri, A. (2009). Modelling longevity dynamics for pensions and annuity business. Oxford: Oxford University Press.
Plat, R. (2009a). On stochastic mortality modeling. Insurance: Mathematics & Economics, 45(3), 393–404.
Plat, R. (2009b). Stochastic portfolio specific mortality and the quantification of mortality basis risk. Insurance: Mathematics and Economics, 45(1), 123–132.
Renshaw, A.E., & Haberman, S. (2003a). LeeCarter mortality forecasting with agespecific enhancement. Insurance: Mathematics & Economics, 33(2), 255–272.
Renshaw, A.E., & Haberman, S. (2006). A cohortbased extension to the LeeCarter model for mortality reduction factors. Insurance: Mathematics and Economics, 38(3), 556–570.
Renshaw, A.E., & Haberman, S. (2008). On simulationbased approaches to risk measurement in mortality with specific reference to Poisson LeeCarter modelling. Insurance: Mathematics & Economics, 42(2), 797–816.
Renshaw, A., & Haberman, S. (2003b). LeeCarter mortality forecasting: a parallel generalized linear modelling approach for England and Wales mortality projections. Journal of the Royal Statistical Society: Series C (Applied Statistics), 52(1), 119–137.
Samuels, J.D., & Sekkel, R.M. (2017). Model confidence sets and forecast combination. International Journal of Forecasting, 33(1), 48–60.
Shang, H.L. (2012). Point and interval forecasts of agespecific life expectancies: a model averaging approach. Demographic Research, 27, 593–644.
Shang, H.L. (2015). Statistically tested ccomparison of the accuracy of forecasting methods for agespecific and sexspecific mortality and life expectancy. Population Studies, 69(3), 317–335.
Shang, H.L. (2016). Mortality and life expectancy forecasting for a group of populations in developed countries: a multilevel functional data method. The Annals of Applied Statistics, 10(3), 1639–1672.
Shang, H.L., Booth, H., Hyndman, R.J. (2011). Point and interval forecasts of mortality rates and life expectancy: a comparison of ten principal component methods. Demographic Research, 25, 173–214.
Shang, H.L., & Haberman, S. (2017). Grouped multivariate and functional time series forecasting: an application to annuity pricing. Insurance: Mathematics & Economics, 75, 166–179.
Shang, H.L, & Hyndman, R.J (2017). Grouped functional time series forecasting: an application to agespecific mortality rates. Journal of Computational and Graphical Statistics, 26(2), 330–343.
Shang, H.L., Wiśniowski, A., Bijak, J., Smith, P.W.F., Raymer, J. (2014). Bayesian functional models for population forecasting. In: Marsili, M., & Capacci, G. (Eds.) In Proceedings of the Sixth Eurostat/UNECE Work Session on Demographic Projections. Istituto Nazionale di Statistica, Rome.
Shang, H.L., & Yang, Y. (2017). Grouped multivariate functional time series method: an application to mortality forecasting. In: Aneiros, G., Bongiorno, E.G., Cao, R., Vieu, P. (Eds.) In Functional Statistics and Related Fields. Springer, Cham, (pp. 233–241).
The Institute and Faculty of Actuaries and the CMI. (2018). Continuous Mortality Investigation. https://www.actuaries.org.uk/learnanddevelop/continuousmortalityinvestigation. Accessed 15 Sept 2018.
Tickle, L., & Booth, H. (2014). The longevity prospects of Australian seniors: an evaluation of forecast method and outcome. AsiaPacific Journal of Risk and Insurance, 8(2), 259–292.
Villegas, A.M., Kaishev, V.K., Millossovich, P. (2018). StMoMo: an R package for stochastic mortality modeling. Journal of Statistical Software, 84(3), 1–38.
Wiśniowski, A., Smith, P.W.F., Bijak, J., Raymer, J., Forster, J. (2015). Bayesian population forecasting: extending the LeeCarter method. Demography, 52(3), 1035–1059.
Zhang, J.T. (2014). Analysis of variance for functional data. Boca Raton: Chapman & Hall.
Zhao, B.B. (2012). A modified LeeCarter model for analysing shortbaseperiod data. Population Studies, 66(1), 39–52.
Zhao, B.B., Liang, X., Zhao, W., Hou, D. (2013). Modeling of groupspecific mortality in China using a modified LeeCarter model. Scandinavian Actuarial Journal, 2013(5), 383–402.
Zivot, E., & Wang, J. (2006). Modeling financial time series with SPLUS. New York: Springer.
Acknowledgements
Thanks to the two reviewers and the Guest Editor for the insightful comments and suggestions. We thank the comments and suggestions received from the S3RI seminar participants, especially Professors Peter W. F. Smith and Zudi Lu, at the Southampton Statistical Sciences Research Institute at the University of Southampton. We also thank Dr. Andres Villegas for answering our queries about the StMoMo package. The first author thanks the hospitalities of the Department of Statistics, Colorado State University; Department of Statistical Science, Cornell University; and Cass Business School, City University of London, when preparing this manuscript.
Funding
Not applicable
Availability of data and materials
The data used in this paper were obtained from Human Mortality Database cited in the paper.
Author information
Affiliations
Contributions
HLS proposed the statistical methodology and implemented computation. SH polished the manuscript and added detailed exposition especially in the RenshawHaberman models. Both authors read and approved the final version of the manuscript.
Corresponding author
Correspondence to Han Lin Shang.
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Additional files
Additional file 1
Code for Shiny application. The R code to produce a Shiny user interface for plotting every series in the Japanese human mortality data. (ZIP 62 kb)
Additional file 2
Detailed point and interval forecast results. While Table 6 presents a summary of the point and interval forecast accuracies, we present the detailed forecast results for ten years in the forecasting period. Calculation for singlepremium fixedterm immediate annuity. The forecasted mortality rate is an essential input for determining temporary annuity prices for various maturities and starting ages of the annuitant. (PDF 149 kb)
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License(http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Shang, H., Haberman, S. Model confidence sets and forecast combination: an application to agespecific mortality. Genus 74, 19 (2018). https://doi.org/10.1186/s4111801800439
Received:
Accepted:
Published:
Keywords
 Equal predictability test
 Japanese human mortality database
 Mean interval score
 Model averaging
 Root mean square forecast error