首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 21 毫秒
1.
In this paper we address the problem of projecting mortality when data are severely affected by random fluctuations, due in particular to a small sample size, or when data are scanty. Such situations may emerge when dealing with small populations, such as small countries (possibly previously part of a larger country), a specific geographic area of a (large) country, a life annuity portfolio or a pension fund, or when the investigation is restricted to the oldest ages. The critical issues arising from the volatility of data due to the small sample size (especially at the highest ages) may be made worse by missing records; this is the case, for example, of a small country previously part of a larger country, or a specific geographic area of a country, given that in some periods mortality data could have been collected just at an aggregate level.We suggest to ‘replicate’ the mortality of the small population by mixing appropriately the mortality data obtained from other populations. We design a two-step procedure. First, we obtain the average mortality of ‘neighboring’ populations. Three alternative approaches are tested for the assessment of the average mortality; conversely, the identification and the weight of the neighboring populations are obtained through (standard) optimization techniques. Then, following a sort of credibility approach, we mix the original mortality data of the small population with the average mortality of the neighboring populations.In principle, the approach described in the paper could be adopted for any population, whatever is its size, aiming at improving mortality projections through information collected from other groups. Through backtesting, we show that the procedure we suggest is convenient for small populations, but not necessarily for large populations, nor for populations not showing noticeable erratic effects in data. This finding can be explained as follows: while the replication of the original data implies the increase of the size of the sample, it also involves a smoothing of data, with a possible loss of specific information relating to the group referred to. In the case of small populations showing major erratic movements in mortality data, the advantages gained from the larger sample size overcome the disadvantages of the smoothing effect.  相似文献   

2.
Forecasting mortality rates is a problem which involves the analysis of high-dimensional time series. Most of usual mortality models propose to decompose the mortality rates into several latent factors to reduce this complexity. These approaches, in particular those using cohort factors, have a good fit, but they are less reliable for forecasting purposes. One of the major challenges is to determine the spatial–temporal dependence structure between mortality rates given a relatively moderate sample size. This paper proposes a large vector autoregressive (VAR) model fitted on the differences in the log-mortality rates, ensuring the existence of long-run relationships between mortality rate improvements. Our contribution is threefold. First, sparsity, when fitting the model, is ensured by using high-dimensional variable selection techniques without imposing arbitrary constraints on the dependence structure. The main interest is that the structure of the model is directly driven by the data, in contrast to the main factor-based mortality forecasting models. Hence, this approach is more versatile and would provide good forecasting performance for any considered population. Additionally, our estimation allows a one-step procedure, as we do not need to estimate hyper-parameters. The variance–covariance matrix of residuals is then estimated through a parametric form. Secondly, our approach can be used to detect nonintuitive age dependence in the data, beyond the cohort and the period effects which are implicitly captured by our model. Third, our approach can be extended to model the several populations in long run perspectives, without raising issue in the estimation process. Finally, in an out-of-sample forecasting study for mortality rates, we obtain rather good performances and more relevant forecasts compared to classical mortality models using the French, US and UK data. We also show that our results enlighten the so-called cohort and period effects for these populations.  相似文献   

3.
There is a burgeoning literature on mortality models for joint lives. In this paper, we propose a new model in which we use time-changed Brownian motion with dependent subordinators to describe the mortality of joint lives. We then employ this model to estimate the mortality rate of joint lives in a well-known Canadian insurance data set. Specifically, we first depict an individual’s death time as the stopping time when the value of the hazard rate process first reaches or exceeds an exponential random variable, and then introduce the dependence through dependent subordinators. Compared with existing mortality models, this model better interprets the correlation of death between joint lives, and allows more flexibility in the evolution of the hazard rate process. Empirical results show that this model yields highly accurate estimations of mortality compared to the baseline non-parametric (Dabrowska) estimation.  相似文献   

4.
The modeling of longitudinal and survival data is an active research area. Most of researches focus on improving the estimating efficiency but ignore many data features frequently encountered in practice. In this article, we develop a joint model that concurrently accounting for longitudinal-survival data with multiple features. Specifically, our joint model handles skewness, limit of detection, missingness and measurement errors in covariates which are typical observed in the collection of longitudinal-survival data from many studies. We employ a Bayesian approach for making inference on the joint model. The proposed model and method are applied to an AIDS study. A few alternative models under different conditions are compared. Some interesting results are reported. Simulation studies are conducted to assess the performance of the proposed methods.  相似文献   

5.
In this article, a conditional likelihood approach is developed for dealing with ordinal data with missing covariates in proportional odds model. Based on the validation data set, we propose the Breslow and Cain (Biometrika 75:11–20, 1988) type estimators using different estimates of the selection probabilities, which may be treated as nuisance parameters. Under the assumption that the observed covariates and surrogate variables are categorical, we present large sample theory for the proposed estimators and show that they are more efficient than the estimator using the true selection probabilities. Simulation results support the theoretical analysis. We also illustrate the approaches using data from a survey of cable TV satisfaction.  相似文献   

6.
We introduce a mixed regression model for mortality data which can be decomposed into a deterministic trend component explained by the covariates age and calendar year, a multivariate Gaussian time series part not explained by the covariates, and binomial risk. Data can be analyzed by means of a simple logistic regression model when the multivariate Gaussian time series component is absent and there is no overdispersion. In this paper we rather allow for overdispersion and the mixed regression model is fitted to mortality data from the United States and Sweden, with the aim to provide prediction and intervals for future mortality and annuity premium, as well as smoothing historical data, using the best linear unbiased predictor. We find that the form of the Gaussian time series has a large impact on the width of the prediction intervals, and it poses some new questions on proper model selection.  相似文献   

7.
Modeling mortality co-movements for multiple populations have significant implications for mortality/longevity risk management. A few two-population mortality models have been proposed to date. They are typically based on the assumption that the forecasted mortality experiences of two or more related populations converge in the long run. This assumption might be justified by the long-term mortality co-integration and thus be applicable to longevity risk modeling. However, it seems too strong to model the short-term mortality dependence. In this paper, we propose a two-stage procedure based on the time series analysis and a factor copula approach to model mortality dependence for multiple populations. In the first stage, we filter the mortality dynamics of each population using an ARMA–GARCH process with heavy-tailed innovations. In the second stage, we model the residual risk using a one-factor copula model that is widely applicable to high dimension data and very flexible in terms of model specification. We then illustrate how to use our mortality model and the maximum entropy approach for mortality risk pricing and hedging. Our model generates par spreads that are very close to the actual spreads of the Vita III mortality bond. We also propose a longevity trend bond and demonstrate how to use this bond to hedge residual longevity risk of an insurer with both annuity and life books of business.  相似文献   

8.
In this article we study a semiparametric generalized partially linear model when the covariates are missing at random. We propose combining local linear regression with the local quasilikelihood technique and weighted estimating equation to estimate the parameters and nonparameters when the missing probability is known or unknown. We establish normality of the estimators of the parameter and asymptotic expansion for the estimators of the nonparametric part. We apply the proposed models and methods to a study of the relation between virologic and immunologic responses in AIDS clinical trials, in which virologic response is classified into binary variables. We also give simulation results to illustrate our approach.  相似文献   

9.
Life expectancy has been increasing sharply around the globe since the second half of the 20th century. Mortality modeling and forecasting have therefore attracted increasing attention from various areas, such as the public pension systems, commercial insurance sectors, as well as actuarial, demographic and epidemiological research. Compared to the aggregate mortality experience, cause-specific mortality rates contain more detailed information, and can help us better understand the ongoing mortality improvements. However, when conducting cause-of-death mortality modeling, it is important to ensure coherence in the forecasts. That is, the forecasts of cause-specific mortality rates should add up to the forecasts of the aggregate mortality rates. In this paper, we propose a novel forecast reconciliation approach to achieve this goal. We use the age-specific mortality experience in the U.S. during 1970–2015 as a case study. Seven major causes of death are considered in this paper. By incorporating both the disaggregate cause-specific data and the aggregate total-level data, we achieve better forecasting results at both levels and coherence across forecasts. Moreover, we perform a cluster analysis on the cause-specific mortality data. It is shown that combining mortality experience from causes with similar mortality patterns can provide additional useful information, and thus further improve forecast accuracy. Finally, based on the proposed reconciliation approach, we conduct a scenario-based analysis to project future mortality rates under the assumption of certain causes being eliminated.  相似文献   

10.
The use of a finite mixture of normal distributions in model-based clustering allows us to capture non-Gaussian data clusters. However, identifying the clusters from the normal components is challenging and in general either achieved by imposing constraints on the model or by using post-processing procedures. Within the Bayesian framework, we propose a different approach based on sparse finite mixtures to achieve identifiability. We specify a hierarchical prior, where the hyperparameters are carefully selected such that they are reflective of the cluster structure aimed at. In addition, this prior allows us to estimate the model using standard MCMC sampling methods. In combination with a post-processing approach which resolves the label switching issue and results in an identified model, our approach allows us to simultaneously (1) determine the number of clusters, (2) flexibly approximate the cluster distributions in a semiparametric way using finite mixtures of normals and (3) identify cluster-specific parameters and classify observations. The proposed approach is illustrated in two simulation studies and on benchmark datasets. Supplementary materials for this article are available online.  相似文献   

11.
??The multivariate response is commonly seen in longitudinal and cross-sectional design. The marginal model is an important tool in discovering the average influence of the covariates on the response. A main feature of the marginal model is that even without specifying the inter-correlation among different components of the response, we still get consistent estimation of the regression parameters. This paper discusses the GMM estimation of marginal model when the covariates are missing at random. Using the inverse probability weighting and different basic working correlation matrices, we obtain a series of estimating equations. We estimate the parameters of interest by minimizing the corresponding quadratic inference function. Asymptotic normality of the proposed estimator is established. Simulation studies are conducted to investigate the finite sample performance of the new estimator. We also apply our proposal to a real data of mathematical achievement from middle school students.  相似文献   

12.
The Birnbaum‐Saunders (BS) distribution is a model that frequently appears in the statistical literature and has proved to be very versatile and efficient across a wide range of applications. However, despite the growing interest in the study of the BS distribution, quantile regression modeling has not been considered for this distribution. To fill this gap, we introduce a class of quantile regression models based on the BS distribution, which allows us to describe positive and asymmetric data when a quantile must be predicted using covariates. We use an approach based on a quantile parameterization to generate the model, permitting us to consider a similar framework to generalized linear models, providing wide flexibility. The methodology proposed includes a thorough study of theoretical properties and practical issues, such as maximum likelihood parameter estimation and diagnostic analytics based on local influence and residuals. The performance of the residuals is evaluated by simulations, whereas an illustrative example of income data is conducted using the methodology to show its potential for applications. The numerical results report an adequate performance of the approach to quantile regression, indicating that the BS distribution is a good modeling choice when dealing with data that have both positive support and asymmetry. The economic implications of our investigation are discussed in the final section. Hence, it can be a valuable addition to the tool kit of applied statisticians and econometricians.  相似文献   

13.
In many applications, some covariates could be missing for various reasons. Regression quantiles could be either biased or under-powered when ignoring the missing data. Multiple imputation and EM-based augment approach have been proposed to fully utilize the data with missing covariates for quantile regression. Both methods however are computationally expensive. We propose a fast imputation algorithm (FI) to handle the missing covariates in quantile regression, which is an extension of the fractional imputation in likelihood based regressions. FI and modified imputation algorithms (FIIPW and MIIPW) are compared to existing MI and IPW approaches in the simulation studies, and applied to part of of the National Collaborative Perinatal Project study.  相似文献   

14.
In most methods for modeling mortality rates, the idiosyncratic shocks are assumed to be homoskedastic. This study investigates the conditional heteroskedasticity of mortality in terms of statistical time series. We start from testing the conditional heteroskedasticity of the period effect in the naïve Lee-Carter model for some mortality data. Then we introduce the Generalized Dynamic Factor method and the multivariate BEKK GARCH model to describe mortality dynamics and the conditional heteroskedasticity of mortality. After specifying the number of static factors and dynamic factors by several variants of information criterion, we compare our model with other two models, namely, the Lee-Carter model and the state space model. Based on several error-based measures of performance, our results indicate that if the number of static factors and dynamic factors is properly determined, the method proposed dominates other methods. Finally, we use our method combined with Kalman filter to forecast the mortality rates of Iceland and period life expectancies of Denmark, Finland, Italy and Netherlands.  相似文献   

15.
In this paper, we study the weighted composite quantile regression (WCQR) for general linear model with missing covariates. We propose the WCQR estimation and bootstrap test procedures for unknown parameters. Simulation studies and a real data analysis are conducted to examine the finite performance of our proposed methods.  相似文献   

16.
With the rapid growth of data sets nowadays, the object sets in an information system may evolve in time when new information arrives. In order to deal with the missing data and incomplete information in real decision problems, this paper presents a matrix based incremental approach in dynamic incomplete information systems. Three matrices (support matrix, accuracy matrix and coverage matrix) under four different extended relations (tolerance relation, similarity relation, limited tolerance relation and characteristic relation), are introduced to incomplete information systems for inducing knowledge dynamically. An illustration shows the procedure of the proposed method for knowledge updating. Extensive experimental evaluations on nine UCI datasets and a big dataset with millions of records validate the feasibility of our proposed approach.  相似文献   

17.
This paper introduces mortality dependence in multi-country mortality modeling using a dynamic copula approach. Specifically, we use time-varying copula models to capture the mortality dependence structure across countries, examining both symmetric and asymmetric dependence structures. In addition, to capture the phenomenon of a heavy tail for the multi-country mortality index, we consider not only the setting of Gaussian innovations but also non-Gaussian innovations under the Lee–Carter framework model. As tests of the goodness of fit of different dynamic copula models, the pattern of mortality dependence, and the distribution of the innovations, we used empirical mortality data from Finland, France, the Netherlands, and Sweden. To understand the effect of mortality dependence on longevity derivatives, we also built a valuation framework for pricing a survivor index swap, then investigated the fair swap rates of a survivor swap numerically. We demonstrate that failing to consider the dynamic copula mortality model and non-Gaussian innovations would lead to serious underestimations of the swap rates and loss reserves.  相似文献   

18.
This article proposes a parsimonious alternative approach for modeling the stochastic dynamics of mortality rates. Instead of the commonly used factor-based decomposition framework, we consider modeling mortality improvements using a random field specification with a given causal structure. Such a class of models introduces dependencies among adjacent cohorts aiming at capturing, among others, the cohort effects and cross generations correlations. It also describes the conditional heteroskedasticity of mortality. The proposed model is a generalization of the now widely used AR-ARCH models for random processes. For such a class of models, we propose an estimation procedure for the parameters. Formally, we use the quasi-maximum likelihood estimator (QMLE) and show its statistical consistency and the asymptotic normality of the estimated parameters. The framework being general, we investigate and illustrate a simple variant, called the three-level memory model, in order to fully understand and assess the effectiveness of the approach for modeling mortality dynamics.  相似文献   

19.
Variational approximations have the potential to scale Bayesian computations to large datasets and highly parameterized models. Gaussian approximations are popular, but can be computationally burdensome when an unrestricted covariance matrix is employed and the dimension of the model parameter is high. To circumvent this problem, we consider a factor covariance structure as a parsimonious representation. General stochastic gradient ascent methods are described for efficient implementation, with gradient estimates obtained using the so-called “reparameterization trick.” The end result is a flexible and efficient approach to high-dimensional Gaussian variational approximation. We illustrate using robust P-spline regression and logistic regression models. For the latter, we consider eight real datasets, including datasets with many more covariates than observations, and another with mixed effects. In all cases, our variational method provides fast and accurate estimates. Supplementary material for this article is available online.  相似文献   

20.
We propose fast and scalable statistical methods for the analysis of hundreds or thousands of high-dimensional vectors observed at multiple visits. The proposed inferential methods do not require loading the entire dataset at once in the computer memory and instead use only sequential access to data. This allows deployment of our methodology on low-resource computers where computations can be done in minutes on extremely large datasets. Our methods are motivated by and applied to a study where hundreds of subjects were scanned using Magnetic Resonance Imaging (MRI) at two visits roughly five years apart. The original data possess over ten billion measurements. The approach can be applied to any type of study where data can be unfolded into a long vector including densely observed functions and images. Supplemental materials are provided with source code for simulations, some technical details and proofs, and additional imaging results of the brain study.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号