首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到17条相似文献,搜索用时 203 毫秒
1.
负二项回归模型的推广及其在分类费率厘定中的应用   总被引:1,自引:0,他引:1  
分类费率厘定中最常使用的模型之一是泊松回归模型,但当损失次数数据存在过离散特征时,通常会采用负二项回归模型。本文将两参数的负二项回归模型推广到了三参数情况,并用它来解决分类费率厘定中的过离散(over-dispersion)问题。本文通过对一组汽车保险损失数据的拟合表明,三参数的负二项分布回归模型可以有效改善对实际损失数据的拟合效果。  相似文献   

2.
零膨胀广义泊松回归模型与保险费率厘定   总被引:1,自引:0,他引:1  
在保险产品的分类费率厘定中,最常使用的模型之一是泊松回归模型.当损失数据存在零膨胀(zero-in flated)特征时,通常会采用零膨胀泊松回归模型.在零膨胀泊松回归模型中,一般假设结构零的比例参数φ为常数,不受费率因子的影响,这有可能背离实际情况.假设参数φ与费率因子之间存在一定关系,并在此基础上建立了零膨胀广义泊松回归模型,即Z IGP(τ)回归模型.通过对一组汽车保险损失数据的拟合表明,Z IGP(τ)回归模型可以有效地改善对实际数据的拟合效果,从而提高费率厘定结果的合理性.  相似文献   

3.
针对车险索赔次数数据经常出现的过度离散问题,采用数值模拟的方法,分别使用泊松模型(Poisson)、负二项回归模型(NB)以及广义泊松模型(GP)对不同程度的过度离散车险索赔次数数据进行拟合,并用均方误差、偏差以及AIC和BIC准则对Poisson、NB、GP三种模型的优良性进行比较分析,得到了不同条件下三种模型的优良性,并针对不同的条件给出了模型选择的建议.  相似文献   

4.
过离散次数分布模型的尾部特征   总被引:1,自引:0,他引:1  
在保险精算和生物统计等领域,离散型次数分布模型的应用十分广泛.当实际数据的尾部较长(即过离散),且零点的概率较大时,许多模型的拟合效果往往欠佳.本文通过计算概率之比的极限和偏度系数,对混合泊松分布和复合泊松分布的右尾特征和零点概率进行了比较,给出了它们的尾部排列顺序,以及尾部长短与零点概率的关系,从而为模型的构造或选择提供了一种指导.本文最后应用一组实际数据说明了在构造或选择次数分布模型时如何考虑尾部特征,从而改善对实际数据的拟合效果.  相似文献   

5.
来源于不同总体的数据异质性较大,数据“零取值”较多且离散度大,可利用零膨胀泊松(ZIP)混合回归模型建模分析,然而混合模型中自变量较多.为了筛选出重要变量,本文利用自适应LASSO对ZIP混合回归模型进行变量选择,即在似然函数中加入惩罚项,再利用EM算法估计参数.通过模拟,验证了该方法在变量选择和参数估计中的有效性.同时,将ZIP混合回归模型应用于预测借贷失败次数的实际数据分析,筛选出对借贷失败有重要影响的因素.最后,通过比较各模型的预测效果,得到ZIP混合回归模型优于泊松(Poisson),负二项(NB)和ZIP回归模型.  相似文献   

6.
本文研究泊松逆高斯回归模型的贝叶斯统计推断.基于应用Gibbs抽样,Metropolis-Hastings算法以及Multiple-Try Metropolis算法等MCMC统计方法计算模型未知参数和潜变量的联合贝叶斯估计,并引入两个拟合优度统计量来评价提出的泊松逆高斯回归模型的合理性.若干模拟研究与一个实证分析说明方法的可行性.  相似文献   

7.
在制造缺陷、专利申请、道路安全和公共卫生等应用领域,经常会出现较多的零观测值和一观测值.采用传统的泊松回归或负二项回归模型往往会过低地估计零观测值和一观测值出现的概率,数据拟合的效果欠佳.文章提出了0-1膨胀几何分布回归模型,巧妙地引入隐变量并进行极大似然估计和贝叶斯估计,基于数据扩充策略分别采用最大期望(EM)算法和Metropolis-Hastings抽样算法对回归参数向量进行估计.在不同的样本容量下进行数值模拟,并对两种估计方法的性能进行评价.研究表明,对于博士研究生发表论文数量的数据集,0-1膨胀几何分布回归模型能够达到更好的拟合效果.  相似文献   

8.
《数理统计与管理》2019,(2):235-246
零膨胀计数数据是当今数据分析的热点问题之一,该类数据的特点是零点过多,目前对这类数据的研究已经比较全面。另外还有些计数数据不仅会出现零点过多的现象,也会同时存在零、一点都过多的情形,如果再用零膨胀计数数据的统计方法去研究,产生的误差较大。目前国内外对零和一都膨胀的数据的研究还比较少,针对这种现象,本文引入零一膨胀泊松回归模型,并用局部多项式核回归法这种非参数统计分析方法对零一膨胀泊松回归模型进行参数估计,这是本文的创新点也是难点,并在求解参数的过程中引进了EM算法和Newton-Raphson迭代对参数近似求解。通过模拟结果可以得出此方法的可行性,最后通过对糖尿病患者数据的实例分析,可以验证此方法的有效性。  相似文献   

9.
本文利用齐次泊松过程的可加性,研究了复合泊松过程的可加性及其性质。作为应用,讨论了单个理赔额服从指数分布的复合泊松风险模型在第n次索赔时发生负盈余的概率。  相似文献   

10.
针对服从二项、泊松、几何、负二项、超几何、负超几何以及对数级数分布等离散型随机变量,给出了求其高阶原点矩的一个较为简单的递推计算方法.不仅非常容易地求出这些离散型随机变量的高阶原点矩,避免了计算阶乘矩或求导等复杂的运算,而且便于学生理解.论文还给出了这些离散型随机变量的3阶和4阶原点矩的表达式.  相似文献   

11.
Our paper presents an empirical analysis of the association between firm attributes in electronic retailing and the adoption of information initiatives in mobile retailing. In our attempt to analyze the collected data, we find that the count of information initiatives exhibits underdispersion. Also, zero‐truncation arises from our study design. To tackle the two issues, we test four zero‐truncated (ZT) count data models—binomial, Poisson, Conway–Maxwell–Poisson, and Consul's generalized Poisson. We observe that the ZT Poisson model has a much inferior fit when compared with the other three models. Interestingly, even though the ZT binomial distribution is the only model that explicitly takes into account the finite range of our count variable, it is still outperformed by the other two Poisson mixtures that turn out to be good approximations. Further, despite the rising popularity of the Conway–Maxwell–Poisson distribution in recent literature, the ZT Consul's generalized Poisson distribution shows the best fit among all candidate models and suggests support for one hypothesis. Because underdispersion is rarely addressed in IT and electronic commerce research, our study aims to encourage empirical researchers to adopt a flexible regression model in order to make a robust assessment on the impact of explanatory variables. Copyright © 2014 John Wiley & Sons, Ltd.  相似文献   

12.
Poisson回归模型广泛应用于分析计数型数据 ,Dean&Lawless(1989)和Dean(1992 )讨论了非重复测量得到的计数型数据的偏大离差存在性的检验问题 .本文分别利用随机系数模型和对数非线性模型讨论了基于重复测量得到的计数型数据的偏大离差的检验问题 ,得到了检验的score统计量 .  相似文献   

13.
The theory of tree-growing (RECPAM approach) is developed for outcome variables which are distributed as the canonical exponential family. The general RECPAM approach (consisting of three steps: recursive partition, pruning and amalgamation), is reviewed. This is seen as constructing a partition with maximal information content about a parameter to be predicted, followed by simplification by the elimination of ‘negligible’ information. The measure of information is defined for an exponential family outcome as a deviance difference, and appropriate modifications of pruning and amalgamation rules are discussed. It is further shown how the proposed approach makes it possible to develop tree-growing for situations usually treated by generalized linear models (GLIM). In particular, Poisson and logistic regression can be tree-structured. Moreover, censored survival data can be treated, as in GLIM, by observing a formal equivalence of the likelihood under random censoring and an appropriate Poisson model. Three examples are given of application to Poisson, binary and censored survival data.  相似文献   

14.
The conditional distribution of Y given X=x, where X and Y are non-negative integer-valued random variables, is characterized in terms of the regression function of X on Y and the marginal distribution of X which is assumed to be of a power series form. Characterizations are given for a binomial conditional distribution when X follows a Poisson, binomial or negative binomial, for a hypergeometric conditional distribution when X is binomial and for a negative hypergeometric conditional distribution when X follows a negative binomial.  相似文献   

15.
In this paper, we elaborate how Poisson regression models of different complexity can be used in order to model absolute transaction price changes of an exchange‐traded security. When combined with an adequate autoregressive conditional duration model, our modelling approach can be used to construct a complete modelling framework for a security's absolute returns at transaction level, and thus for a model‐based quantification of intraday volatility and risk. We apply our approach to absolute price changes of an option on the XETRA DAX index based on quote‐by‐quote data from the EUREX exchange and find that within our Bayesian framework a Poisson generalized linear model (GLM) with a latent AR(1) process in the mean is the best model for our data according to the deviance information criterion (DIC). While, according to our modelling results, the price development of the underlying, the intrinsic value of the option at the time of the trade, the number of new quotations between two price changes, the time between two price changes and the Bid–Ask spread have significant effects on the size of the price changes, this is not the case for the remaining time to maturity of the option. Copyright © 2006 John Wiley & Sons, Ltd.  相似文献   

16.
One of the major challenges associated with the measurement of customer lifetime value is selecting an appropriate model for predicting customer future transactions. Among such models, the Pareto/negative binomial distribution (Pareto/NBD) is the most prevalent in noncontractual relationships characterized by latent customer defections; ie, defections are not observed by the firm when they happen. However, this model and its applications have some shortcomings. Firstly, a methodological shortcoming is that the Pareto/NBD, like all lifetime transaction models based on statistical distributions, assumes that the number of transactions by a customer follows a Poisson distribution. However, many applications have an empirical distribution that does not fit a Poisson model. Secondly, a computational concern is that the implementation of Pareto/NBD model presents some estimation challenges specifically related to the numerous evaluation of the Gaussian hypergeometric function. Finally, the model provides 4 parameters as output, which is insufficient to link the individual purchasing behavior to socio‐demographic information and to predict the behavior of new customers. In this paper, we model a customer's lifetime transactions using the Conway‐Maxwell‐Poisson distribution, which is a generalization of the Poisson distribution, offering more flexibility and a better fit to real‐world discrete data. To estimate parameters, we propose a Markov chain Monte Carlo algorithm, which is easy to implement. Use of this Bayesian paradigm provides individual customer estimates, which help link purchase behavior to socio‐demographic characteristics and an opportunity to target individual customers.  相似文献   

17.
Abstract

This article proposes an algorithm for generating over-dispersed and under-dispersed binomial variates with specified mean and variance. The over-dispersed/under-dispersed distributions are derived from correlated binary variables with an underlying continuous multivariate distribution. Different multivariate distributions or different correlation matrices result in different over-dispersed (or under-dispersed) distributions. The over-dispersed binomial distributions that are generated from three different correlation matrices of a multivariate normal are compared with the beta-binomial distribution for various mean and over-dispersion parameters by quantile-quantile (Q-Q) plots. The two distributions appear to be similar. The under-dispersed binomial distribution is simulated to model an example data set that exhibits under-dispersed binomial variation.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号