首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
In modeling marked point processes, it is convenient to assume a separable or multiplicative form for the conditional intensity, as this assumption typically allows one to estimate each component of the model individually. Tests have been proposed in the simple marked point process case, to investigate whether the mark distribution is separable from the spatial–temporal characteristics of the point process. Here, we extend these tests to the case of a marked point process with covariates, and where one is interested in testing the separability of each of the covariates, as well as the mark and the coordinates of the point process. The extension is not at all trivial, and covariates must be treated in a fundamentally different way than marks and coordinates of the process, especially when the covariates are not uniformly distributed. An application is given to point process models for forecasting wildfire hazard in Los Angeles County, California, and solutions are proposed to the problem of how to proceed when the separability hypothesis is rejected.  相似文献   

2.
We consider the estimation of three-dimensional ROC surfaces for continuous tests given covariates.Three way ROC analysis is important in our motivating example where patients with Alzheimer’s disease are usually classified into three categories and should receive different category-specific medical treatment.There has been no discussion on how covariates affect the three way ROC analysis.We propose a regression framework induced from the relationship between test results and covariates.We consider several practical cases and the corresponding inference procedures.Simulations are conducted to validate our methodology.The application on the motivating example illustrates clearly the age and sex effects on the accuracy for Mini-Mental State Examination of Alzheimer’s disease.  相似文献   

3.
Finite mixture regression (FMR) models are frequently used in statistical modeling, often with many covariates with low significance. Variable selection techniques can be employed to identify the covariates with little influence on the response. The problem of variable selection in FMR models is studied here. Penalized likelihood-based approaches are sensitive to data contamination, and their efficiency may be significantly reduced when the model is slightly misspecified. We propose a new robust variable selection procedure for FMR models. The proposed method is based on minimum-distance techniques, which seem to have some automatic robustness to model misspecification. We show that the proposed estimator has the variable selection consistency and oracle property. The finite-sample breakdown point of the estimator is established to demonstrate its robustness. We examine small-sample and robustness properties of the estimator using a Monte Carlo study. We also analyze a real data set.  相似文献   

4.
Relative-risk models are often used to characterize the relationship between survival time and time-dependent covariates. When the covariates are observed, the estimation and asymptotic theory for parameters of interest are available; challenges remain when missingness occurs. A popular approach at hand is to jointly model survival data and longitudinal data. This seems efficient, in making use of more information, but the rigorous theoretical studies have long been ignored. For both additive risk models and relative-risk models, we consider the missing data nonignorable. Under general regularity conditions, we prove asymptotic normality for the nonparametric maximum likelihood estimators.  相似文献   

5.
6.
Generalized linear mixed models (GLMM) are used in situations where a number of characteristics (covariates) affect a nonnormal response variable and the responses are correlated due to the existence of clusters or groups. For example, the responses in biological applications may be correlated due to common genetic factors or environmental factors. The clustering or grouping is addressed by introducing cluster effects to the model; the associated parameters are often treated as random effects parameters. In many applications, the magnitude of the variance components corresponding to one or more of the sets of random effects parameters are of interest, especially the point null hypothesis that one or more of the variance components is zero. A Bayesian approach to test the hypothesis is to use Bayes factors comparing the models with and without the random effects in question—this work reviews a number of approaches for estimating the Bayes factor. We perform a comparative study of the different approaches to compute Bayes factors for GLMMs by applying them to two different datasets. The first example employs a probit regression model with a single variance component to data from a natural selection study on turtles. The second example uses a disease mapping model from epidemiology, a Poisson regression model with two variance components. Bridge sampling and a recent improvement known as warp bridge sampling, importance sampling, and Chib's marginal likelihood calculation are all found to be effective. The relative advantages of the different approaches are discussed.  相似文献   

7.
The covariate-specific receiver operating characteristic (ROC) curve is an important tool for evaluating the classification accuracy of a diagnostic test when it is associated with certain covariates. In this paper, a weighted Wilcoxon estimator is constructed for estimating this curve under the framework of location-scale model for the test result. The asymptotic normality is established, both for the regression parameter estimator and the estimator for the covariate-specific ROC curve at a fixed false positive point. Simulation results show that the Wilcoxon estimator compares favorably to its main competitors in terms of the standard error, especially when outliers exist in the covariates. As an illustration, the new procedure is applied to the dementia data from the national Alzheimer’s coordinating center.  相似文献   

8.
In analyses of bivariate ordered polytomous cataract data from atomic-bomb survivors, we compared two methods, the univariate worse-eye method, and the bivariate generalized estimating equations (GEE’s) method using global odds ratio by Williamson et al. (Journal of the American Statistical Association, 90, 1432–1437, 1995). When the association was large and only subject level covariates were used, model selection in the univariate and bivariate methods resulted in the same mean model and similar risk estimates. We showed that the mean parameter and the standard error (SE) in the univariate model are emphasized relative to those in the bivariate model, the biases of which are negligible when the association between both eyes is large. Large sample simulation studies indicated that the univariate Wald statistics are slightly conservative. The simulations also showed that, in bivariate cases, irrespective of the degree of association, the independence estimating equations method with robust SE, and the GEE method with model-based and robust SE are almost fully efficient in parameter estimation when only subject level covariates are included in the mean.  相似文献   

9.
风险差是流行病学中重要的指标之一,常用来比较两种治疗或两种诊断的有效性.因此,风险差区间的精确估计对流行病病情的诊断以及治疗方案的选择有很重要的意义.结合Poisson抽样的优点以及慢性病发病周期长和发病率低的特点,利用鞍点逼近方法来构造了Poisson分布下风险差的置信区间.同时,通过实例和Monte Carlo模拟对传统的四种区间构造方法进行评价.模拟结果表明:在小样本情况下,鞍点逼近方法得到的置信区间大多数能保证覆盖率近似于期望的置信水平并且使得区间长度最短,是一种很好的置信区间构造方法.  相似文献   

10.
This article concerns the statistical inference for the upper tail of the conditional distribution of a response variable Y given a covariate X = x based on n random vectors within the parametric extreme value framework. Pioneering work in this field was done by Smith (Stat Sci 4:367–393, 1989) and Smith and Shively (Atmos Environ 29:3489–3499, 1995). We propose to base the inference on a conditional distribution of the point process of exceedances given the point process of covariates. It is of importance that the conditional distribution merely depends on the conditional distribution of the response variable given the covariates. In the special case of Poisson processes such a result may be found in Reiss (1993). Our results are valid within the broader model where the response variables are conditionally independent given the covariates. It is numerically exemplified that the maximum likelihood principle leads to more accurate estimators within the conditional approach than in the previous one.  相似文献   

11.
In this paper, we analyze ovarian cancer cases from six hospitals in China, screen the prognostic factors and predict the survival rate. The data has the feature that all the covariates are categorical. We use three methods to estimate the survival rate–the traditional Cox regression, the two-step Cox regression and a method based on conditional inference tree. By comparison, we know that they are all effective and can predict the survival curve reasonably. The analysis results show that the survival rate is determined by a combination of risk factors, where clinical stage is the most important prognosis factor.  相似文献   

12.
Degradation data have been widely used to estimate product reliability. Because of technology advancement, time‐varying usage and environmental variables, which are called dynamic covariates, can be easily recorded nowadays, in addition to the traditional degradation measurements. The use of dynamic covariates is appealing because they have the potential to explain more variability in degradation paths. We propose a class of general path models to incorporate dynamic covariates for modeling of degradation paths. Physically motivated nonlinear functions are used to describe the degradation paths, and random effects are used to describe unit‐to‐unit variability. The covariate effects are modeled by shape‐restricted splines. The estimation of unknown model parameters is challenging because of the involvement of nonlinear relationships, random effects, and shaped‐restricted splines. We develop an efficient procedure for parameter estimations. The performance of the proposed method is evaluated by simulations. An outdoor coating weathering dataset is used to illustrate the proposed method. Copyright © 2015 John Wiley & Sons, Ltd.  相似文献   

13.
Discovering the preferences and the behaviour of consumers is a key challenge in marketing. Information about such topics can be gathered through surveys in which the respondents must assign a score to a number of items. A strategy based on different latent class models can be used to analyze such data and achieve this objective: it consists in identifying groups of consumers whose response patterns are similar and characterizing them in terms of preferences and covariates. The basic latent class model can be extended by including covariates to model differences in (1) latent class probabilities and (2) conditional probabilities. A strategy for fitting and choosing a suitable model among them is proposed taking into account identifiability issues, the identification of potential covariates and the checking of goodness-of-fit. The tools to perform this analysis are implemented in the R package covLCA available from CRAN. We illustrate and explain the application of this strategy using data about the preferences of Belgian households for supermarkets.  相似文献   

14.
Datasets involving repeated measurements over time are common in medical trials and epidemiological cohort studies. The outcomes and covariates are usually observed from randomly selected subjects, each at a set of possibly unequally spaced time design points. One useful approach for evaluating the effects of covariates is to consider linear models at a specific time, but the coefficients are smooth curves over time. We show that kernel estimators of the coefficients that are based on ordinary local least squares may be subject to large biases when the covariates are time-dependent. As a modification, we propose a two-step kernel method that first centers the covariates and then estimates the curves based on some local least squares criteria and the centered covariates. The practical superiority of the two-step kernel method over the ordinary least squares kernel method is shown through a fetal growth study and simulations. Theoretical properties of both the two-step and ordinary least squares kernel estimators are developed through their large sample mean squared risks.  相似文献   

15.
** Email: e.zwane{at}imperial.ac.uk Registrations in epidemiological studies suffer from incompleteness,thus a general consensus is to use capture–recapture models.Log-linear models are typically used when the registrationsmeasure the same population and the covariates are measuredby all registrations. This article shows how data can be analysedif some covariates are unobserved in some registrations andthe registrations do not all measure the whole population.  相似文献   

16.
We deal with the problem of finding a suitable model to predict survival of patients suffering from glial tumours as a function of several covariates. Estimation is based upon a retrospective study on 192 patients. Data were collected in the Hospital of Bordeaux and are analysed by Commenges and Dartigues1 using a Cox model. In the present paper we use dynamic Bayesian models which allow effects of the covariates to change with time through a stochastic structure. The survival function at one year is also calculated as a function of the covariates with the highest prognostic values and two factors (linear combinations of the covariates) are identified which synthesize information related to the general state of the patient (age, first symptom, etc.) and the characteristics of the tumour (diameter, localization, etc.), respectively. Survival at one year is then calculated as function of the two factors. Results are reported in tabular and graphic forms.  相似文献   

17.
We propose a two-component graphical chain model, the discrete regression distribution, where a set of discrete random variables is modeled as a response to a set of categorical and continuous covariates. The proposed model is useful for modeling a set of discrete variables measured at multiple sites along with a set of continuous and/or discrete covariates. The proposed model allows for joint examination of the dependence structure of the discrete response and observed covariates and also accommodates site-to-site variability. We develop the graphical model properties and theoretical justifications of this model. Our model has several advantages over the traditional logistic normal model used to analyze similar compositional data, including site-specific random effect terms and the incorporation of discrete and continuous covariates.  相似文献   

18.
Trace regression models are widely used in applications involving panel data, images, genomic microarrays, etc., where high-dimensional covariates are often involved. However, the existing research involving high-dimensional covariates focuses mainly on the condition mean model. In this paper, we extend the trace regression model to the quantile trace regression model when the parameter is a matrix of simultaneously low rank and row (column) sparsity. The convergence rate of the penalized estimator is derived under mild conditions. Simulations, as well as a real data application, are also carried out for illustration.  相似文献   

19.
The use of maximum likelihood methods in analysing times to failure in the presence of unobserved randomly changing covariates requires constrained optimization procedures. An alternative approach using a generalized version of the EM-algorithm requires smoothed estimates of covariate values. Similar estimates are needed in evaluating past exposures to hazardous chemicals, radiation or other toxic materials when health effects only become evident long after their use. In this paper, two kinds of equation for smoothing estimates of unobserved covariates in survival problems are derived. The first shows how new information may be used to update past estimates of the covariates' values. The second can be used to project the covariates' trajectory from the present to the past. If the hazard function is quadratic in form, both types of smoothing equation can be derived in a closed analytical form. Examples of both types of equation are presented. Use of these equations in the extended EM-algorithm, and in estimating past exposures to hazardous materials, are discussed. © 1997 by John Wiley & Sons, Ltd.  相似文献   

20.
We introduce a mixed regression model for mortality data which can be decomposed into a deterministic trend component explained by the covariates age and calendar year, a multivariate Gaussian time series part not explained by the covariates, and binomial risk. Data can be analyzed by means of a simple logistic regression model when the multivariate Gaussian time series component is absent and there is no overdispersion. In this paper we rather allow for overdispersion and the mixed regression model is fitted to mortality data from the United States and Sweden, with the aim to provide prediction and intervals for future mortality and annuity premium, as well as smoothing historical data, using the best linear unbiased predictor. We find that the form of the Gaussian time series has a large impact on the width of the prediction intervals, and it poses some new questions on proper model selection.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号