首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 625 毫秒
1.

This paper is concerned with the error density estimation in high-dimensional sparse linear model, where the number of variables may be larger than the sample size. An improved two-stage refitted cross-validation procedure by random splitting technique is used to obtain the residuals of the model, and then traditional kernel density method is applied to estimate the error density. Under suitable sparse conditions, the large sample properties of the estimator including the consistency and asymptotic normality, as well as the law of the iterated logarithm are obtained. Especially, we gave the relationship between the sparsity and the convergence rate of the kernel density estimator. The simulation results show that our error density estimator has a good performance. A real data example is presented to illustrate our methods.

  相似文献   

2.
In most of the earlier research for multiple zeros, in order to obtain a new iteration function from the existing scheme, the usual practice is to make no change at the first substep. In this paper, we explore the idea that what are the advantages if the flexibility of choice is also given at the first substep. Therefore, we present a new two-point sixth-order scheme for multiple roots (m>1). The main advantages of our scheme over the existing schemes are flexibility at both substeps, simple body structure, smaller residual error, smaller error difference between two consecutive iterations, and smaller asymptotic error constant. The development of the scheme is based on midpoint formula and weight functions of two variables. We compare our methods with the existing methods of the same order with real-life applications as well as standard test problems. From the numerical results, we find that our methods can be considered as better alternates for the existing methods of the same order. Finally, dynamical study of the proposed schemes is presented that confirms the theoretical results.  相似文献   

3.

We consider hypothesis testing for high-dimensional covariance structures in which the covariance matrix is a (i) scaled identity matrix, (ii) diagonal matrix, or (iii) intraclass covariance matrix. Our purpose is to systematically establish a nonparametric approach for testing the high-dimensional covariance structures (i)–(iii). We produce a new common test statistic for each covariance structure and show that the test statistic is an unbiased estimator of its corresponding test parameter. We prove that the test statistic establishes the asymptotic normality. We propose a new test procedure for (i)–(iii) and evaluate its asymptotic size and power theoretically when both the dimension and sample size increase. We investigate the performance of the proposed test procedure in simulations. As an application of testing the covariance structures, we give a test procedure to identify an eigenvector. Finally, we demonstrate the proposed test procedure by using a microarray data set.

  相似文献   

4.

This paper is devoted to the nonparametric estimation of the derivative of the regression function in a nonparametric regression model. We implement a very efficient and easy to handle statistical procedure based on the derivative of the recursive Nadaraya–Watson estimator. We establish the almost sure convergence as well as the asymptotic normality for our estimates. We also illustrate our nonparametric estimation procedure on simulated data and real life data associated with sea shores water quality and valvometry.

  相似文献   

5.
The aim of this paper is to introduce a new methodology for operational risk management, based on Bayesian copulae. One of the main problems related to operational risk management is understanding the complex dependence structure of the associated variables. In order to model this structure in a flexible way, we construct a method based on copulae. This allows us to split the joint multivariate probability distribution of a random vector of losses into individual components characterized by univariate marginals. Thus, copula functions embody all the information about the correlation between variables and provide a useful technique for modelling the dependency of a high number of marginals. Another important problem in operational risk modelling is the lack of loss data. This suggests the use of Bayesian models, computed via simulation methods and, in particular, Markov chain Monte Carlo. We propose a new methodology for modelling operational risk and for estimating the required capital. This methodology combines the use of copulae and Bayesian models.   相似文献   

6.
Heatwaves are defined as a set of hot days and nights that cause a marked short-term increase in mortality. Obtaining accurate estimates of the probability of an event lasting many days is important. Previous studies of temporal dependence of extremes have assumed either a first-order Markov model or a particularly strong form of extremal dependence, known as asymptotic dependence. Neither of these assumptions is appropriate for the heatwaves that we observe for our data. A first-order Markov assumption does not capture whether the previous temperature values have been increasing or decreasing and asymptotic dependence does not allow for asymptotic independence, a broad class of extremal dependence exhibited by many processes including all non-trivial Gaussian processes. This paper provides a kth-order Markov model framework that can encompass both asymptotic dependence and asymptotic independence structures. It uses a conditional approach developed for multivariate extremes coupled with copula methods for time series. We provide novel methods for the selection of the order of the Markov process that are based upon only the structure of the extreme events. Under this new framework, the observed daily maximum temperatures at Orleans, in central France, are found to be well modelled by an asymptotically independent third-order extremal Markov model. We estimate extremal quantities, such as the probability of a heatwave event lasting as long as the devastating European 2003 heatwave event. Critically our method enables the first reliable assessment of the sensitivity of such estimates to the choice of the order of the Markov process.  相似文献   

7.
We observe n events occurring in (0, T] taken from a Poisson process. The intensity function of the process is assumed to be a step function with multiple changepoints. This article proposes a Bayesian binary segmentation procedure for locating the changepoints and the associated heights of the intensity function. We conduct a sequence of nested hypothesis tests using the Bayes factor or the BIC approximation to the Bayes factor. At each comparison in the binary segmentation steps, we need only to compare a singlechangepoint model to a no-changepoint model. Therefore, this method circumvents the computational complexity we would normally face in problems with an unknown (large) number of dimensions. A simulation study and an analysis on a real dataset are given to illustrate our methods.  相似文献   

8.
We consider a modified two-stage procedure for constructing a fixed-width confidence interval for the mean of a U-statistic. First, we discuss a few asymptotic results with the associated rates of convergence. The main result gives the rate of convergence for the coverage probability of our proposed confidence interval which is seen to be slower than that for the purely sequential procedure.  相似文献   

9.
In order to solve the location problem in the p-median form we present an approximation algorithm with time complexity O(n 2) and the results of its probabilistic analysis. The input data are defined on a complete graph with distances between the vertices expressed by the independent random variables with the same uniform distribution. The value of the objective function produced by the algorithm amounts to a certain sum of the random variables that we analyze basing on an estimate for the probabilities of large deviations of these sums. We use a limit theorem in the form of the Petrov inequalities, taking into account the dependence of the random variables in the sum. The probabilistic analysis yields some estimates for the relative error and the failure probability of our algorithm, as well as conditions for its asymptotic exactness.  相似文献   

10.
We propose a model of an age-structured population divided into N geographical patches. We distinguish two time scales, at the fast time scale we have the migration dynamics and at the slow time scale the demographic dynamics. The demographic process is described using the classical McKendrick-von Foerster model for each patch, and a simple matrix model including the transfer rates between patches depicts the migration process.Assuming that 0 is a simple strictly dominant eigenvalue for the migration matrix, we transform the model (an e.d.p. problem with N state variables) into a classical McKendrick-von Foerster model (scalar e.d.p. problem) for the global variable: total population density. We prove, under certain assumptions, that the semigroup associated to our problem has the property of positive asynchronous exponential growth and so we compare its asymptotic behaviour to that of the transformed scalar model. This type of study can be included in the so-called aggregation methods, where a large scale dynamical system is approximately described by a reduced system. Aggregation methods have been already developed for systems of ordinary differential equations and for discrete time models.An application of the results to the study of the dynamics of the Sole larvae is also provided.  相似文献   

11.

We propose a new class of convex approximations for two-stage mixed-integer recourse models, the so-called generalized alpha-approximations. The advantage of these convex approximations over existing ones is that they are more suitable for efficient computations. Indeed, we construct a loose Benders decomposition algorithm that solves large problem instances in reasonable time. To guarantee the performance of the resulting solution, we derive corresponding error bounds that depend on the total variations of the probability density functions of the random variables in the model. The error bounds converge to zero if these total variations converge to zero. We empirically assess our solution method on several test instances, including the SIZES and SSLP instances from SIPLIB. We show that our method finds near-optimal solutions if the variability of the random parameters in the model is large. Moreover, our method outperforms existing methods in terms of computation time, especially for large problem instances.

  相似文献   

12.
We analyze the error in finite element methods in approximating, so-called, free or natural convection problems. We also include the effects of conducting solid walls in our analysis. Under a uniqueness condition on the Rayleigh and Prandtl numbers (which we derive), we give direct, quasioptimal error estimates for “div-stable” finite element spaces for the fluid variables and general conforming finite element spaces for the temperature. At larger Rayleigh numbers, we give analogous, asymptotic error estimates, basing this analysis upon local uniqueness properties of the true solution (u p T), which we also establish.  相似文献   

13.
Testing point null hypotheses is a very common activity in various applied situations. However, the existing Bayesian testing procedure may give evidence which does not agree with the classical frequentist p-value in many point null testing situations. A typical example for this is the well known Lindley’s paradox (Lindley in Biometrika 44:187–192, 1957). In this paper we propose an alternative testing procedure in the Bayesian framework. It is shown that for many classical testing examples, the Bayesian evidence derived by our new testing procedure is not contradictory to its frequentist counterpart any more. In fact, the new Bayesian evidence under the noninformative prior is usually coincident with the frequentist observed significance level.  相似文献   

14.
We introduce a test for the lack of dependence between two random variables valued into real Hilbert spaces. Here, we consider lack of dependence in the broader sense, that is, non-correlation. The test statistic is similar to the one proposed by Kokoszka et al. (2008) for testing for no effect in the linear functional model. The asymptotic distribution under the null hypothesis of this statistic is obtained as well as a consistency result for the proposed test. Applications to the case of functional variables are indicated and simulations show, in this context, the performance of the proposed method.  相似文献   

15.
The traditional approach to multivariate extreme values has been through the multivariate extreme value distribution G, characterised by its spectral measure H and associated Pickands’ dependence function A. More generally, for all asymptotically dependent variables, H determines the probability of all multivariate extreme events. When the variables are asymptotically dependent and under the assumption of unit Fréchet margins, several methods exist for the estimation of G, H and A which use variables with radial component exceeding some high threshold. For each of these characteristics, we propose new asymptotically consistent nonparametric estimators which arise from Heffernan and Tawn’s approach to multivariate extremes that conditions on variables with marginal values exceeding some high marginal threshold. The proposed estimators improve on existing estimators in three ways. First, under asymptotic dependence, they give self-consistent estimators of G, H and A; existing estimators are not self-consistent. Second, these existing estimators focus on the bivariate case, whereas our estimators extend easily to describe dependence in the multivariate case. Finally, for asymptotically independent cases, our estimators can model the level of asymptotic independence; whereas existing estimators for the spectral measure treat the variables as either being independent, or asymptotically dependent. For asymptotically dependent bivariate random variables, the new estimators are found to compare favourably with existing estimators, particularly for weak dependence. The method is illustrated with an application to finance data.  相似文献   

16.
Markov chain Monte Carlo (MCMC) methods for Bayesian computation are mostly used when the dominating measure is the Lebesgue measure, the counting measure, or a product of these. Many Bayesian problems give rise to distributions that are not dominated by the Lebesgue measure or the counting measure alone. In this article we introduce a simple framework for using MCMC algorithms in Bayesian computation with mixtures of mutually singular distributions. The idea is to find a common dominating measure that allows the use of traditional Metropolis-Hastings algorithms. In particular, using our formulation, the Gibbs sampler can be used whenever the full conditionals are available. We compare our formulation with the reversible jump approach and show that the two are closely related. We give results for three examples, involving testing a normal mean, variable selection in regression, and hypothesis testing for differential gene expression under multiple conditions. This allows us to compare the three methods considered: Metropolis-Hastings with mutually singular distributions, Gibbs sampler with mutually singular distributions, and reversible jump. In our examples, we found the Gibbs sampler to be more precise and to need considerably less computer time than the other methods. In addition, the full conditionals used in the Gibbs sampler can be used to further improve the estimates of the model posterior probabilities via Rao-Blackwellization, at no extra cost.  相似文献   

17.
This research paper simulated hypothesis testing of the differences of means, when the conventional assumption of independence within one of the samples had been violated. The study ran separate Monte Carlo simulations in which both samples came from uniform and normal populations. Dependence was introduced by multiplying the randomly generated scores within one sample by a predetermined factor. Then the simulation collected data on 10,000 paired samples with factors ranging from 1.0 (independence) to 2.0 (the highest level of dependence). A separate study calculated the mean autocorrelation associated with different conditions of dependence and linked this autocorrelation to the adjusted level of Type I error ( level). The results demonstrated a systematic increase in Type I error as the level of autocorrelation increased. The level that our study found for certain levels of dependence (with n = 30) far exceeded the asymptotic level of adjusted , suggesting that we further explore the effects of autocorrelation on conventional hypothesis testing.  相似文献   

18.
Daw  Andrew  Pender  Jamol 《Queueing Systems》2019,91(3-4):367-401

Queues that feature multiple entities arriving simultaneously are among the oldest models in queueing theory, and are often referred to as “batch” (or, in some cases, “bulk”) arrival queueing systems. In this work, we study the effect of batch arrivals on infinite server queues. We assume that the arrival epochs occur according to a Poisson process, with treatment of both stationary and non-stationary arrival rates. We consider both exponentially and generally distributed service durations, and we analyze both fixed and random arrival batch sizes. In addition to deriving the transient mean, variance, and moment-generating function for time-varying arrival rates, we also find that the steady-state distribution of the queue is equivalent to the sum of scaled Poisson random variables with rates proportional to the order statistics of its service distribution. We do so through viewing the batch arrival system as a collection of correlated sub-queues. Furthermore, we investigate the limiting behavior of the process through a batch scaling of the queue and through fluid and diffusion limits of the arrival rate. In the course of our analysis, we make important connections between our model and the harmonic numbers, generalized Hermite distributions, and truncated polylogarithms.

  相似文献   

19.

Many methods have been developed for analyzing survival data which are commonly right-censored. These methods, however, are challenged by complex features pertinent to the data collection as well as the nature of data themselves. Typically, biased samples caused by left-truncation (or length-biased sampling) and measurement error often accompany survival analysis. While such data frequently arise in practice, little work has been available to simultaneously address these features. In this paper, we explore valid inference methods for handling left-truncated and right-censored survival data with measurement error under the widely used Cox model. We first exploit a flexible estimator for the survival model parameters which does not require specification of the baseline hazard function. To improve the efficiency, we further develop an augmented nonparametric maximum likelihood estimator. We establish asymptotic results and examine the efficiency and robustness issues for the proposed estimators. The proposed methods enjoy appealing features that the distributions of the covariates and of the truncation times are left unspecified. Numerical studies are reported to assess the finite sample performance of the proposed methods.

  相似文献   

20.
Stimulated by a study in Bangladesh about the first birth interval, we propose a semivarying-coefficient model for cluster data analysis. We consider the estimation procedure for the proposed model and establish the asymptotic results of the proposed estimators. Furthermore, we employ the cross-validation (CV) to identify the constant coefficients. The associated asymptotic properties are rigorously examined. Simulation studies are conducted to investigate the performance of the proposed estimation and the CV-based model selection procedure for finite sample size. Finally, our methods are used to analyse the aforementioned data set to explore how several factors affect the first birth interval in Bangladesh.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号