首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Exhaustive enumeration of Steiner Triple Systems is not feasible, due to the combinatorial explosion of instances. The next‐best hope is to quickly find a sample that is representative of isomorphism classes. Stinson's Hill‐Climbing algorithm [ 20 ] is widely used to produce random Steiner Triple Systems, and certainly finds a sample of systems quickly, but the sample is not uniformly distributed with respect to the isomorphism classes of STS with ν ≤ 19, and, in particular, we find that isomorphism classes with a large number of Pasch configurations are under‐represented. No analysis of the non‐uniformity of the distribution with respect to isomorphism classes or the intractability of obtaining a representative sample for ν > 19 is known. We also exhibit a modification to hill‐climbing that makes the sample if finds closer to the uniform distribution over isomorphism classes in return for a modest increase in running time. © 2007 Wiley Periodicals, Inc. J Combin Designs 15: 405–419, 2007  相似文献   

2.
We use proprietary data collected by SVB Analytics, an affiliate of Silicon Valley Bank, to forecast the retained earnings of privately held companies. Combining methods of principal component analysis (PCA) and L1/quantile regression, we build multivariate linear models that feature excellent in‐sample fit and strong out‐of‐sample predictive accuracy. The combined PCA and L1 technique effectively deals with multicollinearity and non‐normality of the data, and also performs favorably when compared against a variety of other models. Additionally, we propose a variable ranking procedure that explains which variables from the current quarter are most predictive of the next quarter's retained earnings. We fit models to the top five variables identified by the ranking procedure and thereby, discover interpretable models with excellent out‐of‐sample performance. Copyright © 2013 John Wiley & Sons, Ltd.  相似文献   

3.
Like ordinary Brownian motion, super‐Brownian motion, a central object in the theory of superprocesses, is a universal object arising in a variety of settings. Schilder‐type theorems and Cramér‐type theorems are two of the major topics for large‐deviation theory. A Schilder‐type (which is also a Cramér‐type) sample large deviation for super‐Brownian motions with a good rate function represented by a variation formula was established in 1993 and 1994; since then there have been very valuable contributions for giving an affirmative answer to the question of whether this sample large deviation holds with an explicit good rate function. In this paper, thanks to previous results on this issue and the Brownian snake, we establish such a large deviation for nonzero finite initial measures. © 2010 Wiley Periodicals, Inc.  相似文献   

4.
A discrete‐time mover‐stayer (MS) model is an extension of a discrete‐time Markov chain, which assumes a simple form of population heterogeneity. The individuals in the population are either stayers, who never leave their initial states or movers who move according to a Markov chain. We, in turn, propose an extension of the MS model by specifying the stayer's probability as a logistic function of an individual's covariates. Such extension has been recently discussed for a continuous time MS but has not been considered before for a discrete time one. This extension allows for an in‐sample classification of subjects who never left their initial states into stayers or movers. The parameters of an extended MS model are estimated using the expectation‐maximization algorithm. A novel bootstrap procedure is proposed for out of sample validation of the in‐sample classification. The bootstrap procedure is also applied to validate the in‐sample classification with respect to a more general dichotomy than the MS one. The developed methods are illustrated with the data set on installment loans. But they can be applied more broadly in credit risk area, where prediction of creditworthiness of a loan borrower or lessee is of major interest.  相似文献   

5.
In this paper, we introduce a robust extension of the three‐factor model of Diebold and Li (J. Econometrics, 130: 337–364, 2006) using the class of symmetric scale mixtures of normal distributions. Specific distributions examined include the multivariate normal, Student‐t, slash, and variance gamma distributions. In the presence of non‐normality in the data, these distributions provide an appealing robust alternative to the routine use of the normal distribution. Using a Bayesian paradigm, we developed an efficient MCMC algorithm for parameter estimation. Moreover, the mixing parameters obtained as a by‐product of the scale mixture representation can be used to identify outliers. Our results reveal that the Diebold–Li models based on the Student‐t and slash distributions provide significant improvement in in‐sample fit and out‐of‐sample forecast to the US yield data than the usual normal‐based model. Copyright © 2011 John Wiley & Sons, Ltd.  相似文献   

6.
In normal samples, the sample variance and sample mean are independent random variables, and the sample variance has, when properly normalized, a X2‐distribution. This paper gives elementary proofs by induction for these two facts.  相似文献   

7.
Count data, most often modeled by a Poisson distribution, are common in statistical process control. They are traditionally monitored by frequentist c or u charts, by cumulative sum and by exponentially weighted moving average charts. These charts all assume that the in‐control true mean is known, a common fiction that is addressed by gathering a large Phase I sample and using it to estimate the mean. “Self‐starting” proposals that ameliorate the need for a large Phase I sample have also appeared. All these methods are frequentist, ie, they allow only retrospective inference during Phase I, and they have no coherent way to incorporate less‐than‐perfect prior information about the in‐control mean. In this paper, we introduce a Bayesian procedure that can incorporate prior information, allow online inference, and should be particularly attractive for short‐run settings where large Phase I calibration exercises are impossible or unreasonable.  相似文献   

8.
Various charts such as |S|, W, and G are used for monitoring process dispersion. Most of these charts are based on the normality assumption, while exact distribution of the control statistic is unknown, and thus limiting distribution of control statistic is employed which is applicable for large sample sizes. In practice, the normality assumption of distribution might be violated, while it is not always possible to collect large sample size. Furthermore, to use control charts in practice, the in‐control state usually has to be estimated. Such estimation has a negative effect on the performance of control chart. Non‐parametric bootstrap control charts can be considered as an alternative when the distribution is unknown or a collection of large sample size is not possible or the process parameters are estimated from a Phase I data set. In this paper, non‐parametric bootstrap multivariate control charts |S|, W, and G are introduced, and their performances are compared against Shewhart‐type control charts. The proposed method is based on bootstrapping the data used for estimating the in‐control state. Simulation results show satisfactory performance for the bootstrap control charts. Ultimately, the proposed control charts are applied to a real case study.  相似文献   

9.
Statistical surveillance is a noteworthy endeavor in many health‐care areas such as epidemiology, hospital quality, infection control, and patient safety. For monitoring hospital adverse events, the Shewhart u‐control chart is the most used methodology. One possible issue of the u‐chart is that in health‐care applications the lower control limit (LCL) is often conventionally set to zero as the adverse events are rare and the sample sizes are not sufficiently large to obtain LCL greater than zero. Consequently, the control chart loses any ability to signal improvements. Furthermore, as the area of opportunity (sample size) is not constant over time, the in‐control and out‐of‐control run length performances of the monitoring scheme are unknown. In this article, on the basis of a real case and through an intensive simulation study, we first investigate the in‐control statistical properties of the u‐chart. Then we set up several alternative monitoring schemes with the same in‐control performances and their out‐of‐control properties are studied and compared. The aim is to identify the most suitable control chart considering jointly: the ability to detect unexpected changes (usually worsening), the ability to test the impact of interventions (usually improvements), and the ease of use and clarity of interpretation. The results indicate that the exponentially weighted moving average control chart derived under the framework of weighted likelihood ratio test has the best overall performance.  相似文献   

10.
In the tensor completion problem, one seeks to estimate a low‐rank tensor based on a random sample of revealed entries. In terms of the required sample size, earlier work revealed a large gap between estimation with unbounded computational resources (using, for instance, tensor nuclear norm minimization) and polynomial‐time algorithms. Among the latter, the best statistical guarantees have been proved, for third‐order tensors, using the sixth level of the sum‐of‐squares (sos ) semidefinite programming hierarchy. However, the sos approach does not scale well to large problem instances. By contrast, spectral methods—based on unfolding or matricizing the tensor—are attractive for their low complexity, but have been believed to require a much larger sample size. This paper presents two main contributions. First, we propose a new method, based on unfolding, which outperforms naive ones for symmetric kth‐order tensors of rank r. For this result we make a study of singular space estimation for partially revealed matrices of large aspect ratio, which may be of independent interest. For third‐order tensors, our algorithm matches the sos method in terms of sample size (requiring about rd3/2 revealed entries), subject to a worse rank condition (rd3/4 rather than rd3/2). We complement this result with a different spectral algorithm for third‐order tensors in the overcomplete (rd) regime. Under a random model, this second approach succeeds in estimating tensors of rank drd3/2 from about rd3/2 revealed entries. © 2018 Wiley Periodicals, Inc.  相似文献   

11.
During the sampling of particulate mixtures, samples taken are analyzed for their mass concentration, which generally has non‐zero sample‐to‐sample variance. Bias, variance, and mean squared error (MSE) of a number of variance estimators, derived by Geelhoed, were studied in this article. The Monte Carlo simulation was applied using an observable first‐order Markov Chain with transition probabilities that served as a model for the sample drawing process. Because the bias and variance of a variance estimator could depend on the specific circumstances under which it is applied, Monte Carlo simulation was performed for a wide range of practically relevant scenarios. Using the ‘smallest mean squared error’ as a criterion, an adaptation of an estimator based on a first‐order Taylor linearization of the sample concentration is the best. An estimator based on the Horvitz–Thompson estimator is not practically applicable because of the potentially high MSE for the cases studied. The results indicate that the Poisson estimator leads to a biased estimator for the variance of fundamental sampling error (up to 428% absolute value of relative bias) in case of low levels of grouping and segregation. The uncertainty of the results obtained by the simulations was also addressed and it was found that the results were not significantly affected. The potentials of a recently described other approach are discussed for extending the first‐order Markov Chain described here to account also for higher levels of grouping and segregation. Copyright © 2013 John Wiley & Sons, Ltd.  相似文献   

12.
Bandura (1997) contends that when compared to other sources of efficacy, mastery experiences, when presented appropriately, have the most powerful influence on self‐efficacy. The purpose of this study was to investigate the effects of community‐based service learning (CBSL) experiences on preservice elementary teachers' personal self‐efficacy beliefs about equitable science teaching and learning. Data were collected using pretests‐posttests and post‐questionnaires with the study sample. Findings from this study support Bandura's assertion. CBSL experiences were an important source of personal self‐efficacy and significantly influenced preservice elementary teachers' personal self‐efficacy beliefs about equitable science teaching and learning.  相似文献   

13.
A sample of 134 sixth‐grade students who were using the Connected Mathematics curriculum were administered an open‐ended item entitled, Vet Club (Balanced Assessment, 2000). This paper explores the role of misconceptions and naïve conceptions in the acquisition of statistical thinking for middle grades students. Students exhibited misconceptions and naïve conceptions regarding representing data graphically, interpreting the meaning of typicality, and plotting 0 above the x‐axis.  相似文献   

14.
A fundamental problem in financial trading is the correct and timely identification of turning points in stock value series. This detection enables to perform profitable investment decisions, such as buying‐at‐low and selling‐at‐high. This paper evaluates the ability of sequential smoothing methods to detect turning points in financial time series. The novel idea is to select smoothing and alarm coefficients on the gain performance of the trading strategy. Application to real data shows that recursive smoothers outperform two‐sided filters at the out‐of‐sample level. Copyright © 2012 John Wiley & Sons, Ltd.  相似文献   

15.
The mixed inverse Gaussian given by Whitmore (biScand. J. Statist., 13 , 1986, 211–220) provides a convenient way for testing the goodness‐of‐fit of a pure inverse Gaussian distribution. The test is a one‐sided score test with the null hypothesis being the pure inverse Gaussian (i.e. the mixing parameter is zero) and the alternative a mixture. We devise a simple score test and study its finite sample properties. Monte Carlo results show that it compares favourably with the smooth test of Ducharme ( Test , 10 , 2001, 271‐290). In practical applications, when the pure inverse Gaussian distribution is rejected, one is interested in making inference about the general values of the mixing parameter. However, as it is well known that the inverse Gaussian mixture is a defective distribution; hence, the standard likelihood inference cannot be applied. We propose several alternatives and provide score tests for the mixing parameter. Finite sample properties of these tests are examined by Monte Carlo simulation. Copyright © 2010 John Wiley & Sons, Ltd.  相似文献   

16.
We use computational phylogenetic techniques to solve a central problem in inferential network monitoring. More precisely, we design a novel algorithm for multicast‐based delay inference, that is, the problem of reconstructing delay characteristics of a network from end‐to‐end delay measurements on network paths. Our inference algorithm is based on additive metric techniques used in phylogenetics. It runs in polynomial time and requires a sample of size only poly(log n). We also show how to recover the topology of the routing tree. © 2010 Wiley Periodicals, Inc. Random Struct. Alg., 2010  相似文献   

17.
In this paper, we propose a stochastic conditional range model with leverage effect (henceforth SCRL) for volatility forecasting. A maximum likelihood method based on the particle filters is developed to estimate the parameters of the SCRL model. Simulation results show that the proposed methodology performs well. We apply the proposed model and methodology to four stock market indices, the Shanghai Stock Exchange Composite Index of China, the Hang Seng Index of Hong Kong, the Nikkei 225 Index of Japan, and the S&P 500 Index of US. Empirical results highlight the value of incorporating leverage effect into range modeling and forecasting. In particular, the results show that our SCRL model outperforms the conditional autoregressive range model, the conditional autoregressive range model with leverage effect, and the stochastic conditional range model in both in‐sample fit and out‐of‐sample forecast.  相似文献   

18.
We consider elliptic and parabolic problems in unbounded domains. We give general existence and regularity results in Besov spaces and semi‐explicit representation formulas via operator‐valued fundamental solutions which turn out to be a powerful tool to derive a series of qualitative results about the solutions. We give a sample of possible applications including asymptotic behavior in the large, singular perturbations, exact boundary conditions on artificial boundaries and validity of maximum principles. (© 2004 WILEY‐VCH Verlag GmbH & Co. KGaA, Weinheim)  相似文献   

19.
In the on‐line nearest‐neighbor graph (ONG), each point after the first in a sequence of points in ?d is joined by an edge to its nearest neighbor amongst those points that precede it in the sequence. We study the large‐sample asymptotic behavior of the total power‐weighted length of the ONG on uniform random points in (0,1)d. In particular, for d = 1 and weight exponent α > 1/2, the limiting distribution of the centered total weight is characterized by a distributional fixed‐point equation. As an ancillary result, we give exact expressions for the expectation and variance of the standard nearest‐neighbor (directed) graph on uniform random points in the unit interval. © 2007 Wiley Periodicals, Inc. Random Struct. Alg., 2008  相似文献   

20.
The analysis of multivariate time series is a common problem in areas like finance and economics. The classical tools for this purpose are vector autoregressive models. These however are limited to the modeling of linear and symmetric dependence. We propose a novel copula‐based model that allows for the non‐linear and non‐symmetric modeling of serial as well as between‐series dependencies. The model exploits the flexibility of vine copulas, which are built up by bivariate copulas only. We describe statistical inference techniques for the new model and discuss how it can be used for testing Granger causality. Finally, we use the model to investigate inflation effects on industrial production, stock returns and interest rates. In addition, the out‐of‐sample predictive ability is compared with relevant benchmark models. Copyright © 2014 John Wiley & Sons, Ltd.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号