首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 125 毫秒
1.
Dimension reduction techniques are at the core of the statistical analysis of high-dimensional and functional observations. Whether the data are vector- or function-valued, principal component techniques, in this context, play a central role. The success of principal components in the dimension reduction problem is explained by the fact that, for any \(K\le p\), the K first coefficients in the expansion of a p-dimensional random vector \(\mathbf{X}\) in terms of its principal components is providing the best linear K-dimensional summary of \(\mathbf X\) in the mean square sense. The same property holds true for a random function and its functional principal component expansion. This optimality feature, however, no longer holds true in a time series context: principal components and functional principal components, when the observations are serially dependent, are losing their optimal dimension reduction property to the so-called dynamic principal components introduced by Brillinger in 1981 in the vector case and, in the functional case, their functional extension proposed by Hörmann, Kidziński and Hallin in 2015.  相似文献   

2.
Although some of the earliest Estimation of Distribution Algorithms (EDAs) utilized bivariate marginal distribution models, up to now, all discrete bivariate EDAs had one serious limitation: they were constrained to exploiting only a limited O(d) subset out of all possible \(O(d^{2})\) bivariate dependencies. As a first we present a family of discrete bivariate EDAs that can learn and exploit all \(O(d^{2})\) dependencies between variables, and yet have the same run-time complexity as their more limited counterparts. This family of algorithms, which we label DICE (DIscrete Correlated Estimation of distribution algorithms), is rigorously based on sound statistical principles, and particularly on a modelling technique from statistical physics: dichotomised multivariate Gaussian distributions. Initially (Lane et al. in European Conference on the Applications of Evolutionary Computation, Springer, 1999), DICE was trialled on a suite of combinatorial optimization problems over binary search spaces. Our proposed dichotomised Gaussian (DG) model in DICE significantly outperformed existing discrete bivariate EDAs; crucially, the performance gap increasingly widened as dimensionality of the problems increased. In this comprehensive treatment, we generalise DICE by successfully extending it to multary search spaces that also allow for categorical variables. Because correlation is not wholly meaningful for categorical variables, interactions between such variables cannot be fully modelled by correlation-based approaches such as in the original formulation of DICE. Therefore, here we extend our original DG model to deal with such situations. We test DICE on a challenging test suite of combinatorial optimization problems, which are defined mostly on multary search spaces. While the two versions of DICE outperform each other on different problem instances, they both outperform all the state-of-the-art bivariate EDAs on almost all of the problem instances. This further illustrates that these innovative DICE methods constitute a significant step change in the domain of discrete bivariate EDAs.  相似文献   

3.
The first part of this paper introduces a class of discrete multivariate phase-type (MPH) distributions. Recursive formulas are found for joint probabilities. Explicit expressions are obtained for means, variances and co-variances. The discrete MPH-distributions are used in the second part of the paper to study multivariate insurance claim processes in risk analysis, where claims may arrive in batches, the arrivals of different types of batches may be correlated and the amounts of different types of claims in a batch may be dependent. Under certain conditions, it is shown that the total amounts of claims accumulated in some random time horizon are discrete MPH random vectors. Matrix-representations of the discrete MPH-distributions are constructed explicitly. Efficient computational methods are developed for computing risk measures of the total claims of different types of claim batches and individual types of claims (e.g., joint distribution, mean, variance, correlation and value at risk.)  相似文献   

4.
Both marginal and dependence features must be described when modelling the extremes of a stationary time series. There are standard approaches to marginal modelling, but long- and short-range dependence of extremes may both appear. In applications, an assumption of long-range independence often seems reasonable, but short-range dependence, i.e., the clustering of extremes, needs attention. The extremal index 0 < ?? ≤ 1 is a natural limiting measure of clustering, but for wide classes of dependent processes, including all stationary Gaussian processes, it cannot distinguish dependent processes from independent processes with ?? = 1. Eastoe and Tawn (Biometrika 99, 43–55 2012) exploit methods from multivariate extremes to treat the subasymptotic extremal dependence structure of stationary time series, covering both 0 < ?? < 1 and ?? = 1, through the introduction of a threshold-based extremal index. Inference for their dependence models uses an inefficient stepwise procedure that has various weaknesses and has no reliable assessment of uncertainty. We overcome these issues using a Bayesian semiparametric approach. Simulations and the analysis of a UK daily river flow time series show that the new approach provides improved efficiency for estimating properties of functionals of clusters.  相似文献   

5.
The Foster–Stuart test is often used in the analysis of time series to determine trends. The test is based on calculating the lower and upper time series records x1, …, xn. Unlike other tests of randomness, tests based on records are not invariant under a reversal of the direction of the time variable. To construct invariant round-trip tests, it is necessary to count the records in both forward and backward time variables. Thus far, the construction of this test has been impossible because the variances of test statistics were not known when the null hypothesis was true. Variances for Foster–Stuart round-trip test statistics D and S have been found in the present paper. The test was designed to identify positive and negative trends in means and variances of x1, …, xn time series. Dispersions of S and D test statistics have been found under the H0 randomness null hypothesis. Asymptotic approximations were obtained for dispersions. Thus, it has turned out to be possible to construct a full-fledged invariant test for two-sided alternatives. The Foster–Stuart round-trip test application example has been reviewed.  相似文献   

6.
Many companies use firm orders-to-date to make forecasts of units to be shipped at a future time t that is j periods away, j = 1, 2,..., h. A number of methods for making these forecasts were developed and evaluated using simulation. The time series for bookings was decomposed into a shipment time series and a time series for factors representing the fraction of shipments booked j periods ahead. Separate techniques were used for the shipments series (namely, naive, exponential smoothing and Bayesian procedures) and for the factors (naive and exponential smoothing procedures). The accuracy of these approaches, as well as an ARIMA model that ignored orders-to-date, was evaluated by using several simulated patterns of bookings. No approach was dominant, but one of the simplest approaches (naive/smoothing) did comparatively well.  相似文献   

7.
There exist two ways of defining regular variation of a time series in a star-shaped metric space: either by the distributions of finite stretches of the series or by viewing the whole series as a single random element in a sequence space. The two definitions are shown to be equivalent. The introduction of a norm-like function, called modulus, yields a polar decomposition similar to the one in Euclidean spaces. The angular component of the time series, called angular or spectral tail process, captures all aspects of extremal dependence. The stationarity of the underlying series induces a transformation formula of the spectral tail process under time shifts.  相似文献   

8.
A generalization of the block replacement policy (BRP) is proposed and analysed for a multi-unit system which has the specific multivariate distribution. Under such a policy an operating system is preventively replaced at times kT (k = 1, 2, 3,...), as in the ordinary BRP, and the replacement of the failed system at failure is not mandatory; instead, a minimal repair to the component of the system can be made. The choice of these two possible actions is based on some random mechanism which is age-dependent. The cost of the ith minimal repair of the component at age y depends on the random part C(y) and the deterministic part Ci(y). The aim of the paper is to find the optimal block interval T which minimizes the long-run expected cost per unit time of the policy.  相似文献   

9.
This paper examines the effects of permanent changes in the variance of the errors on routine applications of standard t-ratio test in regression models. It is shown the asymptotic distribution of t-ratio test is not invariant to non-stationary in variance, and the phenomenon of spurious regression will occur independently of the structure assumed for these time series. The intuition behind this is that the non-stationary volatility can increase persistency in the level of regression errors, which then leads to spurious correlation. Monte Carlo experiment evidence indicates that, in contrast to the broken level/trend case, the presence of spurious relationship critically depends on the location and magnitude of changes, regardless of the sample size. Finally, some real data sets from the Shanghai stock database are reported for illustration.  相似文献   

10.
We say that a convex set K in ? d strictly separates the set A from the set B if A ? int(K) and B ? cl K = ø. The well-known Theorem of Kirchberger states the following. If A and B are finite sets in ? d with the property that for every T ? A?B of cardinality at most d + 2, there is a half space strictly separating T ? A and T ? B, then there is a half space strictly separating A and B. In short, we say that the strict separation number of the family of half spaces in ? d is d + 2.In this note we investigate the problem of strict separation of two finite sets by the family of positive homothetic (resp., similar) copies of a closed, convex set. We prove Kirchberger-type theorems for the family of positive homothets of planar convex sets and for the family of homothets of certain polyhedral sets. Moreover, we provide examples that show that, for certain convex sets, the family of positive homothets (resp., the family of similar copies) has a large strict separation number, in some cases, infinity. Finally, we examine how our results translate to the setting of non-strict separation.  相似文献   

11.
Consider a k-out-of-n system where the lifetimes of the components are independent and identically distributed exponential (λ) random variables. Each component has its own repair facility, the repair times being independent and identically distributed exponential (μ) random variables, independent of the failure times. The mean operating time and mean repair time during the cycle between two successive breakdowns are found using renewal theory and the expression for the system availability. Using these, the mean first-passage times from any of the operating states of the system to the down state, and the mean first-passage times from any of the down states to the operating state are found recursively.  相似文献   

12.
In this work, we deal with hyperbolic Eisenstein series and more particularly, given a degenerating family S l (l ≥ 0) of Riemann Surfaces with their canonical hyperbolic metrics, we work out the degeneration of hyperbolic Eisenstein series associated to the pinching geodesics in S l . Our principal results are theorems 2.2, 4.1 et 4.2.  相似文献   

13.
An analysis taking into account the dependencies in the departure process from the first station of the M/E k /1→./M/1 system is conducted. Arrivals to the second station are approximated as a general independent distribution, and the waiting times in this station are compared to those found through computer simulation.  相似文献   

14.
We propose a new method to impute missing values in mixed data sets. It is based on a principal component method, the factorial analysis for mixed data, which balances the influence of all the variables that are continuous and categorical in the construction of the principal components. Because the imputation uses the principal axes and components, the prediction of the missing values is based on the similarity between individuals and on the relationships between variables. The properties of the method are illustrated via simulations and the quality of the imputation is assessed using real data sets. The method is compared to a recent method (Stekhoven and Buhlmann Bioinformatics 28:113–118, 2011) based on random forest and shows better performance especially for the imputation of categorical variables and situations with highly linear relationships between continuous variables.  相似文献   

15.
Let K be a field of characteristic p>0 and let f(t 1,…,t d ) be a power series in d variables with coefficients in K that is algebraic over the field of multivariate rational functions K(t 1,…,t d ). We prove a generalization of both Derksen’s recent analogue of the Skolem–Mahler–Lech theorem in positive characteristic and a classical theorem of Christol, by showing that the set of indices (n 1,…,n d )∈? d for which the coefficient of \(t_{1}^{n_{1}}\cdots t_{d}^{n_{d}}\) in f(t 1,…,t d ) is zero is a p-automatic set. Applying this result to multivariate rational functions leads to interesting effective results concerning some Diophantine equations related to S-unit equations and more generally to the Mordell–Lang Theorem over fields of positive characteristic.  相似文献   

16.
For a lattice L with 0 and 1, let Princ(L) denote the set of principal congruences of L. Ordered by set inclusion, it is a bounded ordered set. In 2013, G. Grätzer proved that every bounded ordered set is representable as Princ(L); in fact, he constructed L as a lattice of length 5. For {0, 1}-sublattices \({A \subseteq B}\) of L, congruence generation defines a natural map Princ(A) \({\longrightarrow}\) Princ(B). In this way, every family of {0, 1}-sublattices of L yields a small category of bounded ordered sets as objects and certain 0-separating {0, 1}-preserving monotone maps as morphisms such that every hom-set consists of at most one morphism. We prove the converse: every small category of bounded ordered sets with these properties is representable by principal congruences of selfdual lattices of length 5 in the above sense. As a corollary, we can construct a selfdual lattice L in G. Grätzer's above-mentioned result.  相似文献   

17.
We consider a class of mixture models for positive continuous data and the estimation of an underlying parameter θ of the mixing distribution. With a unified approach, we obtain classes of dominating estimators under squared error loss of an unbiased estimator, which include smooth estimators. Applications include estimating noncentrality parameters of chi-square and F-distributions, as well as ρ 2/(1 ? ρ 2), where ρ is amultivariate correlation coefficient in a multivariate normal set-up. Finally, the findings are extended to situations, where there exists a lower bound constraint on θ.  相似文献   

18.
Universal classes of Abelian groups are classified in terms of sets of finitely generated groups closed with respect to the discrimination operator. The notions of a principal universal class and a canonical group for such a class are introduced. For any universal class K, the class Kec of existentially closed groups generated by the universal theory of K is described. It is proved that Kec is axiomatizable and, therefore, the universal theory of K has a model companion.  相似文献   

19.
The paper announces some new inequalities that refer to broken lines, curves and real and complex functions. Their derivation is based on a new principle of angles and lengths for curves. The inequality in the complex analysis, called the principle of derivatives, is valid for analytic functions in arbitrary domains and extends to a broad class of sufficiently smooth complex functions. A new inequality follows for real functions of two variables concerning their level sets. For all cases mentioned above, both for curves and functions, we obtain some analogues of the second fundamental theorem in Nevanlinna theory of meromorphic functions. At the end we discuss a new point-domain inequality dealing with finite point sets in an arbitrary domain.  相似文献   

20.
Tensor-compressed numerical solution of elliptic multiscale-diffusion and high frequency scattering problems is considered. For either problem class, solutions exhibit multiple length scales governed by the corresponding scale parameter: the scale of oscillations of the diffusion coefficient or smallest wavelength, respectively. As is well-known, this imposes a scale-resolution requirement on the number of degrees of freedom required to accurately represent the solutions in standard finite-element (FE) discretizations. Low-order FE methods are by now generally perceived unsuitable for high-frequency coefficients in diffusion problems and high wavenumbers in scattering problems. Accordingly, special techniques have been proposed instead (such as numerical homogenization, heterogeneous multiscale method, oversampling, etc.) which require, in some form, a-priori information on the microstructure of the solution. We analyze the approximation properties of tensor-formatted, conforming first-order FE methods for scale resolution in multiscale problems without a-priori information. The FE methods are based on the dynamic extraction of principal components from stiffness matrices, load and solution vectors by the quantized tensor train (QTT) decomposition. For prototypical model problems, we prove that this approach, by means of the QTT reparametrization of the FE space, allows to identify effective degrees of freedom to replace the degrees of freedom of a uniform “virtual” (i.e. never directly accessed) mesh, whose number may be prohibitively large to realize computationally. Precisely, solutions of model elliptic homogenization and high-frequency acoustic scattering problems are proved to admit QTT-structured approximations whose number of effective degrees of freedom required to reach a prescribed approximation error scales polylogarithmically with respect to the reciprocal of the target Sobolev-norm accuracy ε with only a mild dependence on the scale parameter. No a-priori information on the nature of the problems and intrinsic length scales of the solution is required in the numerical realization of the presently proposed QTT-structured approach. Although only univariate model multiscale problems are analyzed in the present paper, QTT structured algorithms are applicable also in several variables. Detailed numerical experiments confirm the theoretical bounds. As a corollary of our analysis, we prove that for the mentioned model problems, the Kolmogorov n-widths of solution sets are exponentially small for analytic data, independently of the problems’ scale parameters. That implies, in particular, the exponential convergence of reduced basis techniques which is scale-robust, i.e., independent of the scale parameter in the problem.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号