首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Sufficient dimension reduction (SDR) is a paradigm for reducing the dimension of the predictors without losing regression information. Most SDR methods require inverting the covariance matrix of the predictors. This hinders their use in the analysis of contemporary datasets where the number of predictors exceeds the available sample size and the predictors are highly correlated. To this end, by incorporating the seeded SDR idea and the sequential dimension-reduction framework, we propose a SDR method for high-dimensional data with correlated predictors. The performance of the proposed method is studied via extensive simulations. To demonstrate its use, an application to microarray gene expression data where the response is the production rate of riboflavin (vitamin B2) is presented.  相似文献   

2.
We present first methodology for dimension reduction in regressions with predictors that, given the response, follow one-parameter exponential families. Our approach is based on modeling the conditional distribution of the predictors given the response, which allows us to derive and estimate a sufficient reduction of the predictors. We also propose a method of estimating the forward regression mean function without requiring an explicit forward regression model. Whereas nearly all existing estimators of the central subspace are limited to regressions with continuous predictors only, our proposed methodology extends estimation to regressions with all categorical or a mixture of categorical and continuous predictors. Supplementary materials including the proofs and the computer code are available from the JCGS website.  相似文献   

3.
In this article, we review, consolidate and extend a theory for sufficient dimension reduction in regression settings. This theory provides a powerful context for the construction, characterization and interpretation of low-dimensional displays of the data, and allows us to turn graphics into a consistent and theoretically motivated methodological body. In this spirit, we propose an iterative graphical procedure for estimating the meta-parameter which lies at the core of sufficient dimension reduction; namely, the central dimension-reduction subspace.  相似文献   

4.
In this article, we consider sufficient dimension folding for the regression mean function when predictors are matrix- or array-valued. We propose a new concept named central mean dimension folding subspace and its two local estimation methods: folded outer product of gradients estimation (folded-OPG) and folded minimum average variance estimation (folded-MAVE). We establish the asymptotic properties for folded-MAVE. A modified BIC criterion is used to determine the dimensions of the central mean dimension folding subspace. We evaluate the performances of the two local estimation methods by simulated examples and demonstrate the efficacy of folded-MAVE in finite samples. And in particular, we apply our methods to analyze a longitudinal study of primary biliary cirrhosis. Supplementary materials for this article are available online.  相似文献   

5.
The conditional mean of the response given the predictors is often of interest in regression problems. The central mean subspace, recently introduced by Cook and Li, allows inference about aspects of the mean function in a largely nonparametric context. We propose a marginal fourth moments method for estimating directions in the central mean subspace that might be missed by existing methods such as ordinary least squares (OLS) and principal Hessian directions (pHd). Our method, targeting higher order trends, particularly cubics, complements OLS and pHd because there is no inclusion among them. Theory, estimation and inferences as well as illustrative examples are presented.  相似文献   

6.
The problem of finding the Euclidean distance between two convex polyhedra can be reduced to the combinatorial optimization problem of finding the minimum distance between their faces. This paper presents a global optimality criterion for this problem. An algorithm (QLDPA) for the fast computation of the distance between convex and bounded polyhedra is proposed as an application of it. Computer experiments show its fast performance, especially when the total number of vertices is large.  相似文献   

7.
马氏距离聚类分析中协方差矩阵估算的改进   总被引:1,自引:0,他引:1  
本文考虑了变量权重和样本类别的影响,建立了马氏距离聚类过程中评估协方差矩阵的迭代法。以Fisher的iris数据为样本,运用欧氏距离一般聚类、主成分聚类、改进前后的马氏距离聚类方法,进行实证分析和比较,结果表明本文所提出的新方法准确率至少提高了6.63%。最后,运用该方法对35个国家的相关指标数据进行聚类分析,确定了各国的卫生保健状况等级。  相似文献   

8.
In our previous work, we have given an algorithm for segmenting a simplex in the n-dimensional space into rt n+ 1 polyhedrons and provided map F which maps the n-dimensional unit cube to these polyhedrons. In this paper, we prove that the map F is a one to one correspondence at least in lower dimensional spaces (n _〈 3). Moreover, we propose the approximating subdivision and the interpolatory subdivision schemes and the estimation of computational complexity for triangular Bézier patches on a 2-dimensional space. Finally, we compare our schemes with Goldman's in computational complexity and speed.  相似文献   

9.
Sliced inverse regression (SIR) is an important method for reducing the dimensionality of input variables. Its goal is to estimate the effective dimension reduction directions. In classification settings, SIR is closely related to Fisher discriminant analysis. Motivated by reproducing kernel theory, we propose a notion of nonlinear effective dimension reduction and develop a nonlinear extension of SIR called kernel SIR (KSIR). Both SIR and KSIR are based on principal component analysis. Alternatively, based on principal coordinate analysis, we propose the dual versions of SIR and KSIR, which we refer to as sliced coordinate analysis (SCA) and kernel sliced coordinate analysis (KSCA), respectively. In the classification setting, we also call them discriminant coordinate analysis and kernel discriminant coordinate analysis. The computational complexities of SIR and KSIR rely on the dimensionality of the input vector and the number of input vectors, respectively, while those of SCA and KSCA both rely on the number of slices in the output. Thus, SCA and KSCA are very efficient dimension reduction methods.  相似文献   

10.
考虑响应变量随机缺失情形下的非线性EV模型.给出了未知参数的降维估计,有效避免了高维核估计带来的维数灾祸问题.所构造的统计量渐近于x~2分布,所得结果可以用来构造未知参数的置信域.  相似文献   

11.
Let{(t);t∈R_ ~N}be a d-dimensional N-parameter generalized Brownian sheet.Necessaryand sufficient conditions for a compact set E×F to be a polar set for(t,(t))are proved.It is also provedthat if 2N≤αd,then for any compact set ER_>~N,d-2/2 Dim E≤inf{dimF:F ∈ B(R~d),P{(E)∩F≠φ}>0}≤d-2/β DimE,and if 2N>αd,then for any compact set FR~d\{0},α/2(d-DimF)≤inf{dimE:E∈B(R_>~N),P{(E)∩F≠φ}>0}≤β/2(d-DimF),where B(R~d)and B(R_>~N)denote the Borel σ-algebra in R~d and in R_>~N respectively,dim and Dim are Hausdorffdimension and Packing dimension respectively.  相似文献   

12.
线性空间中次子空间的基和维数   总被引:1,自引:0,他引:1  
给出了线性空间中次子空间的基和维数的概念及性质,并以此刻画了非齐次线性方程组解的结构.  相似文献   

13.
In this paper, we propose a new estimate for dimension reduction, called the weighted variance estimate (WVE), which includes Sliced Average Variance Estimate (SAVE) as a special case. Bootstrap method is used to select the best estimate from the WVE and to estimate the structure dimension. And this selected best estimate usually performs better than the existing methods such as Sliced Inverse Regression (SIR), SAVE, etc. Many methods such as SIR, SAVE, etc. usually put the same weight on each observation to estimate central subspace (CS). By introducing a weight function, WVE puts different weights on different observations according to distance of observations from CS. The weight function makes WVE have very good performance in general and complicated situations, for example, the distribution of regressor deviating severely from elliptical distribution which is the base of many methods, such as SIR, etc. And compared with many existing methods, WVE is insensitive to the distribution of the regressor. The consistency of the WVE is established. Simulations to compare the performances of WVE with other existing methods confirm the advantage of WVE. This work was supported by National Natural Science Foundation of China (Grant No. 10771015)  相似文献   

14.
本文讨论了能量距离的两个问题.类似Brownian协方差的讨论提出了Brownian距离的定义,并证明了Brownian距离与能量距离的一致性.给出了配对变量的能量距离的表示,并探讨了将能量距离用于配对样本同分布的检验问题时原假设下的渐近分布理论.最后通过一个简单的数值模拟说明基于能量距离的配对样本的分布差异的检验方法比传统的t检验及Wilcoxon符号秩检验更有效.  相似文献   

15.
16.
The definition of Brownian distance is presented and it's proved that Brownian distance coincides with the energy distance with respect to Brownian motion. Energy distance for dependent random vectors is also given and the asymptotic distribution is derived under null hypothesis. A simple numerical simulation result shows that the method for paired-sample test based on energy distance can distinguish the distributions of the paired variables more effectively than the classical t-test and Wilcoxon signed rank test.  相似文献   

17.
18.
In this paper, we discuss distance measures for a number of different combinatorial optimization problems of which the solutions are best represented as permutations of items, sometimes composed of several permutation (sub)sets. The problems discussed include single-machine and multiple-machine scheduling problems, the traveling salesman problem, vehicle routing problems, and many others. Each of these problems requires a different distance measure that takes the specific properties of the representation into account. The distance measures discussed in this paper are based on a general distance measure for string comparison called the edit distance. We introduce several extensions to the simple edit distance, that can be used when a solution cannot be represented as a simple permutation, and develop algorithms to calculate them efficiently.  相似文献   

19.
20.
We modify and generalize the idea of covariance kernels for Borel probability measures on Rd, and study the relation between the central limit theorem in the total variation distance and the convergence of covariance kernels.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号