首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 250 毫秒
1.
针对天波超视距雷达杂波呈现非均匀非平稳分布的特性以及由虚假杂波多而带来的虚假航迹多等问题,在Viterbi算法中引入了自适应杂波密度模型,在此基础上提出了一种基于自适应杂波密度模型的Viterbi数据关联算法(ACM-VDA).仿真结果表明,与VDA算法及 VDA-AI算法相比,ACM-VDA算法大大降低了虚假航迹的条数.  相似文献   

2.
A multiresolutional approach for measurement decomposition and system modeling is presented in this paper. The decomposition is performed in both spatial and time domains and provides an excellent platform for developing computationally efficient algorithms. Using multiresolutional decomposition and modeling, a multiresolutional joint probabilistic data association (MR-JPDA) algorithm is developed for multiple target tracking. Monte Carlo simulations demonstrate that the computation of the MRJPDA algorithm is much less than the traditional joint probabilistic data association (JPDA) algorithm with a comparable performance.  相似文献   

3.
Vertex coloring problem is a combinatorial optimization problem in graph theory in which a color is assigned to each vertex of graph such that no two adjacent vertices have the same color. In this paper a new hybrid algorithm which is obtained from combination of cellular learning automata (CLA) and memetic algorithm (MA) is proposed for solving the vertex coloring problem. CLA is an effective probabilistic learning model combining cellular automata and learning automaton (LA). Irregular CLA (ICLA) is a generalization of CLA in which the restriction of rectangular grid structure in CLA is removed. The proposed algorithm is based on the irregular open CLA (IOCLA) that is an extension of ICLA in which the evolution of CLA is influenced by both local and global environments. Similar to other IOCLA-based algorithms, in the proposed algorithm, local environment is constituted by neighboring LAs of any cell and the global environment consists of a pool of memes in which each meme corresponds to a certain local search method. Each meme is represented by a set of LAs from which the history of the corresponding local search method can be extracted. To show the superiority of the proposed algorithm over some well-known algorithms, several computer experiments have been conducted. The results show that the proposed algorithm performs better than other methods in terms of running time of algorithm and the required number of colors.  相似文献   

4.
Consistency of regularized spectral clustering   总被引:1,自引:0,他引:1  
Clustering is a widely used technique in machine learning, however, relatively little research in consistency of clustering algorithms has been done so far. In this paper we investigate the consistency of the regularized spectral clustering algorithm, which has been proposed recently. It provides a natural out-of-sample extension for spectral clustering. The presence of the regularization term makes our situation different from that in previous work. Our approach is mainly an elaborate analysis of a functional named the clustering objective. Moreover, we establish a convergence rate. The rate depends on the approximation property and the capacity of the reproducing kernel Hilbert space measured by covering numbers. Some new methods are exploited for the analysis since the underlying setting is much more complicated than usual. Some new methods are exploited for the analysis since the underlying setting is much more complicated than usual.  相似文献   

5.
Probabilistic proximity searching algorithms based on compact partitions   总被引:1,自引:0,他引:1  
The main bottleneck of the research in metric space searching is the so-called curse of dimensionality, which makes the task of searching some metric spaces intrinsically difficult, whatever algorithm is used. A recent trend to break this bottleneck resorts to probabilistic algorithms, where it has been shown that one can find 99% of the relevant objects at a fraction of the cost of the exact algorithm. These algorithms are welcome in most applications because resorting to metric space searching already involves a fuzziness in the retrieval requirements. In this paper, we push further in this direction by developing probabilistic algorithms on data structures whose exact versions are the best for high dimensions. As a result, we obtain probabilistic algorithms that are better than the previous ones. We give new insights on the problem and propose a novel view based on time-bounded searching. We also propose an experimental framework for probabilistic algorithms that permits comparing them in offline mode.  相似文献   

6.
Feature Selection (FS) is an important pre-processing step in data mining and classification tasks. The aim of FS is to select a small subset of most important and discriminative features. All the traditional feature selection methods assume that the entire input feature set is available from the beginning. However, online streaming features (OSF) are an integral part of many real-world applications. In OSF, the number of training examples is fixed while the number of features grows with time as new features stream in. A critical challenge for online streaming feature selection (OSFS) is the unavailability of the entire feature set before learning starts. Several efforts have been made to address the OSFS problem, however they all need some prior knowledge about the entire feature space to select informative features. In this paper, the OSFS problem is considered from the rough sets (RS) perspective and a new OSFS algorithm, called OS-NRRSAR-SA, is proposed. The main motivation for this consideration is that RS-based data mining does not require any domain knowledge other than the given dataset. The proposed algorithm uses the classical significance analysis concepts in RS theory to control the unknown feature space in OSFS problems. This algorithm is evaluated extensively on several high-dimensional datasets in terms of compactness, classification accuracy, run-time, and robustness against noises. Experimental results demonstrate that the algorithm achieves better results than existing OSFS algorithms, in every way.  相似文献   

7.
In this paper, we propose a testing-coverage software reliability model that considers not only the imperfect debugging (ID) but also the uncertainty of operating environments based on a non-homogeneous Poisson process (NHPP). Software is usually tested in a given control environment, but it may be used in different operating environments by different users, which are unknown to the developers. Many NHPP software reliability growth models (SRGMs) have been developed to estimate the software reliability measures, but most of the underlying common assumptions of these models are that the operating environment is the same as the developing environment. But in fact, due to the unpredictability of the uncertainty in the operating environments for the software, environments may considerably influence the reliability and software's performance in an unpredictable way. So when a software system works in a field environment, its reliability is usually different from the theory reliability, and also from all its similar applications in other fields. In this paper, a new model is proposed with the consideration of the fault detection rate based on the testing coverage and examined to cover ID subject to the uncertainty of operating environments. We compare the performance of the proposed model with several existing NHPP SRGMs using three sets of real software failure data based on seven criteria. Improved normalized criteria distance (NCD) method is also used to rank and select the best model in the context of a set of goodness-of-fit criteria taken all together. All results demonstrate that the new model can give a significant improved goodness-of-fit and predictive performance. Finally, the optimal software release time based on cost and reliability requirement and its sensitivity analysis are discussed.  相似文献   

8.
We consider high-dimensional data which contains a linear low-dimensional non-Gaussian structure contaminated with Gaussian noise, and discuss a method to identify this non-Gaussian subspace. For this problem, we provided in our previous work a very general semi-parametric framework called non-Gaussian component analysis (NGCA). NGCA has a uniform probabilistic bound on the error of finding the non-Gaussian components and within this framework, we presented an efficient NGCA algorithm called Multi-index Projection Pursuit. The algorithm is justified as an extension of the ordinary projection pursuit (PP) methods and is shown to outperform PP particularly when the data has complicated non-Gaussian structure. However, it turns out that multi-index PP is not optimal in the context of NGCA. In this article, we therefore develop an alternative algorithm called iterative metric adaptation for radial kernel functions (IMAK), which is theoretically better justifiable within the NGCA framework. We demonstrate that the new algorithm tends to outperform existing methods through numerical examples.  相似文献   

9.
Many multiple attribute decision analysis (MADA) problems are characterised by both quantitative and qualitative attributes with various types of uncertainties. Incompleteness (or ignorance) and vagueness (or fuzziness) are among the most common uncertainties in decision analysis. The evidential reasoning (ER) approach has been developed in the 1990s and in the recent years to support the solution of MADA problems with ignorance, a kind of probabilistic uncertainty. In this paper, the ER approach is further developed to deal with MADA problems with both probabilistic and fuzzy uncertainties.In this newly developed ER approach, precise data, ignorance and fuzziness are all modelled under the unified framework of a distributed fuzzy belief structure, leading to a fuzzy belief decision matrix. A utility-based grade match method is proposed to transform both numerical data and qualitative (fuzzy) assessment information of various formats into the fuzzy belief structure. A new fuzzy ER algorithm is developed to aggregate multiple attributes using the information contained in the fuzzy belief matrix, resulting in an aggregated fuzzy distributed assessment for each alternative. Different from the existing ER algorithm that is of a recursive nature, the new fuzzy ER algorithm provides an analytical means for combining all attributes without iteration, thus providing scope and flexibility for sensitivity analysis and optimisation. A numerical example is provided to illustrate the detailed implementation process of the new ER approach and its validity and wide applicability.  相似文献   

10.
An approach to non-convex multi-objective optimization problems is considered where only the values of objective functions are required by the algorithm. The proposed approach is a generalization of the probabilistic branch-and-bound approach well applicable to complicated problems of single-objective global optimization. In the present paper the concept of probabilistic branch-and-bound based multi-objective optimization algorithms is discussed, and some illustrations are presented.  相似文献   

11.
A new algorithm based on evolutionary computation concepts is presented in this paper. This algorithm is a non linear evolutive filter known as the Evolutive Localization Filter (ELF) which is able to solve the global localization problem in a robust and efficient way. The proposed algorithm searches stochastically along the state space for the best robot pose estimate. The set of pose solutions (the population) represents the most likely areas according to the perception and motion information up to date. The population evolves by using the log-likelihood of each candidate pose according to the observation and the motion error derived from the comparison between observed and predicted data obtained from the probabilistic perception and motion model. The algorithm has been tested on a mobile robot equipped with a laser range finder to demonstrate the effectiveness, robustness and computational efficiency of the proposed approach.  相似文献   

12.
In this paper, we suggest a new steganographic spatial domain algorithm based on a single chaotic map. Unlike most existing steganographic algorithms, the proposed algorithm uses one chaotic map to determine the pixel position of the host color image, the channel (red, green or blue) and the bit position of the targeted value in which a sensitive information bit can be hidden. Furthermore, this algorithm can be regarded as a variable-sized embedding algorithm. Experimental results demonstrate that this algorithm can defeat many existing steganalytic attacks. In comparison with existing steganographic spatial domain based algorithms, the suggested algorithm is shown to have some advantages over existing ones, namely, larger key space and a higher level of security against some existing attacks.  相似文献   

13.
A deterministic global optimization algorithm for box-constrained problems is presented. The proposed approach is based on well-known non-uniform space covering technique. In the paper this approach is further elaborated. We propose a new techniques that enables a significant reduction of the search space by means of dropping parts of processed boxes. Also a new quadratic underestimation for the objective function based on hessian eigenvalues bounds is presented. It is shown how this underestimation can be improved by exploiting the first-order optimality conditions. In the experimental section we compare the proposed approach with existing methods and programming tools. Numerical tests indicate that the proposed algorithm is highly competitive with considered methods.  相似文献   

14.
In real-life projects, both the trade-off between the project cost and the project completion time, and the uncertainty of the environment are considerable aspects for decision-makers. However, the research on the time-cost trade-off problem seldom concerns stochastic environments. Besides, optimizing the expected value of the objective is the exclusive decision-making criterion in the existing models for the stochastic time-cost trade-off problem. In this paper, two newly developed alternative stochastic time-cost trade-off models are proposed, in which the philosophies of chance-constrained programming and dependent-chance programming are adopted for decision-making. In addition, a hybrid intelligent algorithm integrating stochastic simulations and genetic algorithm is designed to search the quasi-optimal schedules under different decision-making criteria. The goal of the paper is to reveal how to obtain the optimal balance of the project completion time and the project cost in stochastic environments.  相似文献   

15.
话题发现是网络社交平台上进行热点话题预测的一个重要研究问题。针对已有话题发现算法大多基于传统余弦相似度衡量文本数据间的相似性,无法识别各维度取值成比例变化时数据对象间的差异,文本数据相似度计算结果不准确,影响话题发现正确率的问题,提出基于双向改进余弦相似度的话题发现算法(TABOC),首先从方向和取值两个角度改进余弦相似度,提出双向改进余弦相似度,能够区分各维度取值成比例变化的数据对象,保留传统余弦相似度在方向判别上的优势,提高衡量文本相似度的准确性;进一步定义集合的双向改进余弦特征向量和双向改进余弦特征向量的加法等相关定义定理,舍弃无关信息,直接计算新合并集合的特征向量,减小话题发现过程中的时间和空间消耗;还结合增量聚类框架,高效处理新增数据。采用百度贴吧数据进行实验表明,TABOC算法进行话题发现是有效可行的,算法正确率和时间效率总体上优于其他对比算法。  相似文献   

16.
研究了广泛存在于物流作业中一类新型的装箱问题,主要特征体现在箱子使用费用是关于装载率的凹函数。为求解问题,提出了一种基于分组编码策略的改进差分进化算法,以避免常规实数和整数编码方法存在放大搜索空间的不足。针对分组编码策略,定制化设计了以促进优秀基因传播为导向的新型变异和交叉操作,另外还嵌入了以物品置换为邻域的自适应局部搜索操作以增强局部搜索能力。对以往文献给出算例在不同凹费用函数下进行测试,实验结果显示所提出的算法明显优于BFD启发式算法,并且较遗传算法也有显著性改进。  相似文献   

17.
Storing XML documents in relational databases has drawn much attention in recent years because it can leverage existing investments in relational database technologies. Different algorithms have been proposed to map XML DTD/Schema to relational schema in order to store XML data in relational databases. However, most work defines mapping rules based on heuristics without considering application characteristics, hence fails to produce efficient relational schema for various applications. In this paper, we propose a workload-aware approach to generate relational schema from XML data and user specified workload. Our approach adopts the genetic algorithm to find optimal mappings. An elegant encoding method and related operations are proposed to manipulate mappings using bit strings. Various techniques for optimization can be applied to the XML to relational mapping problem based on this representation. We implemented the proposed algorithm and our experiment results showed that our algorithm was more robust and produced better mappings than existing work.  相似文献   

18.
AbstractIn this paper, a new superlinearly convergent algorithm of sequential systems of linear equations (SSLE) for nonlinear optimization problems with inequality constraints is proposed. Since the new algorithm only needs to solve several systems of linear equations having a same coefficient matrix per iteration, the computation amount of the algorithm is much less than that of the existing SQP algorithms per iteration. Moreover, for the SQP type algorithms, there exist so-called inconsistent problems, i.e., quadratic programming subproblems of the SQP algorithms may not have a solution at some iterations, but this phenomenon will not occur with the SSLE algorithms because the related systems of linear equations always have solutions. Some numerical results are reported.  相似文献   

19.
王竹芳  缪文清 《运筹与管理》2012,(1):142-146,179
本文通过对B运输问题建立数学模型,提出了一种求解B运输问题的改进解法。改进解法首先通过最小元素法求出初始解,然后进行变量闭回路法调整,直到求出最优解,并给出了一个计算实例证明了解法的有效性。文章还对改进解法和另外两种现有的算法进行了综合的分析,由于改进解法计算过程中采用的变量闭回路法省略了求检验数的环节,使得新算法比两种现有的算法更简便。  相似文献   

20.
叠前逆时偏移方法可以精确成像复杂地下构造,利用高阶有限差分方法求解声波方程,并给出了满足稳定性条件的采样间隔的选取方式.利用GPU/CPU加速技术实现地震资料的叠前逆时偏移算法,极大地提高了计算效率,算法也采用随机边界条件,节约了大量存储空间.分析了速度模型变化对成像结果的影响.复杂地震数据成像的测试结果表明,所述的叠前逆时偏移算法可清晰成像陡倾角成像清晰,对盐丘边界和内部构造成像效果也较好.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号