首页 | 本学科首页   官方微博 | 高级检索  
     检索      

Deleting Outliers in Robust Regression with Mixed Integer Programming
作者姓名:GeorgiosZioutas  AntoniosAvramidis
作者单位:Division of Computational Methods and Computer Programming,General Department,Faculty of Technology,Aristotle University of Thessaloniki,541 24 Thessaloniki,Greece,Division of Computational Methods and Computer Programming,General Department,Faculty of Technology,Aristotle University of Thessaloniki,541 24 Thessaloniki,Greece
摘    要:In robust regression we often have to decide how many are the unusual observations, which should be removed from the sample in order to obtain better fitting for the rest of the observations. Generally, we use the basic principle of LTS, which is to fit the majority of the data, identifying as outliers those points that cause the biggest damage to the robust fit. However, in the LTS regression method the choice of default values for high break down-point affects seriously the efficiency of the estimator. In the proposed approach we introduce penalty cost for discarding an outlier, consequently, the best fit for the majority of the data is obtained by discarding only catastrophic observations. This penalty cost is based on robust design weights and high break down-point residual scale taken from the LTS estimator. The robust estimation is obtained by solving a convex quadratic mixed integer programming problem, where in the objective function the sum of the squared residuals and penalties for discarding observations is minimized. The proposed mathematical programming formula is suitable for small-sample data. Moreover, we conduct a simulation study to compare other robust estimators with our approach in terms of their efficiency and robustness.

关 键 词:稳健回归  二次混合积分规划  最小平衡平方  离群删除
收稿时间:5 March 2003

Deleting Outliers in Robust Regression with Mixed Integer Programming
GeorgiosZioutas AntoniosAvramidis.Deleting Outliers in Robust Regression with Mixed Integer Programming[J].Acta Mathematicae Applicatae Sinica,2005,21(2):323-334.
Authors:Georgios Zioutas  Antonios Avramidis
Institution:(1) Division of Computational Methods and Computer Programming, General Department, Faculty of Technology, Aristotle University of Thessaloniki, 541 24 Thessaloniki, Greece
Abstract:In robust regression we often have to decide how many are the unusual observations, which should be removed from the sample in order to obtain better fitting for the rest of the observations. Generally, we use the basic principle of LTS, which is to fit the majority of the data, identifying as outliers those points that cause the biggest damage to the robust fit. However, in the LTS regression method the choice of default values for high break down-point affects seriously the efficiency of the estimator. In the proposed approach we introduce penalty cost for discarding an outlier, consequently, the best fit for the majority of the data is obtained by discarding only catastrophic observations. This penalty cost is based on robust design weights and high break down-point residual scale taken from the LTS estimator. The robust estimation is obtained by solving a convex quadratic mixed integer programming problem, where in the objective function the sum of the squared residuals and penalties for discarding observations is minimized. The proposed mathematical programming formula is suitable for small-sample data. Moreover, we conduct a simulation study to compare other robust estimators with our approach in terms of their efficiency and robustness.
Keywords:Robust regression  quadratic mixed integer programming  least trimmed squares  deleting outliers  penalty methods
本文献已被 CNKI 维普 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号