首页 | 本学科首页   官方微博 | 高级检索  
     检索      


Balancing principle in supervised learning for a general regularization scheme
Institution:1. School of Mathematical Sciences, Fudan University, Shanghai, 200433, China;2. Weierstrass Institute for Applied Analysis and Stochastics, Mohrenstrasse 39, 10117 Berlin, Germany;3. Johann Radon Institute for Computational and Applied Mathematics, Altenbergerstrasse 69, A-4040 Linz, Austria;1. Georgia State Business School, GSU, GA, United States of America;2. School of Electrical and Computer Engineering at Information Technology University, Lahore, Pakistan;3. Department of Mathematics and College of Computer and Information Science, Northeastern University, MA, United States of America;4. Department of Computational and Applied Mathematics, Rice University, TX, United States of America;1. Laboratoire de Mathématiques et Applications, UMR-CNRS 7348, Université de Poitiers, Téléport 2-BP30179, Boulevard Marie et Pierre Curie, 86962 Chasseneuil, France;2. Avignon Université, Laboratoire de Mathématiques d''Avignon (EA 2151), F-84018 Avignon, France;3. Inria, BIGS, Villers-lès-Nancy, F-54600, France;1. Department of Mathematics, and Department of Computational Mathematics, Science, and Engineering (CMSE), Michigan State University, East Lansing, MI, 48824, USA;2. Department of Mathematics, University of California San Diego, La Jolla, CA 92093, USA;3. Department of Mathematics and Statistics, University of Michigan – Dearborn, Dearborn, MI, 48128, USA
Abstract:We discuss the problem of parameter choice in learning algorithms generated by a general regularization scheme. Such a scheme covers well-known algorithms as regularized least squares and gradient descent learning. It is known that in contrast to classical deterministic regularization methods, the performance of regularized learning algorithms is influenced not only by the smoothness of a target function, but also by the capacity of a space, where regularization is performed. In the infinite dimensional case the latter one is usually measured in terms of the effective dimension. In the context of supervised learning both the smoothness and effective dimension are intrinsically unknown a priori. Therefore we are interested in a posteriori regularization parameter choice, and we propose a new form of the balancing principle. An advantage of this strategy over the known rules such as cross-validation based adaptation is that it does not require any data splitting and allows the use of all available labeled data in the construction of regularized approximants. We provide the analysis of the proposed rule and demonstrate its advantage in simulations.
Keywords:Supervised learning  General smoothness  Balancing principle
本文献已被 ScienceDirect 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号