首页 | 本学科首页   官方微博 | 高级检索  
     


A stochastic extra-step quasi-Newton method for nonsmooth nonconvex optimization
Authors:Yang  Minghan  Milzarek  Andre  Wen  Zaiwen  Zhang  Tong
Affiliation:1.Beijing International Center for Mathematical Research, BICMR, Peking University, Beijing, China
;2.School of Data Science SDS, The Chinese University of Hong Kong - Shenzhen, Shenzhen, Guangdong, China
;3.Shenzhen Research Institute of Big Data, SRIBD, Shenzhen, Guangdong, China
;4.Shenzhen Institute of Artificial Intelligence and Robotics for Society, AIRS, Shenzhen, Guangdong, China
;5.Center for Data Science, Peking University, Beijing, China
;6.Hong Kong University of Science and Technology, Hong Kong, China
;
Abstract:

In this paper, a novel stochastic extra-step quasi-Newton method is developed to solve a class of nonsmooth nonconvex composite optimization problems. We assume that the gradient of the smooth part of the objective function can only be approximated by stochastic oracles. The proposed method combines general stochastic higher order steps derived from an underlying proximal type fixed-point equation with additional stochastic proximal gradient steps to guarantee convergence. Based on suitable bounds on the step sizes, we establish global convergence to stationary points in expectation and an extension of the approach using variance reduction techniques is discussed. Motivated by large-scale and big data applications, we investigate a stochastic coordinate-type quasi-Newton scheme that allows to generate cheap and tractable stochastic higher order directions. Finally, numerical results on large-scale logistic regression and deep learning problems show that our proposed algorithm compares favorably with other state-of-the-art methods.

Keywords:
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号