DocumentCode :
730589
Title :
Rapid: Rapidly accelerated proximal gradient algorithms for convex minimization
Author :
Ziming Zhang ; Saligrama, Venkatesh
Author_Institution :
ECE, Boston Univ., Boston, MA, USA
fYear :
2015
fDate :
19-24 April 2015
Firstpage :
3796
Lastpage :
3800
Abstract :
In this paper, we propose a new algorithm to speed-up the convergence of accelerated proximal gradient (APG) methods. In order to minimize a convex function f(x), our algorithm introduces a simple line search step after each proximal gradient step in APG so that a biconvex function f(θx) is minimized over scalar variable θ > 0 while fixing variable x. We propose two new ways of constructing the auxiliary variables in APG based on the intermediate solutions of the proximal gradient and the line search steps. We prove that at arbitrary iteration step t(t ≥ 1), our algorithm can achieve a smaller upper-bound for the gap between the current and optimal objective values than those in the traditional APG methods such as FISTA [1], making it converge faster in practice. We apply our algorithm to many important convex optimization problems such as sparse linear regression. Our experimental results demonstrate that our algorithm converges faster than APG, even comparable to some sophisticated solvers.
Keywords :
compressed sensing; gradient methods; regression analysis; accelerated proximal gradient methods; biconvex function; convex minimization; sparse linear regression; Acceleration; Convergence; Convex functions; Gradient methods; Minimization; Signal processing algorithms; Upper bound;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location :
South Brisbane, QLD
Type :
conf
DOI :
10.1109/ICASSP.2015.7178681
Filename :
7178681
Link To Document :
بازگشت