DocumentCode :
358905
Title :
Stochastic approximation for global random optimization
Author :
Maryak, John L. ; Chin, Daniel C.
Author_Institution :
Appl. Phys. Lab., Johns Hopkins Univ., Laurel, MD, USA
Volume :
5
fYear :
2000
fDate :
2000
Firstpage :
3294
Abstract :
A desire with iterative optimization techniques is that the algorithm reach the global optimum rather than get stranded at a local optimum value. One method used to try to assure global convergence is the injection of extra noise terms into the recursion, which may allow the algorithm to escape local optimum points. The amplitude of the injected noise is decreased over time (a process called “annealing”), so that the algorithm can finally converge when it reaches the global optimum point. In this context, we examine a certain “gradient free” stochastic approximation algorithm called “SPSA,” that has performed well in complex optimization problems. We discuss conditions under which SPSA will converge globally using injected noise. In a separate section, we show that, under different conditions, “basic” SPSA (i.e., without injected noise) can achieve a standard type of convergence to a global optimum. The discussion is supported by a numerical study
Keywords :
approximation theory; convergence; iterative methods; optimisation; SPSA; annealing; complex optimization problems; global optimum; global random optimization; injected noise; iterative optimization techniques; stochastic approximation; Annealing; Approximation algorithms; Convergence; Iterative algorithms; Laboratories; Loss measurement; Noise level; Physics; Stochastic processes; Stochastic resonance;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
American Control Conference, 2000. Proceedings of the 2000
Conference_Location :
Chicago, IL
ISSN :
0743-1619
Print_ISBN :
0-7803-5519-9
Type :
conf
DOI :
10.1109/ACC.2000.879174
Filename :
879174
Link To Document :
بازگشت