Activity Number:
|
143
|
Type:
|
Topic Contributed
|
Date/Time:
|
Monday, August 12, 2002 : 2:00 PM to 3:50 PM
|
Sponsor:
|
Section on Statistical Computing*
|
Abstract - #301530 |
Title:
|
Global Random Optimization by Simultaneous Perturbation Stochastic Approximation
|
Author(s):
|
John Maryak*+ and Daniel Chin
|
Affiliation(s):
|
Johns Hopkins University and Johns Hopkins University
|
Address:
|
11100 Johns Hopkins Road, Laurel, Maryland, 20723-6099, USA
|
Keywords:
|
Stochastic optimization ; Global convergence ; SPSA ; Stochastic approximation ; Recursive annealing
|
Abstract:
|
A desire with iterative optimization techniques is that the algorithm reach the global optimum rather than get stranded at a local optimum value. Here, we examine the global convergence properties of a "gradient free" stochastic approximation algorithm called "SPSA," that has performed well in complex optimization problems. We establish two theorems on the global convergence of SPSA. The first provides conditions under which SPSA will converge in probability to a global optimum using the well-known method of injected noise. In the second theorem, we show that, under different conditions, "basic" SPSA without injected noise can achieve convergence in probability to a global optimum. This latter result can have important benefits in the setup (tuning) and performance of the algorithm. The discussion is supported by numerical studies showing favorable comparisons of SPSA to simulated annealing and genetic algorithms.
|