Approximation of Hessian Matrix for Second-order SPSA Algorithm ...
Approximation of Hessian Matrix for Second-order SPSA Algorithm ...
Approximation of Hessian Matrix for Second-order SPSA Algorithm ...
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
2.2 THE <strong>SPSA</strong> ALGORITHM RECURSIONS<br />
1st-<strong>SPSA</strong> [17]:<br />
θ ˆ = θˆ<br />
− a gˆ<br />
( θˆ<br />
), 0,1,2,...<br />
(2.1)<br />
k + 1 k k k k<br />
k =<br />
2nd-<strong>SPSA</strong> [18]:<br />
ˆ ˆ<br />
−1<br />
θ = θ − a H gˆ<br />
( ˆ ), H = f ( H )<br />
(2.2 a)<br />
k + 1 k k k k<br />
θ<br />
k k k k<br />
= k<br />
1<br />
H H<br />
ˆ<br />
1<br />
+ H , = 0,1,2,...<br />
k + 1<br />
− k + 1<br />
k<br />
(2.2 b)<br />
k k<br />
k<br />
where<br />
a<br />
k and a<br />
k are the scalar gain series that satisfy certain SA conditions [18], ĝ<br />
k<br />
is<br />
the SP estimate <strong>of</strong> the loss function gradient that depends on the gain sequence<br />
c<br />
k<br />
(representing a difference interval <strong>of</strong> the perturbations),<br />
Hˆ<br />
k<br />
is the SP estimate <strong>of</strong> the <strong>Hessian</strong><br />
matrix, and<br />
f<br />
k maps the usual non-positive-definite H<br />
k<br />
to a positive-definite pxp matrix.<br />
The two recursions are showed in Fig. 2.1. Let<br />
∆<br />
k be a user-generated mean zero random<br />
vector <strong>of</strong> dimension p with its components being independent random variables.<br />
Fig. 2.1. The two-recursions in 2nd-<strong>SPSA</strong> algorithm<br />
(solid-line eq. 2.2 a, dashed-line eq. 2.2 b).<br />
The i-th element <strong>of</strong> the loss function gradient is given by [18].<br />
( gˆ<br />
) = (2c<br />
∆<br />
k<br />
i<br />
k<br />
ki<br />
−1<br />
) [ y( ˆ θ + c ∆ ) − y( ˆ θ −c<br />
∆ )], i=1, 2, … , p (2.3)<br />
k<br />
k<br />
k<br />
k<br />
k<br />
k<br />
21