Processing math: 100%

AdaGrad SGD optimization

adagrad(stepsize = 0.05, epsilon = 1e-08)

Arguments

stepsize

stepsize for SGD

epsilon

epsilon for numerical stability

Value

a list of control variables for optimization (used in control_opt function)

Details

The update rule for AdaGrad is: vt=vt1+g2t xt+1=xtstepsizegtvt+ϵ