TY - UNPB
T1 - Yet another fast variant of Newton's method for nonconvex optimization
AU - Gratton, Serge
AU - Jerad, Sadok
AU - TOINT, Philippe
PY - 2023/2/21
Y1 - 2023/2/21
N2 - A second-order algorithm is proposed for minimizing smooth nonconvexfunctions that alternates between regularized Newton and negativecurvature steps. In most cases, the Hessian matrix is regularized withthe square root of the current gradient and an additional term takingmoderate negative curvature into account, a negative curvature stepbeing taken only exceptionnally. As a consequence, the proposed methodonly requires the solution of a single linear system at nearly alliterations. We establish that at most O(\log(epsilon)|epsilon^{-3/2})evaluations of the problem's objective function and derivatives areneeded for this algorithm to obtain an epsilon-approximate first-orderminimizer, and at most O(|\log(epsilon)| epsilon^{-3}) to obtain asecond-order one. Initial numerical experiments with two variants ofthe new method are finally presented.
AB - A second-order algorithm is proposed for minimizing smooth nonconvexfunctions that alternates between regularized Newton and negativecurvature steps. In most cases, the Hessian matrix is regularized withthe square root of the current gradient and an additional term takingmoderate negative curvature into account, a negative curvature stepbeing taken only exceptionnally. As a consequence, the proposed methodonly requires the solution of a single linear system at nearly alliterations. We establish that at most O(\log(epsilon)|epsilon^{-3/2})evaluations of the problem's objective function and derivatives areneeded for this algorithm to obtain an epsilon-approximate first-orderminimizer, and at most O(|\log(epsilon)| epsilon^{-3}) to obtain asecond-order one. Initial numerical experiments with two variants ofthe new method are finally presented.
KW - Newton's method, nonconvex optimization, negative curvature, adaptive regularization methods, evaluation complexity
M3 - Working paper
VL - 2203.10065
BT - Yet another fast variant of Newton's method for nonconvex optimization
PB - Arxiv
ER -