Grunnleggende konsepter
Efficient algorithm for nonconvex minimization with inexact evaluations.
Sammendrag
The content introduces a randomized algorithm for nonconvex minimization with inexact oracle access to gradient and Hessian. It focuses on achieving approximate second-order optimality without requiring access to the function value. The algorithm incorporates Rademacher randomness in negative curvature steps and allows for gradient and Hessian inexactness. The convergence analysis includes both expectation and high-probability bounds. The algorithm's complexity results show improved gradient sample complexity for empirical risk minimization problems.
- Introduction
- Seeks local minimizer of a smooth nonconvex function.
- Defines approximate second-order point optimality conditions.
- Data Extraction
- "Our complexity results include both expected and high-probability stopping times for the algorithm."
- Prior Work
- Discusses approximate second-order points in nonconvex functions.
- Compares iteration and operation complexity of different algorithms.
- Inexact Derivatives
- Examines settings with inexact gradient and Hessian oracles.
- Reviews stochastic and general inexact settings.
- Notation
- Defines Lipschitz continuity and key mathematical notations.
- Algorithm and Assumptions
- Defines the algorithm and assumptions on function, gradients, and Hessians.
- Describes the step types and convergence analysis.
- High-Probability Bound
- States and proves the main result on the number of iterations for algorithm termination.
- Discusses various choices of parameters and their impact on iteration complexity.
Statistikk
"Our complexity results include both expected and high-probability stopping times for the algorithm."
Sitater
"A distinctive feature of our method is that it 'flips a coin' to decide whether to move in a positive or negative sense along a direction of negative curvature for an approximate Hessian."