EURO-Online login
- New to EURO? Create an account
- I forgot my username and/or my password.
- Help with cookies
(important for IE8 users)
3271. Accelerated Adaptive Cubic Regularized Quasi-Newton Methods
Invited abstract in session WB-32: Beyond First-Order Optimization Methods, stream Advances in large scale nonlinear optimization.
Wednesday, 10:30-12:00Room: 41 (building: 303A)
Authors (first author is the speaker)
1. | Dmitry Kamzolov
|
Machine Learning, Mohamed bin Zayed University of Artificial Intelligence (MBZUAI) |
Abstract
In this paper, we propose the first Quasi-Newton method with a global convergence rate of $O(\frac{L_1R^2}{k})$ for general convex functions. Quasi-Newton methods, such as BFGS, SR-1, are well-known for their impressive practical performance. However, they are theoretically slower than gradient descent for general convex functions. This gap between impressive practical performance and poor theoretical guarantees was an open question for a long period of time. In this paper, we make a significant step to close this gap. We improve upon the existing rate and propose the Cubic Regularized Quasi-Newton Method with a convergence rate of $O(\frac{L_1R^2}{k})$. The key to achieving this improvement is to use the Cubic Regularized Newton Method over the Damped Newton Method as an outer method, where the Quasi-Newton update is an inexact Hessian approximation. Using this approach, we propose the first Accelerated Quasi-Newton method with a global convergence rate of $O(\frac{L_1R^2}{k^2})$ for general convex functions. In special cases where we have access to additional computations, for example Hessian-vector products, we can improve the inexact Hessian approximation and achieve a global convergence rate of $O(\frac{L_2R^3}{k^{3}})$, which make it intermediate second-order method. To make these methods practical, we introduce the Adaptive Inexact Cubic Regularized Newton Method and its accelerated version, which provide real-time control of the approximation error. We show that the p
Keywords
- Large Scale Optimization
- Continuous Optimization
- Convex Optimization
Status: accepted
Back to the list of papers