论文标题
具有复杂性的非线性共轭梯度法及其在非convex回归中的应用
A nonlinear conjugate gradient method with complexity guarantees and its application to nonconvex regression
论文作者
论文摘要
非线性共轭梯度是解决连续优化问题的最流行技术之一。尽管这些方案长期以来从全球收敛的角度进行了研究,但它们最差的复杂性属性尚未完全理解,尤其是在非convex环境中。特别是,尚不清楚非线性共轭梯度方法是否比一阶方法(例如梯度下降)拥有更好的保证。同时,最近的实验表明,即使与赋予最佳已知复杂性保证的方法相比,即使在某些非凸问题上,标准非线性共轭梯度技术的表现令人印象深刻。 在本文中,我们提出了一个基于简单的线路搜索范式和修改的重新启动条件的非线性共轭梯度方案。这两种成分允许监视搜索方向的特性,这有助于获得复杂性保证。我们的复杂性结果说明了非线性共轭梯度方法与经典梯度下降之间的可能差异。对非凸的鲁棒回归问题以及标准基准的数值研究表明,重新启动条件可以跟踪标准实现的行为。
Nonlinear conjugate gradients are among the most popular techniques for solving continuous optimization problems. Although these schemes have long been studied from a global convergence standpoint, their worst-case complexity properties have yet to be fully understood, especially in the nonconvex setting. In particular, it is unclear whether nonlinear conjugate gradient methods possess better guarantees than first-order methods such as gradient descent. Meanwhile, recent experiments have shown impressive performance of standard nonlinear conjugate gradient techniques on certain nonconvex problems, even when compared with methods endowed with the best known complexity guarantees. In this paper, we propose a nonlinear conjugate gradient scheme based on a simple line-search paradigm and a modified restart condition. These two ingredients allow for monitoring the properties of the search directions, which is instrumental in obtaining complexity guarantees. Our complexity results illustrate the possible discrepancy between nonlinear conjugate gradient methods and classical gradient descent. A numerical investigation on nonconvex robust regression problems as well as a standard benchmark illustrate that the restarting condition can track the behavior of a standard implementation.