Restarted Nonconvex Accelerated Gradient Descent: No More Polylogarithmic Factor in the O(-7/4) Complexity

JOURNAL OF MACHINE LEARNING RESEARCH(2023)

引用 0|浏览1
暂无评分
摘要
This paper studies accelerated gradient methods for nonconvex optimization with Lipschitz continuous gradient and Hessian. We propose two simple accelerated gradient methods, restarted accelerated gradient descent (AGD) and restarted heavy ball (HB) method, and establish that our methods achieve an is an element of-approximate first-order stationary point within O(is an element of(-7/4)) number of gradient evaluations by elementary proofs. Theoretically, our complexity does not hide any polylogarithmic factors, and thus it improves over the best known one by the O(log1/& varepsilon;) factor. Our algorithms are simple in the sense that they only consist of Nesterov's classical AGD or Polyak's HB iterations, as well as a restart mechanism. They do not invoke negative curvature exploitation or minimization of regularized surrogate functions as the subroutines. In contrast with existing analysis, our elementary proofs use less advanced techniques and do not invoke the analysis of strongly convex AGD or HB.
更多
查看译文
关键词
accelerated gradient descent,heavy ball method,restart,nonconvex optimization,first-order stationary point
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要