首页    期刊浏览 2024年12月04日 星期三
登录注册

文章基本信息

  • 标题:Stochastic optimization with momentum: Convergence, fluctuations, and traps avoidance
  • 本地全文:下载
  • 作者:Anas Barakat ; Pascal Bianchi ; Walid Hachem
  • 期刊名称:Electronic Journal of Statistics
  • 印刷版ISSN:1935-7524
  • 出版年度:2021
  • 卷号:15
  • 期号:2
  • 页码:3892-3947
  • DOI:10.1214/21-EJS1880
  • 语种:English
  • 出版社:Institute of Mathematical Statistics
  • 摘要:In this paper, a general stochastic optimization procedure is studied, unifying several variants of the stochastic gradient descent such as, among others, the stochastic heavy ball method, the Stochastic Nesterov Accelerated Gradient algorithm (S-NAG), and the widely used Adam algorithm. The algorithm is seen as a noisy Euler discretization of a non-autonomous ordinary differential equation, recently introduced by Belotto da Silva and Gazeau, which is analyzed in depth. Assuming that the objective function is non-convex and differentiable, the stability and the almost sure convergence of the iterates to the set of critical points are established. A noteworthy special case is the convergence proof of S-NAG in a non-convex setting. Under some assumptions, the convergence rate is provided under the form of a Central Limit Theorem. Finally, the non-convergence of the algorithm to undesired critical points, such as local maxima or saddle points, is established. Here, the main ingredient is a new avoidance of traps result for non-autonomous settings, which is of independent interest.
  • 关键词:34A12; 60F99; 62L20; 68T99; ADAM; adaptive gradient methods with momentum; avoidance of traps; dynamical systems; Nesterov accelerated gradient; stochastic approximation
国家哲学社会科学文献中心版权所有