首页    期刊浏览 2024年10月05日 星期六
登录注册

文章基本信息

  • 标题:A mean field view of the landscape of two-layer neural networks
  • 本地全文:下载
  • 作者:Song Mei ; Andrea Montanari ; Phan-Minh Nguyen
  • 期刊名称:Proceedings of the National Academy of Sciences
  • 印刷版ISSN:0027-8424
  • 电子版ISSN:1091-6490
  • 出版年度:2018
  • 卷号:115
  • 期号:33
  • 页码:E7665-E7671
  • DOI:10.1073/pnas.1806579115
  • 语种:English
  • 出版社:The National Academy of Sciences of the United States of America
  • 摘要:Multilayer neural networks are among the most powerful models in machine learning, yet the fundamental reasons for this success defy mathematical understanding. Learning a neural network requires optimizing a nonconvex high-dimensional objective (risk function), a problem that is usually attacked using stochastic gradient descent (SGD). Does SGD converge to a global optimum of the risk or only to a local optimum? In the former case, does this happen because local minima are absent or because SGD somehow avoids them? In the latter, why do local minima reached by SGD have good generalization properties? In this paper, we consider a simple case, namely two-layer neural networks, and prove that—in a suitable scaling limit—SGD dynamics is captured by a certain nonlinear partial differential equation (PDE) that we call distributional dynamics (DD). We then consider several specific examples and show how DD can be used to prove convergence of SGD to networks with nearly ideal generalization error. This description allows for “averaging out” some of the complexities of the landscape of neural networks and can be used to prove a general convergence result for noisy SGD.
  • 关键词:neural networks ; stochastic gradient descent ; gradient flow ; Wasserstein space ; partial differential equations
国家哲学社会科学文献中心版权所有