首页    期刊浏览 2025年02月18日 星期二
登录注册

文章基本信息

  • 标题:A mathematical theory of semantic development in deep neural networks
  • 本地全文:下载
  • 作者:Andrew M. Saxe ; Andrew M. Saxe ; James L. McClelland
  • 期刊名称:Proceedings of the National Academy of Sciences
  • 印刷版ISSN:0027-8424
  • 电子版ISSN:1091-6490
  • 出版年度:2019
  • 卷号:116
  • 期号:23
  • 页码:11537-11546
  • DOI:10.1073/pnas.1820226116
  • 出版社:The National Academy of Sciences of the United States of America
  • 摘要:An extensive body of empirical research has revealed remarkable regularities in the acquisition, organization, deployment, and neural representation of human semantic knowledge, thereby raising a fundamental conceptual question: What are the theoretical principles governing the ability of neural networks to acquire, organize, and deploy knowledge by integrating across many individual experiences? We address this question by mathematically analyzing the nonlinear dynamics of learning in deep linear networks. We find exact solutions to this learning dynamics that yield a conceptual explanation for the prevalence of many disparate phenomena in semantic cognition, including the hierarchical differentiation of concepts through rapid developmental transitions, the ubiquity of semantic illusions between such transitions, the emergence of item typicality and category coherence as factors controlling the speed of semantic processing, changing patterns of inductive projection over development, and the conservation of semantic similarity in neural representations across species. Thus, surprisingly, our simple neural model qualitatively recapitulates many diverse regularities underlying semantic development, while providing analytic insight into how the statistical structure of an environment can interact with nonlinear deep-learning dynamics to give rise to these regularities.
  • 关键词:semantic cognition ; deep learning ; neural networks ; generative models
国家哲学社会科学文献中心版权所有