首页    期刊浏览 2025年08月03日 星期日
登录注册

文章基本信息

  • 标题:Survey of Neural Text Representation Models
  • 本地全文:下载
  • 作者:Karlo Babić ; Sanda Martinčić-Ipšić ; Ana Meštrović
  • 期刊名称:Information
  • 电子版ISSN:2078-2489
  • 出版年度:2020
  • 卷号:11
  • 期号:11
  • 页码:511-542
  • DOI:10.3390/info11110511
  • 出版社:MDPI Publishing
  • 摘要:In natural language processing, text needs to be transformed into a machine-readable representation before any processing. The quality of further natural language processing tasks greatly depends on the quality of those representations. In this survey, we systematize and analyze 50 neural models from the last decade. The models described are grouped by the architecture of neural networks as shallow, recurrent, recursive, convolutional, and attention models. Furthermore, we categorize these models by representation level, input level, model type, and model supervision. We focus on task-independent representation models, discuss their advantages and drawbacks, and subsequently identify the promising directions for future neural text representation models. We describe the evaluation datasets and tasks used in the papers that introduced the models and compare the models based on relevant evaluations. The quality of a representation model can be evaluated as its capability to generalize to multiple unrelated tasks. Benchmark standardization is visible amongst recent models and the number of different tasks models are evaluated on is increasing.
  • 关键词:deep learning; embedding; neural language model; neural networks; NLP; text representation deep learning ; embedding ; neural language model ; neural networks ; NLP ; text representation
国家哲学社会科学文献中心版权所有