首页    期刊浏览 2024年10月06日 星期日
登录注册

文章基本信息

  • 标题:Active Learning for Sequence Tagging with Deep Pre-trained Models andBayesian Uncertainty Estimates
  • 本地全文:下载
  • 作者:Artem Shelmanov ; Dmitri Puzyrev ; Lyubov Kupriyanova
  • 期刊名称:Conference on European Chapter of the Association for Computational Linguistics (EACL)
  • 出版年度:2021
  • 卷号:2021
  • 页码:1698-1712
  • DOI:10.18653/v1/2021.eacl-main.145
  • 语种:English
  • 出版社:ACL Anthology
  • 摘要:Annotating training data for sequence tagging of texts is usually very time-consuming. Recent advances in transfer learning for natural language processing in conjunction with active learning open the possibility to significantly reduce the necessary annotation budget. We are the first to thoroughly investigate this powerful combination for the sequence tagging task. We conduct an extensive empirical study of various Bayesian uncertainty estimation methods and Monte Carlo dropout options for deep pre-trained models in the active learning framework and find the best combinations for different types of models. Besides, we also demonstrate that to acquire instances during active learning, a full-size Transformer can be substituted with a distilled version, which yields better computational performance and reduces obstacles for applying deep active learning in practice.
国家哲学社会科学文献中心版权所有