首页    期刊浏览 2024年11月29日 星期五
登录注册

文章基本信息

  • 标题:T2NER: Transformers based Transfer Learning Framework for Named Entity Recognition
  • 本地全文:下载
  • 作者:Saadullah Amin ; Guenter Neumann
  • 期刊名称:Conference on European Chapter of the Association for Computational Linguistics (EACL)
  • 出版年度:2021
  • 卷号:2021
  • 页码:212-220
  • DOI:10.18653/v1/2021.eacl-demos.25
  • 语种:English
  • 出版社:ACL Anthology
  • 摘要:Recent advances in deep transformer models have achieved state-of-the-art in several natural language processing (NLP) tasks, whereas named entity recognition (NER) has traditionally benefited from long-short term memory (LSTM) networks. In this work, we present a Transformers based Transfer Learning framework for Named Entity Recognition (T2NER) created in PyTorch for the task of NER with deep transformer models. The framework is built upon the Transformers library as the core modeling engine and supports several transfer learning scenarios from sequential transfer to domain adaptation, multi-task learning, and semi-supervised learning. It aims to bridge the gap between the algorithmic advances in these areas by combining them with the state-of-the-art in transformer models to provide a unified platform that is readily extensible and can be used for both the transfer learning research in NER, and for real-world applications. The framework is available at: https://github.com/suamin/t2ner.
国家哲学社会科学文献中心版权所有