首页    期刊浏览 2024年11月28日 星期四
登录注册

文章基本信息

  • 标题:T-NER: An All-Round Python Library for Transformer-based Named Entity Recognition
  • 本地全文:下载
  • 作者:Asahi Ushio ; Jose Camacho-Collados
  • 期刊名称:Conference on European Chapter of the Association for Computational Linguistics (EACL)
  • 出版年度:2021
  • 卷号:2021
  • 页码:53-62
  • DOI:10.18653/v1/2021.eacl-demos.7
  • 语种:English
  • 出版社:ACL Anthology
  • 摘要:Language model (LM) pretraining has led to consistent improvements in many NLP downstream tasks, including named entity recognition (NER). In this paper, we present T-NER (Transformer-based Named Entity Recognition), a Python library for NER LM finetuning. In addition to its practical utility, T-NER facilitates the study and investigation of the cross-domain and cross-lingual generalization ability of LMs finetuned on NER. Our library also provides a web app where users can get model predictions interactively for arbitrary text, which facilitates qualitative model evaluation for non-expert programmers. We show the potential of the library by compiling nine public NER datasets into a unified format and evaluating the cross-domain and cross- lingual performance across the datasets. The results from our initial experiments show that in-domain performance is generally competitive across datasets. However, cross-domain generalization is challenging even with a large pretrained LM, which has nevertheless capacity to learn domain-specific features if fine- tuned on a combined dataset. To facilitate future research, we also release all our LM checkpoints via the Hugging Face model hub.
国家哲学社会科学文献中心版权所有