首页    期刊浏览 2024年10月07日 星期一
登录注册

文章基本信息

  • 标题:Benchmarking Transformer-based Language Models forArabic Sentiment and Sarcasm Detection
  • 本地全文:下载
  • 作者:Ibrahim Abu Farha ; Walid Magdy
  • 期刊名称:Conference on European Chapter of the Association for Computational Linguistics (EACL)
  • 出版年度:2021
  • 卷号:2021
  • 页码:21-31
  • 语种:English
  • 出版社:ACL Anthology
  • 摘要:The introduction of transformer-based language models has been a revolutionary step for natural language processing (NLP) research. These models, such as BERT, GPT and ELECTRA, led to state-of-the-art performance in many NLP tasks. Most of these models were initially developed for English and other languages followed later. Recently, several Arabic-specific models started emerging. However, there are limited direct comparisons between these models. In this paper, we evaluate the performance of 24 of these models on Arabic sentiment and sarcasm detection. Our results show that the models achieving the best performance are those that are trained on only Arabic data, including dialectal Arabic, and use a larger number of parameters, such as the recently released MARBERT. However, we noticed that AraELECTRA is one of the top performing models while being much more efficient in its computational cost. Finally, the experiments on AraGPT2 variants showed low performance compared to BERT models, which indicates that it might not be suitable for classification tasks.
国家哲学社会科学文献中心版权所有