ARBERT & MARBERT: Deep bidirectional transformers for Arabic

240Citations
Citations of this article
283Readers
Mendeley users who have this article in their library.

Abstract

Pre-trained language models (LMs) are currently integral to many natural language processing systems. Although multilingual LMs were also introduced to serve many languages, these have limitations such as being costly at inference time and the size and diversity of non-English data involved in their pre-training. We remedy these issues for a collection of diverse Arabic varieties by introducing two powerful deep bidirectional transformer-based models, ARBERT and MARBERT. To evaluate our models, we also introduce ARLUE, a new benchmark for multi-dialectal Arabic language understanding evaluation. ARLUE is built using 42 datasets targeting six different task clusters, allowing us to offer a series of standardized experiments under rich conditions. When fine-tuned on ARLUE, our models collectively achieve new state-of-the-art results across the majority of tasks (37 out of 48 classification tasks, on the 42 datasets). Our best model acquires the highest ARLUE score (77.40) across all six task clusters, outperforming all other models including XLM-RLarge (~ 3.4× larger size). Our models are publicly available at https://github.com/UBCNLP/marbert and ARLUE will be released through the same repository.

References Powered by Scopus

SQuad: 100,000+ questions for machine comprehension of text

4031Citations
N/AReaders
Get full text

A primer in bertology: What we know about how bert works

899Citations
N/AReaders
Get full text

Japanese and Korean voice search

671Citations
N/AReaders
Get full text

Cited by Powered by Scopus

BERT Models for Arabic Text Classification: A Systematic Review

93Citations
N/AReaders
Get full text

AraT5: Text-to-Text Transformers for Arabic Language Generation

85Citations
N/AReaders
Get full text

Arabic offensive and hate speech detection using a cross-corpora multi-task learning model

63Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Abdul-Mageed, M., Elmadany, A. R., & Nagoudi, E. M. B. (2021). ARBERT & MARBERT: Deep bidirectional transformers for Arabic. In ACL-IJCNLP 2021 - 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Proceedings of the Conference (pp. 7088–7105). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.acl-long.551

Readers over time

‘21‘22‘23‘24‘250255075100

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 84

72%

Lecturer / Post doc 13

11%

Researcher 10

9%

Professor / Associate Prof. 9

8%

Readers' Discipline

Tooltip

Computer Science 107

88%

Engineering 6

5%

Linguistics 5

4%

Mathematics 3

2%

Save time finding and organizing research with Mendeley

Sign up for free
0