Improving the Lexical Ability of Pretrained Language Models for Unsupervised Neural Machine Translation

16Citations
Citations of this article
69Readers
Mendeley users who have this article in their library.

Abstract

Successful methods for unsupervised neural machine translation (UNMT) employ cross-lingual pretraining via self-supervision, often in the form of a masked language modeling or a sequence generation task, which requires the model to align the lexical- and high-level representations of the two languages. While cross-lingual pretraining works for similar languages with abundant corpora, it performs poorly in low-resource and distant languages. Previous research has shown that this is because the representations are not sufficiently aligned. In this paper, we enhance the bilingual masked language model pretraining with lexical-level information by using type-level cross-lingual subword embeddings. Empirical results demonstrate improved performance both on UNMT (up to 4.5 BLEU) and bilingual lexicon induction using our method compared to a UNMT baseline.

References Powered by Scopus

Extracting and composing robust features with denoising autoencoders

6008Citations
N/AReaders
Get full text

Neural machine translation of rare words with subword units

4457Citations
N/AReaders
Get full text

A Call for Clarity in Reporting BLEU Scores

1978Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Survey of Low-Resource Machine Translation

104Citations
N/AReaders
Get full text

Improving Word Translation via Two-Stage Contrastive Learning

20Citations
N/AReaders
Get full text

Investigating Unsupervised Neural Machine Translation for Low-resource Language Pair English-Mizo via Lexically Enhanced Pre-Trained Language Models

5Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Chronopoulou, A., Stojanovski, D., & Fraser, A. (2021). Improving the Lexical Ability of Pretrained Language Models for Unsupervised Neural Machine Translation. In NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 173–180). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.naacl-main.16

Readers over time

‘21‘22‘23‘24‘2508162432

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 17

63%

Researcher 6

22%

Professor / Associate Prof. 2

7%

Lecturer / Post doc 2

7%

Readers' Discipline

Tooltip

Computer Science 22

73%

Linguistics 4

13%

Neuroscience 2

7%

Engineering 2

7%

Save time finding and organizing research with Mendeley

Sign up for free
0