Learning language specific sub-network for multilingual machine translation

60Citations
Citations of this article
119Readers
Mendeley users who have this article in their library.

Abstract

Multilingual neural machine translation aims at learning a single translation model for multiple languages. These jointly trained models often suffer from performance degradation on rich-resource language pairs. We attribute this degeneration to parameter interference. In this paper, we propose LaSS to jointly train a single unified multilingual MT model. LaSS learns Language Specific Sub-network (LaSS) for each language pair to counter parameter interference. Comprehensive experiments on IWSLT and WMT datasets with various Transformer architectures show that LaSS obtains gains on 36 language pairs by up to 1.2 BLEU. Besides, LaSS shows its strong generalization performance at easy adaptation to new language pairs and zero-shot translation. LaSS boosts zero-shot translation with an average of 8.3 BLEU on 30 language pairs. Codes and trained models are available at https://github.com/NLP-Playground/LaSS.

References Powered by Scopus

Neural machine translation of rare words with subword units

4457Citations
N/AReaders
Get full text

SentencePiece: A simple and language independent subword tokenizer and detokenizer for neural text processing

2077Citations
N/AReaders
Get full text

Multi-Task learning for multiple language translation

449Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Transformer: A General Framework from Machine Translation to Others

32Citations
N/AReaders
Get full text

Multilingual Machine Translation with Hyper-Adapters

16Citations
N/AReaders
Get full text

Improving Multilingual Neural Machine Translation System for Indic Languages

14Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Lin, Z., Wu, L., Wang, M., & Li, L. (2021). Learning language specific sub-network for multilingual machine translation. In ACL-IJCNLP 2021 - 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Proceedings of the Conference (pp. 293–305). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.acl-long.25

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 30

75%

Researcher 5

13%

Professor / Associate Prof. 3

8%

Lecturer / Post doc 2

5%

Readers' Discipline

Tooltip

Computer Science 40

87%

Linguistics 4

9%

Chemistry 1

2%

Neuroscience 1

2%

Save time finding and organizing research with Mendeley

Sign up for free