Integrating Domain Terminology into Neural Machine Translation

25Citations
Citations of this article
69Readers
Mendeley users who have this article in their library.

Abstract

This paper extends existing work on terminology integration into Neural Machine Translation, a common industrial practice to dynamically adapt translation to a specific domain. Our method, based on the use of placeholders complemented with morphosyntactic annotation, efficiently taps into the ability of the neural network to deal with symbolic knowledge to surpass the surface generalization shown by alternative techniques. We compare our approach to state-of-the-art systems and benchmark them through a well-defined evaluation framework, focusing on actual application of terminology and not just on the overall performance. Results indicate the suitability of our method in the use-case where terminology is used in a system trained on generic data only.

References Powered by Scopus

Neural machine translation of rare words with subword units

4499Citations
N/AReaders
Get full text

OpenNMT: Open-source toolkit for neural machine translation

1016Citations
N/AReaders
Get full text

Six Challenges for Neural Machine Translation

695Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Domain Adaptation and Multi-Domain Adaptation for Neural Machine Translation: A Survey

44Citations
N/AReaders
Get full text

A short guide to post-editing

28Citations
N/AReaders
Get full text

General and Domain-adaptive Chinese Spelling Check with Error-consistent Pretraining

15Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Michon, E., Crego, J., & Senellart, J. (2020). Integrating Domain Terminology into Neural Machine Translation. In COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference (pp. 3925–3937). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.coling-main.348

Readers over time

‘20‘21‘22‘23‘24‘2506121824

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 18

69%

Researcher 4

15%

Lecturer / Post doc 3

12%

Professor / Associate Prof. 1

4%

Readers' Discipline

Tooltip

Computer Science 20

67%

Linguistics 8

27%

Neuroscience 1

3%

Social Sciences 1

3%

Save time finding and organizing research with Mendeley

Sign up for free
0