Language Tags Matter for Zero-Shot Neural Machine Translation
Autor: | Shanbo Cheng, Liwei Wu, Lei Li, Mingxuan Wang |
---|---|
Rok vydání: | 2021 |
Předmět: |
FOS: Computer and information sciences
Computer Science - Computation and Language Machine translation business.industry Computer science Shot (filmmaking) computer.software_genre Translation (geometry) Zero (linguistics) Consistency (database systems) Artificial intelligence business Encoder computer Computation and Language (cs.CL) Natural language processing |
Zdroj: | ACL/IJCNLP (Findings) |
DOI: | 10.48550/arxiv.2106.07930 |
Popis: | Multilingual Neural Machine Translation (MNMT) has aroused widespread interest due to its efficiency. An exciting advantage of MNMT models is that they could also translate between unsupervised (zero-shot) language directions. Language tag (LT) strategies are often adopted to indicate the translation directions in MNMT. In this paper, we demonstrate that the LTs are not only indicators for translation directions but also crucial to zero-shot translation qualities. Unfortunately, previous work tends to ignore the importance of LT strategies. We demonstrate that a proper LT strategy could enhance the consistency of semantic representations and alleviate the off-target issue in zero-shot directions. Experimental results show that by ignoring the source language tag (SLT) and adding the target language tag (TLT) to the encoder, the zero-shot translations could achieve a +8 BLEU score difference over other LT strategies in IWSLT17, Europarl, TED talks translation tasks. Comment: 7 pages, 3 figures, Accepted by the Findings of ACL2021 |
Databáze: | OpenAIRE |
Externí odkaz: |