Pre-training neural machine translation with alignment information via optimal transport.

Autor: Su, Xueping, Zhao, Xingkai, Ren, Jie, Li, Yunhong, Rätsch, Matthias
Předmět:
Zdroj: Multimedia Tools & Applications; May2024, Vol. 83 Issue 16, p48377-48397, 21p
Abstrakt: With the rapid development of globalization, the demand for translation between different languages is also increasing. Although pre-training has achieved excellent results in neural machine translation, the existing neural machine translation has almost no high-quality suitable for specific fields. Alignment information, so this paper proposes a pre-training neural machine translation with alignment information via optimal transport. First, this paper narrows the representation gap between different languages by using OTAP to generate domain-specific data for information alignment, and learns richer semantic information. Secondly, this paper proposes a lightweight model DR-Reformer, which uses Reformer as the backbone network, adds Dropout layers and Reduction layers, reduces model parameters without losing accuracy, and improves computational efficiency. Experiments on the Chinese and English datasets of AI Challenger 2018 and WMT-17 show that the proposed algorithm has better performance than existing algorithms. [ABSTRACT FROM AUTHOR]
Databáze: Complementary Index