4.6 Article

Similarity-aware neural machine translation: reducing human translator efforts by leveraging high-potential sentences with translation memory

Journal

NEURAL COMPUTING & APPLICATIONS
Volume 32, Issue 23, Pages 17623-17635

Publisher

SPRINGER LONDON LTD
DOI: 10.1007/s00521-020-04939-y

Keywords

Neural machine translation; Translation memory; High-potential sentences; Human translator efforts

Funding

  1. Key Technologies Research and Development Program [2018YFC0832100]
  2. National Natural Science Foundation of China [1636203]

Ask authors/readers for more resources

In computer-aided translation tasks, reducing the time of reviewing and post-editing on translations is meaningful for human translators. However, existing studies mainly aim to improve overall translation quality, which only reduces post-editing time. In this work, we firstly identify testing sentences which are highly similar to training set (high-potential sentences) to reduce reviewing time, then we focus on improving corresponding translation quality greatly to reduce post-editing time. From this point, we firstly propose two novel translation memory methods to characterize similarity between sentences on syntactic and template dimensions separately. Based on that, we propose a similarity-aware neural machine translation (similarity-NMT) which consists of two independent modules: (1) Identification Module, which can identify high-potential sentences of testing set according to multi-dimensional similarity information; (2) Translation Module, which can integrate multi-dimensional similarity information of parallel training sentence pairs into an attention-based NMT model by leveraging posterior regularization. Experiments on two Chinese English domains have well-validated the effectiveness and universality of the proposed method of reducing human translator efforts.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available