4.5 Article

Dual Knowledge Distillation for neural machine translation

Journal

COMPUTER SPEECH AND LANGUAGE
Volume 84, Issue -, Pages -

Publisher

ACADEMIC PRESS LTD- ELSEVIER SCIENCE LTD
DOI: 10.1016/j.csl.2023.101583

Keywords

Knowledge distillation; k Nearest Neighbor Knowledge Distillation; Low-resource; Monolingual data

Ask authors/readers for more resources

In this paper, a new knowledge distillation method called Dual Knowledge Distillation (DKD) is proposed to better utilize monolingual and limited bilingual data. By combining self-distillation and consistency regularization strategies, significant improvements are achieved in extracting consistent monolingual representation and forcing the decoder to produce consistent output.
Existing knowledge distillation methods use large amount of bilingual data and focus on mining the corresponding knowledge distribution between the source language and the target language. However, for some languages, bilingual data is not abundant. In this paper, to make better use of both monolingual and limited bilingual data, we propose a new knowledge distillation method called Dual Knowledge Distillation (DKD). For monolingual data, we use a self-distillation strategy which combines self-training and knowledge distillation for the encoder to extract more consistent monolingual representation. For bilingual data, on top of the k Nearest Neighbor Knowledge Distillation (kNN-KD) method, a similar self-distillation strategy is adopted as a consistency regularization method to force the decoder to produce consistent output. Experiments on standard datasets, multi-domain translation datasets, and low-resource datasets show that DKD achieves consistent improvements over state-of-the-art baselines including kNN-KD.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available