Word-based Domain Adaptation for Neural Machine Translation

06/07/2019
by   Shen Yan, et al.
0

In this paper, we empirically investigate applying word-level weights to adapt neural machine translation to e-commerce domains, where small e-commerce datasets and large out-of-domain datasets are available. In order to mine in-domain like words in the out-of-domain datasets, we compute word weights by using a domain-specific and a non-domain-specific language model followed by smoothing and binary quantization. The baseline model is trained on mixed in-domain and out-of-domain datasets. Experimental results on English to Chinese e-commerce domain translation show that compared to continuing training without word weights, it improves MT quality by up to 2.11 1.59 Pre-training a model with word weights improves fine-tuning up to 1.24 absolute and 1.64

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset