Article ID | Journal | Published Year | Pages | File Type |
---|---|---|---|---|
4973671 | Computer Speech & Language | 2017 | 19 Pages |
Abstract
We evaluated our models on the standard task of translating English-to-German and Arabic-to-English TED talks. The NDAM models achieved better perplexities and modest BLEU improvements compared to the baseline NNJM, trained either on in-domain or on a concatenation of in- and out-domain data. On the other hand, the NFM models obtained significant improvements of up to +0.9 and +0.7 BLEU points, respectively. We also demonstrate improvements over existing adaptation methods such as instance weighting, phrasetable fill-up, linear and log-linear interpolations.
Keywords
Related Topics
Physical Sciences and Engineering
Computer Science
Signal Processing
Authors
Shafiq Joty, Nadir Durrani, Hassan Sajjad, Ahmed Abdelali,