Turghun TayirLin LiBei LiJianquan LiuKong Aik Lee
The main purpose of multimodal machine translation is to improve the quality of translation results by taking the corresponding visual context as an additional input. Recently many studies in neural machine translation have attempted to obtain high-quality multimodal representation of encoder or decoder via attention mechanism. However, attention mechanism does not always accurately identify the decisive input for each prediction, which leads to an unsatisfactory multimodal information fusion. To this end, we propose an encoder-decoder calibration method which can automatically calibrate the image and text fusion representation in the encoder, and find the decisive input to the translation in the decoder. We validate our model on the multimodal machine translation dataset Multi30K. Experimental results show that our method significantly outperforms several recent baselines for both English–German and English–French translation tasks in terms of BLEU and METEOR.
Yingbo GaoChristian HeroldZijian YangHermann Ney
Peng‐Jen ChenBowen ShiKelvin NiuAnn LeeWei-Ning Hsu
Jungo KasaiNikolaos PappasHao PengJames CrossNoah A. Smith