Mingyang Zhou, Runxiang Cheng, Yong Jae Lee, Zhou Yu
We introduce a novel multimodal machine translation model that utilizes parallel visual and textual information. Our model jointly optimizes the learning of a shared visual-language embedding and a translator. The model leverages a visual attention grounding mechanism that links the visual semantics with the corresponding textual semantics. Our approach achieves competitive state-of-the-art results on the Multi30K and the Ambiguous COCO datasets. We also collected a new multilingual multimodal product description dataset to simulate a real-world international online shopping scenario. On this dataset, our visual attention grounding model outperforms other methods by a large margin.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Machine Translation | Multi30K | BLEU (EN-DE) | 31.6 | VAG-NMT |
| Machine Translation | Multi30K | Meteor (EN-DE) | 52.2 | VAG-NMT |
| Machine Translation | Multi30K | Meteor (EN-FR) | 70.3 | VAG-NMT |
| Multimodal Machine Translation | Multi30K | BLEU (EN-DE) | 31.6 | VAG-NMT |
| Multimodal Machine Translation | Multi30K | Meteor (EN-DE) | 52.2 | VAG-NMT |
| Multimodal Machine Translation | Multi30K | Meteor (EN-FR) | 70.3 | VAG-NMT |