Supervised Visual Attention For Simultaneous Multimodal Machine Translation · The Large Language Model Bible Contribute to LLM-Bible

Supervised Visual Attention For Simultaneous Multimodal Machine Translation

Haralampieva Veneta, Caglayan Ozan, Specia Lucia. Journal of Artificial Intelligence Research 2022

[Paper]    
Applications Attention Mechanism Fine Tuning Model Architecture Multimodal Models Pretraining Methods Reinforcement Learning Training Techniques Transformer

Recently, there has been a surge in research in multimodal machine translation (MMT), where additional modalities such as images are used to improve translation quality of textual systems. A particular use for such multimodal systems is the task of simultaneous machine translation, where visual context has been shown to complement the partial information provided by the source sentence, especially in the early phases of translation. In this paper, we propose the first Transformer-based simultaneous MMT architecture, which has not been previously explored in the field. Additionally, we extend this model with an auxiliary supervision signal that guides its visual attention mechanism using labelled phrase-region alignments. We perform comprehensive experiments on three language directions and conduct thorough quantitative and qualitative analyses using both automatic metrics and manual inspection. Our results show that (i) supervised visual attention consistently improves the translation quality of the MMT models, and (ii) fine-tuning the MMT with supervision loss enabled leads to better performance than training the MMT from scratch. Compared to the state-of-the-art, our proposed model achieves improvements of up to 2.3 BLEU and 3.5 METEOR points.

Similar Work