Visual Pivoting Unsupervised Multimodal Machine Translation in Low-Resource Distant Language Pairs

Turghun Tayir, Lin Li, Xiaohui Tao, Mieradilijiang Maimaiti, Ming Li, Jianquan Liu


Abstract
Unsupervised multimodal machine translation (UMMT) aims to leverage vision information as a pivot between two languages to achieve better performance on low-resource language pairs. However, there is presently a challenge: how to handle alignment between distant language pairs (DLPs) in UMMT. To this end, this paper proposes a visual pivoting UMMT method for DLPs. Specifically, we first construct a dataset containing two DLPs, including English-Uyghur and Chinese-Uyghur. We then apply the visual pivoting method for both to pre-training and fine-tuning, and we observe that the images on the encoder and decoder of UMMT have noticeable effects on DLPs. Finally, we introduce informative multi-granularity image features to facilitate further alignment of the latent space between the two languages. Experimental results show that the proposed method significantly outperforms several baselines on DLPs and close language pairs (CLPs).
Anthology ID:
2024.findings-emnlp.320
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2024
Month:
November
Year:
2024
Address:
Miami, Florida, USA
Editors:
Yaser Al-Onaizan, Mohit Bansal, Yun-Nung Chen
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
5596–5607
Language:
URL:
https://aclanthology.org/2024.findings-emnlp.320/
DOI:
10.18653/v1/2024.findings-emnlp.320
Bibkey:
Cite (ACL):
Turghun Tayir, Lin Li, Xiaohui Tao, Mieradilijiang Maimaiti, Ming Li, and Jianquan Liu. 2024. Visual Pivoting Unsupervised Multimodal Machine Translation in Low-Resource Distant Language Pairs. In Findings of the Association for Computational Linguistics: EMNLP 2024, pages 5596–5607, Miami, Florida, USA. Association for Computational Linguistics.
Cite (Informal):
Visual Pivoting Unsupervised Multimodal Machine Translation in Low-Resource Distant Language Pairs (Tayir et al., Findings 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.findings-emnlp.320.pdf
Data:
 2024.findings-emnlp.320.data.zip