Zefeng Du
2024
Benchmarking and Improving Long-Text Translation with Large Language Models
Longyue Wang
|
Zefeng Du
|
Wenxiang Jiao
|
Chenyang Lyu
|
Jianhui Pang
|
Leyang Cui
|
Kaiqiang Song
|
Derek Wong
|
Shuming Shi
|
Zhaopeng Tu
Findings of the Association for Computational Linguistics: ACL 2024
Recent studies have illuminated the promising capabilities of large language models (LLMs) in handling long texts. However, their performance in machine translation (MT) of long documents remains underexplored. This paper aims to shed light on how LLMs navigate this complex task, offering a comprehensive evaluation of their capabilities and limitations in long-text MT. First, we collect and construct an instruction-based benchmark dataset, specifically designed for the finetuning and evaluation of LLMs, encompassing multilingual, multi-domain, and document-level parallel data. Second, we conduct a comprehensive comparison between MT and LLM models concerning document-level translation. Our analysis uncovers that LLMs exhibit shortcomings in long-text domains, and their performance diminishes as document size escalates. By exploiting various extrapolation strategies, we enhance the capacity of LLMs to translate longer texts. We release data, code, and models at https://github.com/longyuewangdcu/Document-MT-LLM.
A Paradigm Shift: The Future of Machine Translation Lies with Large Language Models
Chenyang Lyu
|
Zefeng Du
|
Jitao Xu
|
Yitao Duan
|
Minghao Wu
|
Teresa Lynn
|
Alham Fikri Aji
|
Derek F. Wong
|
Longyue Wang
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
Machine Translation (MT) has greatly advanced over the years due to the developments in deep neural networks. However, the emergence of Large Language Models (LLMs) like GPT-4 and ChatGPT is introducing a new phase in the MT domain. In this context, we believe that the future of MT is intricately tied to the capabilities of LLMs. These models not only offer vast linguistic understandings but also bring innovative methodologies, such as prompt-based techniques, that have the potential to further elevate MT. In this paper, we provide an overview of the significant enhancements in MT that are influenced by LLMs and advocate for their pivotal role in upcoming MT research and implementations. We highlight several new MT directions, emphasizing the benefits of LLMs in scenarios such as Long-Document Translation, Stylized Translation, and Interactive Translation. Additionally, we address the important concern of privacy in LLM-driven MT and suggest essential privacy-preserving strategies. By showcasing practical instances, we aim to demonstrate the advantages that LLMs offer, particularly in tasks like translating extended documents. We conclude by emphasizing the critical role of LLMs in guiding the future evolution of MT and offer a roadmap for future exploration in the sector.
Search
Co-authors
- Longyue Wang 2
- Chenyang Lyu 2
- Wenxiang Jiao 1
- Jianhui Pang 1
- Leyang Cui 1
- show all...