Bai Jionghao
2024
Speech-to-Speech Translation with Discrete-Unit-Based Style Transfer
Yongqi Wang
|
Bai Jionghao
|
Rongjie Huang
|
Ruiqi Li
|
Zhiqing Hong
|
Zhou Zhao
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop)
Direct speech-to-speech translation (S2ST) with discrete self-supervised representations has achieved remarkable accuracy, but is unable to preserve the speaker timbre of the source speech. Meanwhile, the scarcity of high-quality speaker-parallel data poses a challenge for learning style transfer during translation. We design an S2ST pipeline with style-transfer capability on the basis of discrete self-supervised speech representations and codec units. The acoustic language model we introduce for style transfer leverages self-supervised in-context learning, acquiring style transfer ability without relying on any speaker-parallel data, thereby overcoming data scarcity. By using extensive training data, our model achieves zero-shot cross-lingual style transfer on previously unseen source languages. Experiments show that our model generates translated speeches with high fidelity and speaker similarity. Audio samples are available at http://stylelm.github.io/ .
2023
ART: rule bAsed futuRe-inference deducTion
Mengze Li
|
Tianqi Zhao
|
Bai Jionghao
|
Baoyi He
|
Jiaxu Miao
|
Wei Ji
|
Zheqi Lv
|
Zhou Zhao
|
Shengyu Zhang
|
Wenqiao Zhang
|
Fei Wu
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing
Deductive reasoning is a crucial cognitive ability of humanity, allowing us to derive valid conclusions from premises and observations. However, existing works mainly focus on language-based premises and generally neglect deductive reasoning from visual observations. In this work, we introduce rule bAsed futuRe-inference deducTion (ART), which aims at deducing the correct future event based on the visual phenomenon (a video) and the rule-based premises, along with an explanation of the reasoning process. To advance this field, we construct a large-scale densely annotated dataset (Video-ART), where the premises, future event candidates, the reasoning process explanation, and auxiliary commonsense knowledge (e.g., actions and appearance) are annotated by annotators. Upon Video-ART, we develop a strong baseline named ARTNet. In essence, guided by commonsense knowledge, ARTNet learns to identify the target video character and perceives its visual clues related to the future event. Then, ARTNet rigorously applies the given premises to conduct reasoning from the identified information to future events, through a non-parametric rule reasoning network and a reasoning-path review module. Empirical studies validate the rationality of ARTNet in deductive reasoning upon visual observations and the effectiveness over existing works.
Search
Co-authors
- Zhou Zhao 2
- Yongqi Wang 1
- Rongjie Huang 1
- Ruiqi Li 1
- Zhiqing Hong 1
- show all...