Songtai Dai
2023
Retrieval-Augmented Domain Adaptation of Language Models
Benfeng Xu
|
Chunxu Zhao
|
Wenbin Jiang
|
PengFei Zhu
|
Songtai Dai
|
Chao Pang
|
Zhuo Sun
|
Shuohuan Wang
|
Yu Sun
Proceedings of the 8th Workshop on Representation Learning for NLP (RepL4NLP 2023)
Language models pretrained on general domain corpora usually exhibit considerable degradation when generalizing to downstream tasks of specialized domains. Existing approaches try to construct PLMs for each specific domains either from scratch or through further pretraining, which not only costs substantial resources, but also fails to cover all target domains at various granularity. In this work, we propose RADA, a novel Retrieval-Augmented framework for Domain Adaptation. We first construct a textual corpora that covers the downstream task at flexible domain granularity and resource availability. We employ it as a pluggable datastore to retrieve informative background knowledge, and integrate them into the standard language model framework to augment representations. We then propose a two-level selection scheme to integrate the most relevant information while alleviating irrelevant noises. Specifically, we introduce a differentiable sampling module as well as an attention mechanism to achieve both passage-level and word-level selection. Such a retrieval-augmented framework enables domain adaptation of language models with flexible domain coverage and fine-grained domain knowledge integration. We conduct comprehensive experiments across biomedical, science and legal domains to demonstrate the effectiveness of the overall framework, and its advantage over existing solutions.
2021
BDKG at MEDIQA 2021: System Report for the Radiology Report Summarization Task
Songtai Dai
|
Quan Wang
|
Yajuan Lyu
|
Yong Zhu
Proceedings of the 20th Workshop on Biomedical Language Processing
This paper presents our winning system at the Radiology Report Summarization track of the MEDIQA 2021 shared task. Radiology report summarization automatically summarizes radiology findings into free-text impressions. This year’s task emphasizes the generalization and transfer ability of participating systems. Our system is built upon a pre-trained Transformer encoder-decoder architecture, i.e., PEGASUS, deployed with an additional domain adaptation module to particularly handle the transfer and generalization issue. Heuristics like ensemble and text normalization are also used. Our system is conceptually simple yet highly effective, achieving a ROUGE-2 score of 0.436 on test set and ranked the 1st place among all participating systems.
Search
Co-authors
- Benfeng Xu 1
- Chunxu Zhao 1
- Wenbin Jiang 1
- Pengfei Zhu 1
- Chao Pang 1
- show all...