@inproceedings{khayrallah-sedoc-2020-smrt,
title = "{SMRT} Chatbots: {I}mproving Non-Task-Oriented Dialog with {S}imulated {M}ultiple {R}eference {T}raining",
author = "Khayrallah, Huda and
Sedoc, Jo{\~a}o",
editor = "Cohn, Trevor and
He, Yulan and
Liu, Yang",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2020",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2020.findings-emnlp.403/",
doi = "10.18653/v1/2020.findings-emnlp.403",
pages = "4489--4505",
abstract = "Non-task-oriented dialog models suffer from poor quality and non-diverse responses. To overcome limited conversational data, we apply Simulated Multiple Reference Training (SMRT; Khayrallah et al., 2020), and use a paraphraser to simulate multiple responses per training prompt. We find SMRT improves over a strong Transformer baseline as measured by human and automatic quality scores and lexical diversity. We also find SMRT is comparable to pretraining in human evaluation quality, and outperforms pretraining on automatic quality and lexical diversity, without requiring related-domain dialog data."
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="khayrallah-sedoc-2020-smrt">
<titleInfo>
<title>SMRT Chatbots: Improving Non-Task-Oriented Dialog with Simulated Multiple Reference Training</title>
</titleInfo>
<name type="personal">
<namePart type="given">Huda</namePart>
<namePart type="family">Khayrallah</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">João</namePart>
<namePart type="family">Sedoc</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2020-11</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Findings of the Association for Computational Linguistics: EMNLP 2020</title>
</titleInfo>
<name type="personal">
<namePart type="given">Trevor</namePart>
<namePart type="family">Cohn</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Yulan</namePart>
<namePart type="family">He</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Yang</namePart>
<namePart type="family">Liu</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Online</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Non-task-oriented dialog models suffer from poor quality and non-diverse responses. To overcome limited conversational data, we apply Simulated Multiple Reference Training (SMRT; Khayrallah et al., 2020), and use a paraphraser to simulate multiple responses per training prompt. We find SMRT improves over a strong Transformer baseline as measured by human and automatic quality scores and lexical diversity. We also find SMRT is comparable to pretraining in human evaluation quality, and outperforms pretraining on automatic quality and lexical diversity, without requiring related-domain dialog data.</abstract>
<identifier type="citekey">khayrallah-sedoc-2020-smrt</identifier>
<identifier type="doi">10.18653/v1/2020.findings-emnlp.403</identifier>
<location>
<url>https://aclanthology.org/2020.findings-emnlp.403/</url>
</location>
<part>
<date>2020-11</date>
<extent unit="page">
<start>4489</start>
<end>4505</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T SMRT Chatbots: Improving Non-Task-Oriented Dialog with Simulated Multiple Reference Training
%A Khayrallah, Huda
%A Sedoc, João
%Y Cohn, Trevor
%Y He, Yulan
%Y Liu, Yang
%S Findings of the Association for Computational Linguistics: EMNLP 2020
%D 2020
%8 November
%I Association for Computational Linguistics
%C Online
%F khayrallah-sedoc-2020-smrt
%X Non-task-oriented dialog models suffer from poor quality and non-diverse responses. To overcome limited conversational data, we apply Simulated Multiple Reference Training (SMRT; Khayrallah et al., 2020), and use a paraphraser to simulate multiple responses per training prompt. We find SMRT improves over a strong Transformer baseline as measured by human and automatic quality scores and lexical diversity. We also find SMRT is comparable to pretraining in human evaluation quality, and outperforms pretraining on automatic quality and lexical diversity, without requiring related-domain dialog data.
%R 10.18653/v1/2020.findings-emnlp.403
%U https://aclanthology.org/2020.findings-emnlp.403/
%U https://doi.org/10.18653/v1/2020.findings-emnlp.403
%P 4489-4505
Markdown (Informal)
[SMRT Chatbots: Improving Non-Task-Oriented Dialog with Simulated Multiple Reference Training](https://aclanthology.org/2020.findings-emnlp.403/) (Khayrallah & Sedoc, Findings 2020)
ACL