@inproceedings{tam-etal-2023-annotating,
title = "Annotating Situated Actions in Dialogue",
author = "Tam, Christopher and
Brutti, Richard and
Lai, Kenneth and
Pustejovsky, James",
editor = "Bonn, Julia and
Xue, Nianwen",
booktitle = "Proceedings of the Fourth International Workshop on Designing Meaning Representations",
month = jun,
year = "2023",
address = "Nancy, France",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.dmr-1.5/",
pages = "45--51",
abstract = "Actions are critical for interpreting dialogue: they provide context for demonstratives and definite descriptions in discourse, and they continually update the common ground. This paper describes how Abstract Meaning Representation (AMR) can be used to annotate actions in multimodal human-human and human-object interactions. We conduct initial annotations of shared task and first-person point-of-view videos. We show that AMRs can be interpreted by a proxy language, such as VoxML, as executable annotation structures in order to recreate and simulate a series of annotated events."
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="tam-etal-2023-annotating">
<titleInfo>
<title>Annotating Situated Actions in Dialogue</title>
</titleInfo>
<name type="personal">
<namePart type="given">Christopher</namePart>
<namePart type="family">Tam</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Richard</namePart>
<namePart type="family">Brutti</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Kenneth</namePart>
<namePart type="family">Lai</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">James</namePart>
<namePart type="family">Pustejovsky</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2023-06</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Fourth International Workshop on Designing Meaning Representations</title>
</titleInfo>
<name type="personal">
<namePart type="given">Julia</namePart>
<namePart type="family">Bonn</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Nianwen</namePart>
<namePart type="family">Xue</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Nancy, France</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Actions are critical for interpreting dialogue: they provide context for demonstratives and definite descriptions in discourse, and they continually update the common ground. This paper describes how Abstract Meaning Representation (AMR) can be used to annotate actions in multimodal human-human and human-object interactions. We conduct initial annotations of shared task and first-person point-of-view videos. We show that AMRs can be interpreted by a proxy language, such as VoxML, as executable annotation structures in order to recreate and simulate a series of annotated events.</abstract>
<identifier type="citekey">tam-etal-2023-annotating</identifier>
<location>
<url>https://aclanthology.org/2023.dmr-1.5/</url>
</location>
<part>
<date>2023-06</date>
<extent unit="page">
<start>45</start>
<end>51</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Annotating Situated Actions in Dialogue
%A Tam, Christopher
%A Brutti, Richard
%A Lai, Kenneth
%A Pustejovsky, James
%Y Bonn, Julia
%Y Xue, Nianwen
%S Proceedings of the Fourth International Workshop on Designing Meaning Representations
%D 2023
%8 June
%I Association for Computational Linguistics
%C Nancy, France
%F tam-etal-2023-annotating
%X Actions are critical for interpreting dialogue: they provide context for demonstratives and definite descriptions in discourse, and they continually update the common ground. This paper describes how Abstract Meaning Representation (AMR) can be used to annotate actions in multimodal human-human and human-object interactions. We conduct initial annotations of shared task and first-person point-of-view videos. We show that AMRs can be interpreted by a proxy language, such as VoxML, as executable annotation structures in order to recreate and simulate a series of annotated events.
%U https://aclanthology.org/2023.dmr-1.5/
%P 45-51
Markdown (Informal)
[Annotating Situated Actions in Dialogue](https://aclanthology.org/2023.dmr-1.5/) (Tam et al., DMR 2023)
ACL
- Christopher Tam, Richard Brutti, Kenneth Lai, and James Pustejovsky. 2023. Annotating Situated Actions in Dialogue. In Proceedings of the Fourth International Workshop on Designing Meaning Representations, pages 45–51, Nancy, France. Association for Computational Linguistics.