@inproceedings{chen-etal-2021-mitigating,
title = "Mitigating Temporal-Drift: A Simple Approach to Keep {NER} Models Crisp",
author = "Chen, Shuguang and
Neves, Leonardo and
Solorio, Thamar",
editor = "Ku, Lun-Wei and
Li, Cheng-Te",
booktitle = "Proceedings of the Ninth International Workshop on Natural Language Processing for Social Media",
month = jun,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.socialnlp-1.14",
doi = "10.18653/v1/2021.socialnlp-1.14",
pages = "163--169",
abstract = "Performance of neural models for named entity recognition degrades over time, becoming stale. This degradation is due to temporal drift, the change in our target variables{'} statistical properties over time. This issue is especially problematic for social media data, where topics change rapidly. In order to mitigate the problem, data annotation and retraining of models is common. Despite its usefulness, this process is expensive and time-consuming, which motivates new research on efficient model updating. In this paper, we propose an intuitive approach to measure the potential trendiness of tweets and use this metric to select the most informative instances to use for training. We conduct experiments on three state-of-the-art models on the Temporal Twitter Dataset. Our approach shows larger increases in prediction accuracy with less training data than the alternatives, making it an attractive, practical solution.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="chen-etal-2021-mitigating">
<titleInfo>
<title>Mitigating Temporal-Drift: A Simple Approach to Keep NER Models Crisp</title>
</titleInfo>
<name type="personal">
<namePart type="given">Shuguang</namePart>
<namePart type="family">Chen</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Leonardo</namePart>
<namePart type="family">Neves</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Thamar</namePart>
<namePart type="family">Solorio</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2021-06</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Ninth International Workshop on Natural Language Processing for Social Media</title>
</titleInfo>
<name type="personal">
<namePart type="given">Lun-Wei</namePart>
<namePart type="family">Ku</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Cheng-Te</namePart>
<namePart type="family">Li</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Online</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Performance of neural models for named entity recognition degrades over time, becoming stale. This degradation is due to temporal drift, the change in our target variables’ statistical properties over time. This issue is especially problematic for social media data, where topics change rapidly. In order to mitigate the problem, data annotation and retraining of models is common. Despite its usefulness, this process is expensive and time-consuming, which motivates new research on efficient model updating. In this paper, we propose an intuitive approach to measure the potential trendiness of tweets and use this metric to select the most informative instances to use for training. We conduct experiments on three state-of-the-art models on the Temporal Twitter Dataset. Our approach shows larger increases in prediction accuracy with less training data than the alternatives, making it an attractive, practical solution.</abstract>
<identifier type="citekey">chen-etal-2021-mitigating</identifier>
<identifier type="doi">10.18653/v1/2021.socialnlp-1.14</identifier>
<location>
<url>https://aclanthology.org/2021.socialnlp-1.14</url>
</location>
<part>
<date>2021-06</date>
<extent unit="page">
<start>163</start>
<end>169</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Mitigating Temporal-Drift: A Simple Approach to Keep NER Models Crisp
%A Chen, Shuguang
%A Neves, Leonardo
%A Solorio, Thamar
%Y Ku, Lun-Wei
%Y Li, Cheng-Te
%S Proceedings of the Ninth International Workshop on Natural Language Processing for Social Media
%D 2021
%8 June
%I Association for Computational Linguistics
%C Online
%F chen-etal-2021-mitigating
%X Performance of neural models for named entity recognition degrades over time, becoming stale. This degradation is due to temporal drift, the change in our target variables’ statistical properties over time. This issue is especially problematic for social media data, where topics change rapidly. In order to mitigate the problem, data annotation and retraining of models is common. Despite its usefulness, this process is expensive and time-consuming, which motivates new research on efficient model updating. In this paper, we propose an intuitive approach to measure the potential trendiness of tweets and use this metric to select the most informative instances to use for training. We conduct experiments on three state-of-the-art models on the Temporal Twitter Dataset. Our approach shows larger increases in prediction accuracy with less training data than the alternatives, making it an attractive, practical solution.
%R 10.18653/v1/2021.socialnlp-1.14
%U https://aclanthology.org/2021.socialnlp-1.14
%U https://doi.org/10.18653/v1/2021.socialnlp-1.14
%P 163-169
Markdown (Informal)
[Mitigating Temporal-Drift: A Simple Approach to Keep NER Models Crisp](https://aclanthology.org/2021.socialnlp-1.14) (Chen et al., SocialNLP 2021)
ACL