@inproceedings{panda-levitan-2021-detecting,
title = "Detecting Multilingual {COVID}-19 Misinformation on Social Media via Contextualized Embeddings",
author = "Panda, Subhadarshi and
Levitan, Sarah Ita",
editor = "Feldman, Anna and
Da San Martino, Giovanni and
Leberknight, Chris and
Nakov, Preslav",
booktitle = "Proceedings of the Fourth Workshop on NLP for Internet Freedom: Censorship, Disinformation, and Propaganda",
month = jun,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.nlp4if-1.19/",
doi = "10.18653/v1/2021.nlp4if-1.19",
pages = "125--129",
abstract = "We present machine learning classifiers to automatically identify COVID-19 misinformation on social media in three languages: English, Bulgarian, and Arabic. We compared 4 multitask learning models for this task and found that a model trained with English BERT achieves the best results for English, and multilingual BERT achieves the best results for Bulgarian and Arabic. We experimented with zero shot, few shot, and target-only conditions to evaluate the impact of target-language training data on classifier performance, and to understand the capabilities of different models to generalize across languages in detecting misinformation online. This work was performed as a submission to the shared task, NLP4IF 2021: Fighting the COVID-19 Infodemic. Our best models achieved the second best evaluation test results for Bulgarian and Arabic among all the participating teams and obtained competitive scores for English."
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="panda-levitan-2021-detecting">
<titleInfo>
<title>Detecting Multilingual COVID-19 Misinformation on Social Media via Contextualized Embeddings</title>
</titleInfo>
<name type="personal">
<namePart type="given">Subhadarshi</namePart>
<namePart type="family">Panda</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Sarah</namePart>
<namePart type="given">Ita</namePart>
<namePart type="family">Levitan</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2021-06</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Fourth Workshop on NLP for Internet Freedom: Censorship, Disinformation, and Propaganda</title>
</titleInfo>
<name type="personal">
<namePart type="given">Anna</namePart>
<namePart type="family">Feldman</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Giovanni</namePart>
<namePart type="family">Da San Martino</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Chris</namePart>
<namePart type="family">Leberknight</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Preslav</namePart>
<namePart type="family">Nakov</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Online</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>We present machine learning classifiers to automatically identify COVID-19 misinformation on social media in three languages: English, Bulgarian, and Arabic. We compared 4 multitask learning models for this task and found that a model trained with English BERT achieves the best results for English, and multilingual BERT achieves the best results for Bulgarian and Arabic. We experimented with zero shot, few shot, and target-only conditions to evaluate the impact of target-language training data on classifier performance, and to understand the capabilities of different models to generalize across languages in detecting misinformation online. This work was performed as a submission to the shared task, NLP4IF 2021: Fighting the COVID-19 Infodemic. Our best models achieved the second best evaluation test results for Bulgarian and Arabic among all the participating teams and obtained competitive scores for English.</abstract>
<identifier type="citekey">panda-levitan-2021-detecting</identifier>
<identifier type="doi">10.18653/v1/2021.nlp4if-1.19</identifier>
<location>
<url>https://aclanthology.org/2021.nlp4if-1.19/</url>
</location>
<part>
<date>2021-06</date>
<extent unit="page">
<start>125</start>
<end>129</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Detecting Multilingual COVID-19 Misinformation on Social Media via Contextualized Embeddings
%A Panda, Subhadarshi
%A Levitan, Sarah Ita
%Y Feldman, Anna
%Y Da San Martino, Giovanni
%Y Leberknight, Chris
%Y Nakov, Preslav
%S Proceedings of the Fourth Workshop on NLP for Internet Freedom: Censorship, Disinformation, and Propaganda
%D 2021
%8 June
%I Association for Computational Linguistics
%C Online
%F panda-levitan-2021-detecting
%X We present machine learning classifiers to automatically identify COVID-19 misinformation on social media in three languages: English, Bulgarian, and Arabic. We compared 4 multitask learning models for this task and found that a model trained with English BERT achieves the best results for English, and multilingual BERT achieves the best results for Bulgarian and Arabic. We experimented with zero shot, few shot, and target-only conditions to evaluate the impact of target-language training data on classifier performance, and to understand the capabilities of different models to generalize across languages in detecting misinformation online. This work was performed as a submission to the shared task, NLP4IF 2021: Fighting the COVID-19 Infodemic. Our best models achieved the second best evaluation test results for Bulgarian and Arabic among all the participating teams and obtained competitive scores for English.
%R 10.18653/v1/2021.nlp4if-1.19
%U https://aclanthology.org/2021.nlp4if-1.19/
%U https://doi.org/10.18653/v1/2021.nlp4if-1.19
%P 125-129
Markdown (Informal)
[Detecting Multilingual COVID-19 Misinformation on Social Media via Contextualized Embeddings](https://aclanthology.org/2021.nlp4if-1.19/) (Panda & Levitan, NLP4IF 2021)
ACL