@inproceedings{dhuliawala-etal-2023-diachronic,
title = "A Diachronic Perspective on User Trust in {AI} under Uncertainty",
author = "Dhuliawala, Shehzaad and
Zouhar, Vil{\'e}m and
El-Assady, Mennatallah and
Sachan, Mrinmaya",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.emnlp-main.339/",
doi = "10.18653/v1/2023.emnlp-main.339",
pages = "5567--5580",
abstract = "In human-AI collaboration, users typically form a mental model of the AI system, which captures the user`s beliefs about when the system performs well and when it does not. The construction of this mental model is guided by both the system`s veracity as well as the system output presented to the user e.g., the system`s confidence and an explanation for the prediction. However, modern NLP systems are seldom calibrated and are often confidently incorrect about their predictions, which violates users' mental model and erodes their trust. In this work, we design a study where users bet on the correctness of an NLP system, and use it to study the evolution of user trust as a response to these trust-eroding events and how the user trust is rebuilt as a function of time after these events. We find that even a few highly inaccurate confidence estimation instances are enough to damage users' trust in the system and performance, which does not easily recover over time. We further find that users are more forgiving to the NLP system if it is unconfidently correct rather than confidently incorrect, even though, from a game-theoretic perspective, their payoff is equivalent. Finally, we find that each user can entertain multiple mental models of the system based on the type of the question. These results highlight the importance of confidence calibration in developing user-centered NLP applications to avoid damaging user trust and compromising the collaboration performance."
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="dhuliawala-etal-2023-diachronic">
<titleInfo>
<title>A Diachronic Perspective on User Trust in AI under Uncertainty</title>
</titleInfo>
<name type="personal">
<namePart type="given">Shehzaad</namePart>
<namePart type="family">Dhuliawala</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Vilém</namePart>
<namePart type="family">Zouhar</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Mennatallah</namePart>
<namePart type="family">El-Assady</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Mrinmaya</namePart>
<namePart type="family">Sachan</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2023-12</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</title>
</titleInfo>
<name type="personal">
<namePart type="given">Houda</namePart>
<namePart type="family">Bouamor</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Juan</namePart>
<namePart type="family">Pino</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Kalika</namePart>
<namePart type="family">Bali</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Singapore</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>In human-AI collaboration, users typically form a mental model of the AI system, which captures the user‘s beliefs about when the system performs well and when it does not. The construction of this mental model is guided by both the system‘s veracity as well as the system output presented to the user e.g., the system‘s confidence and an explanation for the prediction. However, modern NLP systems are seldom calibrated and are often confidently incorrect about their predictions, which violates users’ mental model and erodes their trust. In this work, we design a study where users bet on the correctness of an NLP system, and use it to study the evolution of user trust as a response to these trust-eroding events and how the user trust is rebuilt as a function of time after these events. We find that even a few highly inaccurate confidence estimation instances are enough to damage users’ trust in the system and performance, which does not easily recover over time. We further find that users are more forgiving to the NLP system if it is unconfidently correct rather than confidently incorrect, even though, from a game-theoretic perspective, their payoff is equivalent. Finally, we find that each user can entertain multiple mental models of the system based on the type of the question. These results highlight the importance of confidence calibration in developing user-centered NLP applications to avoid damaging user trust and compromising the collaboration performance.</abstract>
<identifier type="citekey">dhuliawala-etal-2023-diachronic</identifier>
<identifier type="doi">10.18653/v1/2023.emnlp-main.339</identifier>
<location>
<url>https://aclanthology.org/2023.emnlp-main.339/</url>
</location>
<part>
<date>2023-12</date>
<extent unit="page">
<start>5567</start>
<end>5580</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T A Diachronic Perspective on User Trust in AI under Uncertainty
%A Dhuliawala, Shehzaad
%A Zouhar, Vilém
%A El-Assady, Mennatallah
%A Sachan, Mrinmaya
%Y Bouamor, Houda
%Y Pino, Juan
%Y Bali, Kalika
%S Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing
%D 2023
%8 December
%I Association for Computational Linguistics
%C Singapore
%F dhuliawala-etal-2023-diachronic
%X In human-AI collaboration, users typically form a mental model of the AI system, which captures the user‘s beliefs about when the system performs well and when it does not. The construction of this mental model is guided by both the system‘s veracity as well as the system output presented to the user e.g., the system‘s confidence and an explanation for the prediction. However, modern NLP systems are seldom calibrated and are often confidently incorrect about their predictions, which violates users’ mental model and erodes their trust. In this work, we design a study where users bet on the correctness of an NLP system, and use it to study the evolution of user trust as a response to these trust-eroding events and how the user trust is rebuilt as a function of time after these events. We find that even a few highly inaccurate confidence estimation instances are enough to damage users’ trust in the system and performance, which does not easily recover over time. We further find that users are more forgiving to the NLP system if it is unconfidently correct rather than confidently incorrect, even though, from a game-theoretic perspective, their payoff is equivalent. Finally, we find that each user can entertain multiple mental models of the system based on the type of the question. These results highlight the importance of confidence calibration in developing user-centered NLP applications to avoid damaging user trust and compromising the collaboration performance.
%R 10.18653/v1/2023.emnlp-main.339
%U https://aclanthology.org/2023.emnlp-main.339/
%U https://doi.org/10.18653/v1/2023.emnlp-main.339
%P 5567-5580
Markdown (Informal)
[A Diachronic Perspective on User Trust in AI under Uncertainty](https://aclanthology.org/2023.emnlp-main.339/) (Dhuliawala et al., EMNLP 2023)
ACL
- Shehzaad Dhuliawala, Vilém Zouhar, Mennatallah El-Assady, and Mrinmaya Sachan. 2023. A Diachronic Perspective on User Trust in AI under Uncertainty. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 5567–5580, Singapore. Association for Computational Linguistics.