TY - GEN
T1 - Fostering Appropriate Reliance on Large Language Models
T2 - 2025 CHI Conference on Human Factors in Computing Systems, CHI 2025
AU - Kim, Sunnie S.Y.
AU - Vaughan, Jennifer Wortman
AU - Liao, Q. Vera
AU - Lombrozo, Tania
AU - Russakovsky, Olga
N1 - Publisher Copyright:
© 2025 Copyright held by the owner/author(s).
PY - 2025/4/26
Y1 - 2025/4/26
N2 - Large language models (LLMs) can produce erroneous responses that sound fluent and convincing, raising the risk that users will rely on these responses as if they were correct. Mitigating such overreliance is a key challenge. Through a think-aloud study in which participants use an LLM-infused application to answer objective questions, we identify several features of LLM responses that shape users' reliance: explanations (supporting details for answers), inconsistencies in explanations, and sources. Through a large-scale, pre-registered, controlled experiment (N=308), we isolate and study the effects of these features on users' reliance, accuracy, and other measures. We find that the presence of explanations increases reliance on both correct and incorrect responses. However, we observe less reliance on incorrect responses when sources are provided or when explanations exhibit inconsistencies. We discuss the implications of these findings for fostering appropriate reliance on LLMs.
AB - Large language models (LLMs) can produce erroneous responses that sound fluent and convincing, raising the risk that users will rely on these responses as if they were correct. Mitigating such overreliance is a key challenge. Through a think-aloud study in which participants use an LLM-infused application to answer objective questions, we identify several features of LLM responses that shape users' reliance: explanations (supporting details for answers), inconsistencies in explanations, and sources. Through a large-scale, pre-registered, controlled experiment (N=308), we isolate and study the effects of these features on users' reliance, accuracy, and other measures. We find that the presence of explanations increases reliance on both correct and incorrect responses. However, we observe less reliance on incorrect responses when sources are provided or when explanations exhibit inconsistencies. We discuss the implications of these findings for fostering appropriate reliance on LLMs.
KW - Explanations
KW - Human-AI interaction
KW - Inconsistencies
KW - Large language models
KW - Overreliance
KW - Question answering
KW - Sources
UR - http://www.scopus.com/inward/record.url?scp=105005746734&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=105005746734&partnerID=8YFLogxK
U2 - 10.1145/3706598.3714020
DO - 10.1145/3706598.3714020
M3 - Conference contribution
AN - SCOPUS:105005746734
T3 - Conference on Human Factors in Computing Systems - Proceedings
BT - CHI 2025 - Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems
PB - Association for Computing Machinery
Y2 - 26 April 2025 through 1 May 2025
ER -