TY - GEN
T1 - Can Language Models Recognize Convincing Arguments?
AU - Rescala, Paula Dolores
AU - Ribeiro, Manoel Horta
AU - Hu, Tiancheng
AU - West, Robert
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024
Y1 - 2024
N2 - The capabilities of large language models (LLMs) have raised concerns about their potential to create and propagate convincing narratives.Here, we study their performance in detecting convincing arguments to gain insights into LLMs' persuasive capabilities without directly engaging in experimentation with humans.We extend a dataset by Durmus and Cardie (2018) with debates, votes, and user traits and propose tasks measuring LLMs' ability to (1) distinguish between strong and weak arguments, (2) predict stances based on beliefs and demographic characteristics, and (3) determine the appeal of an argument to an individual based on their traits.We show that LLMs perform on par with humans in these tasks and that combining predictions from different LLMs yields significant performance gains, surpassing human performance.The data and code released with this paper contribute to the crucial effort of continuously evaluating and monitoring LLMs' capabilities and potential impact.(https://go.epfl.ch/persuasion-llm).
AB - The capabilities of large language models (LLMs) have raised concerns about their potential to create and propagate convincing narratives.Here, we study their performance in detecting convincing arguments to gain insights into LLMs' persuasive capabilities without directly engaging in experimentation with humans.We extend a dataset by Durmus and Cardie (2018) with debates, votes, and user traits and propose tasks measuring LLMs' ability to (1) distinguish between strong and weak arguments, (2) predict stances based on beliefs and demographic characteristics, and (3) determine the appeal of an argument to an individual based on their traits.We show that LLMs perform on par with humans in these tasks and that combining predictions from different LLMs yields significant performance gains, surpassing human performance.The data and code released with this paper contribute to the crucial effort of continuously evaluating and monitoring LLMs' capabilities and potential impact.(https://go.epfl.ch/persuasion-llm).
UR - http://www.scopus.com/inward/record.url?scp=85217622462&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85217622462&partnerID=8YFLogxK
U2 - 10.18653/v1/2024.findings-emnlp.515
DO - 10.18653/v1/2024.findings-emnlp.515
M3 - Conference contribution
AN - SCOPUS:85217622462
T3 - EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Findings of EMNLP 2024
SP - 8826
EP - 8837
BT - EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Findings of EMNLP 2024
A2 - Al-Onaizan, Yaser
A2 - Bansal, Mohit
A2 - Chen, Yun-Nung
PB - Association for Computational Linguistics (ACL)
T2 - 2024 Findings of the Association for Computational Linguistics, EMNLP 2024
Y2 - 12 November 2024 through 16 November 2024
ER -