Prevalence and Prevention of Large Language Model Use in Crowd Work

  • Veniamin Veselovsky
  • , Manoel Horta Ribeiro
  • , Philip J. Cozzolino
  • , Andrew Gordon
  • , David Rothschild
  • , Robert West

Research output: Contribution to journalArticlepeer-review

10 Scopus citations

Abstract

Crowd work platforms, such as Prolific and Amazon Mechanical Turk, play an important part in academia and industry, empowering the creation, annotation, and summarization of data,11 as well as surveys and experiments.21 At the same time, large language models (LLMs), such as ChatGPT, Gemini, and Claude, promise similar capabilities. They are remarkable data annotators10 and can, in some cases, accurately simulate human behavior, enabling in-silico experiments and surveys that yield human-like results.2 Yet, if crowd workers were to start using LLMs, this could threaten the validity of data generated using crowd work platforms. Sometimes, researchers seek to observe unaided human responses (even if LLMs could provide a good proxy), and LLMs still often fail to accurately simulate human behavior.22 Further, LLM-generated data may degrade subsequent models trained on it.23 Here, we investigate the extent to which crowd workers use LLMs in a text-production task and whether targeted mitigation strategies can prevent LLM use.

Original languageEnglish (US)
Pages (from-to)42-47
Number of pages6
JournalCommunications of the ACM
Volume68
Issue number3
DOIs
StatePublished - Mar 1 2025
Externally publishedYes

All Science Journal Classification (ASJC) codes

  • General Computer Science

Fingerprint

Dive into the research topics of 'Prevalence and Prevention of Large Language Model Use in Crowd Work'. Together they form a unique fingerprint.

Cite this