TY - JOUR
T1 - Assessing the quality and reliability of the Amazon Mechanical Turk (MTurk) data in 2024
AU - Shimoni, Hagar
AU - Axelrod, Vadim
N1 - Publisher Copyright:
© 2025 The Authors.
PY - 2025/7/16
Y1 - 2025/7/16
N2 - Amazon Mechanical Turk (MTurk) has been one of the most popular platforms for online research in psychology and the social sciences in general. While concerns about MTurk data quality have been raised, the platform continues to be widely used. The question is whether the MTurk platform is suitable for research and, if so, whether it is used optimally. We conducted a systematic investigation of MTurk data quality and reliability, including main and replication experiments, with more than 1300 participants subdivided into three cohorts: (i) workers (i.e. participants on the MTurk platform) with master requirement (i.e. high-performing workers selected by MTurk), (ii) workers without master requirement, and (iii) workers without master requirement, but with a 95% or above approval rate. We found that master workers almost never missed attentional checks, exhibited high reliability and showed no tendency towards straightlining, therefore, these workers are recommended, especially when the naivety of participants is not a strong prerequisite and no large sample size is required. In contrast, the workers without restrictions or with a 95% or above approval-rate threshold missed many attentional checks, exhibited low reliability and showed a tendency towards straightlining, raising serious concerns about the suitability of these workers for research.
AB - Amazon Mechanical Turk (MTurk) has been one of the most popular platforms for online research in psychology and the social sciences in general. While concerns about MTurk data quality have been raised, the platform continues to be widely used. The question is whether the MTurk platform is suitable for research and, if so, whether it is used optimally. We conducted a systematic investigation of MTurk data quality and reliability, including main and replication experiments, with more than 1300 participants subdivided into three cohorts: (i) workers (i.e. participants on the MTurk platform) with master requirement (i.e. high-performing workers selected by MTurk), (ii) workers without master requirement, and (iii) workers without master requirement, but with a 95% or above approval rate. We found that master workers almost never missed attentional checks, exhibited high reliability and showed no tendency towards straightlining, therefore, these workers are recommended, especially when the naivety of participants is not a strong prerequisite and no large sample size is required. In contrast, the workers without restrictions or with a 95% or above approval-rate threshold missed many attentional checks, exhibited low reliability and showed a tendency towards straightlining, raising serious concerns about the suitability of these workers for research.
KW - Amazon Mechanical Turk
KW - MTurk
KW - approval rate
KW - attentional check items
KW - data quality
KW - master workers
KW - online experiments
KW - reliability
UR - https://www.scopus.com/pages/publications/105010920857
U2 - 10.1098/rsos.250361
DO - 10.1098/rsos.250361
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
C2 - 40727403
AN - SCOPUS:105010920857
SN - 2054-5703
VL - 12
JO - Royal Society Open Science
JF - Royal Society Open Science
IS - 7
M1 - 250361
ER -