TY - JOUR
T1 - Data quality of platforms and panels for online behavioral research
AU - Peer, Eyal
AU - Rothschild, David
AU - Gordon, Andrew
AU - Evernden, Zak
AU - Damer, Ekaterina
N1 - Publisher Copyright:
© 2021, The Psychonomic Society, Inc.
PY - 2022/8
Y1 - 2022/8
N2 - We examine key aspects of data quality for online behavioral research between selected platforms (Amazon Mechanical Turk, CloudResearch, and Prolific) and panels (Qualtrics and Dynata). To identify the key aspects of data quality, we first engaged with the behavioral research community to discover which aspects are most critical to researchers and found that these include attention, comprehension, honesty, and reliability. We then explored differences in these data quality aspects in two studies (N ~ 4000), with or without data quality filters (approval ratings). We found considerable differences between the sites, especially in comprehension, attention, and dishonesty. In Study 1 (without filters), we found that only Prolific provided high data quality on all measures. In Study 2 (with filters), we found high data quality among CloudResearch and Prolific. MTurk showed alarmingly low data quality even with data quality filters. We also found that while reputation (approval rating) did not predict data quality, frequency and purpose of usage did, especially on MTurk: the lowest data quality came from MTurk participants who report using the site as their main source of income but spend few hours on it per week. We provide a framework for future investigation into the ever-changing nature of data quality in online research, and how the evolving set of platforms and panels performs on these key aspects.
AB - We examine key aspects of data quality for online behavioral research between selected platforms (Amazon Mechanical Turk, CloudResearch, and Prolific) and panels (Qualtrics and Dynata). To identify the key aspects of data quality, we first engaged with the behavioral research community to discover which aspects are most critical to researchers and found that these include attention, comprehension, honesty, and reliability. We then explored differences in these data quality aspects in two studies (N ~ 4000), with or without data quality filters (approval ratings). We found considerable differences between the sites, especially in comprehension, attention, and dishonesty. In Study 1 (without filters), we found that only Prolific provided high data quality on all measures. In Study 2 (with filters), we found high data quality among CloudResearch and Prolific. MTurk showed alarmingly low data quality even with data quality filters. We also found that while reputation (approval rating) did not predict data quality, frequency and purpose of usage did, especially on MTurk: the lowest data quality came from MTurk participants who report using the site as their main source of income but spend few hours on it per week. We provide a framework for future investigation into the ever-changing nature of data quality in online research, and how the evolving set of platforms and panels performs on these key aspects.
KW - Amazon mechanical turk
KW - Attention
KW - Comprehension
KW - Data quality
KW - Honesty
KW - Online research
KW - Prolific
KW - Reliability
UR - http://www.scopus.com/inward/record.url?scp=85124869424&partnerID=8YFLogxK
U2 - 10.3758/s13428-021-01694-3
DO - 10.3758/s13428-021-01694-3
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
C2 - 34590289
AN - SCOPUS:85124869424
SN - 1554-351X
VL - 54
SP - 1643
EP - 1662
JO - Behavior Research Methods
JF - Behavior Research Methods
IS - 4
ER -