Discussion
Loading...

Post

Log in
  • About
  • Code of conduct
  • Privacy
  • Users
  • Instances
  • About Bonfire
Nick Byrd, Ph.D.
Nick Byrd, Ph.D.
@ByrdNick@nerdculture.de  ·  activity timestamp 3 years ago

Repli-scooping some of what I find in a soon-to-be finished paper about correlations and effects between reflective reasoning and philosophical thought experiments across multiple participant samples, the Brauer lab finds
- mTurk workers offered lower quality and lower value data than Prolific workers, students, and even CloudResearch's approved mTurk workers
- Qualtrics panels had the least value but moderate quality
- Students seemed to offer the highest value

#noxP https://doi.org/10.1371/journal.pone.0279720

4 media
Sorry, no caption provided by author
Sorry, no caption provided by author
Sorry, no caption provided by author
Sorry, no caption provided by author
Sorry, no caption provided by author
Sorry, no caption provided by author
Sorry, no caption provided by author
Sorry, no caption provided by author
Sorry, no caption provided by author
Sorry, no caption provided by author
Sorry, no caption provided by author
Sorry, no caption provided by author

Data quality in online human-subjects research: Comparisons between MTurk, Prolific, CloudResearch, Qualtrics, and SONA

  • Copy link
  • Flag this post
  • Block
Nick Byrd, Ph.D.
Nick Byrd, Ph.D.
@ByrdNick@nerdculture.de replied  ·  activity timestamp last week

I forgot to share the #mTurk data quality result that got scooped:

“In late 2020…. Participants from the United States were recruited from Amazon Mechanical Turk, - #CloudResearch, #Prolific, and a #university. One participant source yielded up to 18 times as many low-quality respondents as the other three.”

https://doi.org/10.1093/analys/anaf015

#psychology #philosophy #surveyMethods #quantMethods #dataScience #qualityControl

Figure 1b. Number of observations per sample,
before and after filtering for data quality (N = 460).
Figure 1b. Number of observations per sample, before and after filtering for data quality (N = 460).
Figure 1b. Number of observations per sample, before and after filtering for data quality (N = 460).

Reflection-philosophy order effects and correlations across samples

  • Copy link
  • Flag this comment
  • Block
Nick Byrd, Ph.D.
Nick Byrd, Ph.D.
@ByrdNick@nerdculture.de replied  ·  activity timestamp last week

We've found recruiting people for online #research via #onlineAdvertising yielded good results on overt and covert #dataQuality measures (perhaps because participation incentives aren't financial):

Attention checks passed ≅ 2.6 out of 3

ReCAPTCHA (v3) ≅ 0.94 out of 1.0

Sample size > 5000 (from six continents)

https://doi.org/10.1017/S0034412525000198

#surveyMethods #cogSci #psychology #xPhi #QualityControl #econ #marketing

Analytic atheism and analytic apostasy across cultures

  • Copy link
  • Flag this comment
  • Block
Tom Stafford
Tom Stafford
@tomstafford@mastodon.online replied  ·  activity timestamp last week

@ByrdNick interesting. I love the idea of giving participant non-financial incentives in return for their participation/data. Much more of a partnership model. Harder to do, required more thought, but great to see if yields data quality (which is sure to drop only further https://tomstafford.substack.com/p/faking-survey-responses-with-llms )

Faking survey responses with LLMs

Survey research is a key mechanism by which society knows itself, we should all be worried if can't be trusted.
  • Copy link
  • Flag this comment
  • Block

bonfire.cafe

A space for Bonfire maintainers and contributors to communicate

bonfire.cafe: About · Code of conduct · Privacy · Users · Instances
Bonfire social · 1.0.1-beta.35 no JS en
Automatic federation enabled
Log in
  • Explore
  • About
  • Members
  • Code of Conduct