Crowdsourcing Hypothesis Tests: Making transparent how design choices shape research results

Publikation: Bidrag til tidsskriftTidsskriftartikelForskningfagfællebedømt

Standard

Crowdsourcing Hypothesis Tests : Making transparent how design choices shape research results. / Crowdsourced Hypothesis Tests Collaboration.

I: Psychological Bulletin, Bind 146, Nr. 5, 2020, s. 451-479.

Publikation: Bidrag til tidsskriftTidsskriftartikelForskningfagfællebedømt

Harvard

Crowdsourced Hypothesis Tests Collaboration 2020, 'Crowdsourcing Hypothesis Tests: Making transparent how design choices shape research results', Psychological Bulletin, bind 146, nr. 5, s. 451-479. https://doi.org/10.1037/bul0000220

APA

Crowdsourced Hypothesis Tests Collaboration (2020). Crowdsourcing Hypothesis Tests: Making transparent how design choices shape research results. Psychological Bulletin, 146(5), 451-479. https://doi.org/10.1037/bul0000220

Vancouver

Crowdsourced Hypothesis Tests Collaboration. Crowdsourcing Hypothesis Tests: Making transparent how design choices shape research results. Psychological Bulletin. 2020;146(5):451-479. https://doi.org/10.1037/bul0000220

Author

Crowdsourced Hypothesis Tests Collaboration. / Crowdsourcing Hypothesis Tests : Making transparent how design choices shape research results. I: Psychological Bulletin. 2020 ; Bind 146, Nr. 5. s. 451-479.

Bibtex

@article{e36f20bdc9334ea3b852065cef30d36c,
title = "Crowdsourcing Hypothesis Tests: Making transparent how design choices shape research results",
abstract = "To what extent are research results influenced by subjective decisions that scientists make as they design studies? Fifteen research teams independently designed studies to answer five original research questions related to moral judgments, negotiations, and implicit cognition. Participants from 2 separate large samples (total N = 15,000) were then randomly assigned to complete 1 version of each study. Effect sizes varied dramatically across different sets of materials designed to test the same hypothesis: Materials from different teams rendered statistically significant effects in opposite directions for 4 of 5 hypotheses, with the narrowest range in estimates being d = —0.37 to + 0.26. Meta-analysis and a Bayesian perspective on the results revealed overall support for 2 hypotheses and a lack of support for 3 hypotheses. Overall, practically none of the variability in effect sizes was attributable to the skill of the research team in designing materials, whereas considerable variability was attributable to the hypothesis being tested. In a forecasting survey, predictions of other scientists were significantly correlated with study results, both across and within hypotheses. Crowdsourced testing of research hypotheses helps reveal the true consistency of empirical support for a scientific claim.",
keywords = "conceptual replications, crowdsourcing, forecasting, research robustness, scientific transparency",
author = "Landy, {Justin F.} and Miaolei Liam and Ding, {Isabel L.} and Domenico Viganola and Warren Tierney and Anna Dreber and Magnus Johannesson and Thomas Pfeiffer and Ebersole, {Charles R.} and Gronau, {Quentin F.} and Alexander Ly and Bergh, {Don Van Den} and Maarten Marsman and Koen Derks and Wagenmaker, {Eric Jan} and Andrew Proctor and Bartels, {Daniel M.} and Bauman, {Christopher W.} and Brady, {William J.} and Felix Cheung and Andrei Cimpian and Simone Dohle and Donnellan, {M. Brent} and Adam Hahn and Hall, {Michael P.} and William Jim{\'e}nez-Leal and Johnson, {David J.} and Lucas, {Richard E.} and Beno{\^I}t Monin and Andres Montealegre and Elizabeth Mullen and Jun Pang and Jennifer Ray and Reinero, {Diego A.} and Jesse Reynolds and Walter Sowden and Daniel Storage and Runkun Su and Tworek, {Christina M.} and {Van Bavel}, {Jay J.} and Daniel Walco and Julian Wills and Xiaobing Xu and Yam, {Kai Chi} and Xiaoyu Yang and Cunningham, {William A.} and Martin Schweinsberg and Molly Urwitz and Simon Columbus and Paul Conway and {Crowdsourced Hypothesis Tests Collaboration}",
note = "Publisher Copyright: {\textcopyright} 2020, American Psychological Association.",
year = "2020",
doi = "10.1037/bul0000220",
language = "English",
volume = "146",
pages = "451--479",
journal = "Psychological Bulletin",
issn = "0033-2909",
publisher = "American Psychological Association Inc.",
number = "5",

}

RIS

TY - JOUR

T1 - Crowdsourcing Hypothesis Tests

T2 - Making transparent how design choices shape research results

AU - Landy, Justin F.

AU - Liam, Miaolei

AU - Ding, Isabel L.

AU - Viganola, Domenico

AU - Tierney, Warren

AU - Dreber, Anna

AU - Johannesson, Magnus

AU - Pfeiffer, Thomas

AU - Ebersole, Charles R.

AU - Gronau, Quentin F.

AU - Ly, Alexander

AU - Bergh, Don Van Den

AU - Marsman, Maarten

AU - Derks, Koen

AU - Wagenmaker, Eric Jan

AU - Proctor, Andrew

AU - Bartels, Daniel M.

AU - Bauman, Christopher W.

AU - Brady, William J.

AU - Cheung, Felix

AU - Cimpian, Andrei

AU - Dohle, Simone

AU - Donnellan, M. Brent

AU - Hahn, Adam

AU - Hall, Michael P.

AU - Jiménez-Leal, William

AU - Johnson, David J.

AU - Lucas, Richard E.

AU - Monin, BenoÎt

AU - Montealegre, Andres

AU - Mullen, Elizabeth

AU - Pang, Jun

AU - Ray, Jennifer

AU - Reinero, Diego A.

AU - Reynolds, Jesse

AU - Sowden, Walter

AU - Storage, Daniel

AU - Su, Runkun

AU - Tworek, Christina M.

AU - Van Bavel, Jay J.

AU - Walco, Daniel

AU - Wills, Julian

AU - Xu, Xiaobing

AU - Yam, Kai Chi

AU - Yang, Xiaoyu

AU - Cunningham, William A.

AU - Schweinsberg, Martin

AU - Urwitz, Molly

AU - Columbus, Simon

AU - Conway, Paul

AU - Crowdsourced Hypothesis Tests Collaboration

N1 - Publisher Copyright: © 2020, American Psychological Association.

PY - 2020

Y1 - 2020

N2 - To what extent are research results influenced by subjective decisions that scientists make as they design studies? Fifteen research teams independently designed studies to answer five original research questions related to moral judgments, negotiations, and implicit cognition. Participants from 2 separate large samples (total N = 15,000) were then randomly assigned to complete 1 version of each study. Effect sizes varied dramatically across different sets of materials designed to test the same hypothesis: Materials from different teams rendered statistically significant effects in opposite directions for 4 of 5 hypotheses, with the narrowest range in estimates being d = —0.37 to + 0.26. Meta-analysis and a Bayesian perspective on the results revealed overall support for 2 hypotheses and a lack of support for 3 hypotheses. Overall, practically none of the variability in effect sizes was attributable to the skill of the research team in designing materials, whereas considerable variability was attributable to the hypothesis being tested. In a forecasting survey, predictions of other scientists were significantly correlated with study results, both across and within hypotheses. Crowdsourced testing of research hypotheses helps reveal the true consistency of empirical support for a scientific claim.

AB - To what extent are research results influenced by subjective decisions that scientists make as they design studies? Fifteen research teams independently designed studies to answer five original research questions related to moral judgments, negotiations, and implicit cognition. Participants from 2 separate large samples (total N = 15,000) were then randomly assigned to complete 1 version of each study. Effect sizes varied dramatically across different sets of materials designed to test the same hypothesis: Materials from different teams rendered statistically significant effects in opposite directions for 4 of 5 hypotheses, with the narrowest range in estimates being d = —0.37 to + 0.26. Meta-analysis and a Bayesian perspective on the results revealed overall support for 2 hypotheses and a lack of support for 3 hypotheses. Overall, practically none of the variability in effect sizes was attributable to the skill of the research team in designing materials, whereas considerable variability was attributable to the hypothesis being tested. In a forecasting survey, predictions of other scientists were significantly correlated with study results, both across and within hypotheses. Crowdsourced testing of research hypotheses helps reveal the true consistency of empirical support for a scientific claim.

KW - conceptual replications

KW - crowdsourcing

KW - forecasting

KW - research robustness

KW - scientific transparency

UR - http://www.scopus.com/inward/record.url?scp=85081412411&partnerID=8YFLogxK

U2 - 10.1037/bul0000220

DO - 10.1037/bul0000220

M3 - Journal article

C2 - 31944796

AN - SCOPUS:85081412411

VL - 146

SP - 451

EP - 479

JO - Psychological Bulletin

JF - Psychological Bulletin

SN - 0033-2909

IS - 5

ER -

ID: 294828511