Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Cardiff 12-5-2017

Cardiff 12-5-2017

Invited Colloquium on Designing Efficient and Informative Studies

Daniel Lakens

May 12, 2017
Tweet

Other Decks in Science

Transcript

  1. You need to justify the sample size of a study.

    What goal do you want to achieve?
  2. Statistical power is the long-run probability of observing p <

    α with N participants, assuming a specific effect size.
  3. But 1) You never know the true effect size, and

    the literature is biased, and 2) If you expect a true effect of 0, power is 0
  4. 0% 10% 20% 30% 40% 50% 60% 70% 80% 90%

    100% 10 20 30 40 50 60 70 80 90 100 110 120 130 140 150 160 170 180 190 200 Power Sample Size per condition in a independent t-test d=0.3 d=0.4 d=0.5 d=0.6 d=0.7 d=0.8
  5. My department requires sample size justification before funding a study.

    One justification the IRB accepts is 90% power.
  6. Evidence is always relative. You want a higher likelihood of

    p<0.05 when H1 is true than when H0 is true.
  7. A pilot study does not provide a meaningful effect size

    estimate for planning subsequent studies. Leon, Davis, & Kraemer, 2011
  8. Power analysis based on significant studies need to be based

    on a truncated F distribution. Taylor & Muller, 1996
  9. You can also take into account variability (‘assurance’) – e.g.,

    using safeguard power. Perugini, Gallucci, & Constantini, 2014
  10. Effect sizes from the published literature are always smaller than

    you expect, even when you take into account that effect sizes from the published literature are always smaller than you expect.
  11. Plan for the change you would like to see in

    the world. Ask yourself: What is your smallest effect size of interest?
  12. Requires you to specify H1! That’s a good thing. What

    does you theory predict, or what do you care about if H0 is false?
  13. But ‘I’m not interested in the size of the effect

    – the presence of any effect supports my theory!’ Really?
  14. If you expect a ‘medium’ effect size and plan for

    80% power, d<0.35 will never be significant.
  15. Now you can also reject effects as large as, or

    larger than, your SESOI, using an equivalence test.
  16. My prediction: Publishing a paper that say ‘p > 0.05,

    thus no effect’ will be difficult in 2019.
  17. Extending your statistical tool kit with equivalence tests is an

    easy way to improve your inferences. Lakens, 2017
  18. However, when the true effect size is larger than the

    SESOI, powering for it is inefficient (and possibly wasteful).
  19. 0% 10% 20% 30% 40% 50% 60% 70% 80% 90%

    100% 10 20 30 40 50 60 70 80 90 100 110 120 130 140 150 160 170 180 190 200 Power Sample Size per condition in a independent t-test d=0.3 d=0.4 d=0.5 d=0.6 d=0.7 d=0.8
  20. A user of NHST could always obtain a significant result

    through optional stopping. Wagenmakers, 2007
  21. Use decision rules based on p-values or Bayes factors, but

    check Frequentist properties. Schonbrodt, Wagenmakers, Zehetleitner, & Perugini, 2015
  22. If you think the current reproducibility crisis was bad, wait

    till the theory crisis in psychology starts.