Skip to main content

Equivalence Tests: A Practical Primer for t Tests, Correlations, and Meta-Analyses

๐Ÿ“„ Original study โ†—
Lakens, Daniรซl โ€ข 2017 Current Era โ€ข methodology

๐Ÿ“Œ Appears in:

Plain English Summary

Here's a problem most people don't realize exists: when a scientific study finds "no significant result," that does NOT actually mean they proved nothing is happening. It just means the evidence was murky. Lakens offers a fix called equivalence testing (specifically the TOST procedure), which lets researchers formally demonstrate that an effect is so tiny it's practically zero. You pick the smallest effect you'd care about, then statistically show your result falls inside that "who cares" zone. This is a big deal for parapsychology, where failed replications of psychic experiments often get shrugged off as "inconclusive" rather than "genuine evidence against psi." Lakens provides formulas, spreadsheets, and an R software package so anyone can do this. One fair critique: the recommended way to set your "smallest interesting effect" can be a bit circular, since it depends on how many participants you can recruit rather than what your theory actually predicts.

Actual Paper Abstract

Scientists should be able to provide support for the absence of a meaningful effect. Currently, researchers often incorrectly conclude an effect is absent based a nonsignificant result. A widely recommended approach within a frequentist framework is to test for equivalence. In equivalence tests, such as the two one-sided tests (TOST) procedure discussed in this article, an upper and lower equivalence bound is specified based on the smallest effect size of interest. The TOST procedure can be used to statistically reject the presence of effects large enough to be considered worthwhile. This practical primer with accompanying spreadsheet and R package enables psychologists to easily perform equivalence tests (and power analyses) by setting equivalence bounds based on standardized effect sizes and provides recommendations to prespecify equivalence bounds. Extending your statistical tool kit with equivalence tests is an easy way to improve your statistical and theoretical inferences.

Research Notes

Critical methodological resource for psi research: enables formal testing of 'evidence for no psi' versus 'inconclusive results.' Essential for evaluating replication failures (e.g., Bem FTF replications, Ganzfeld failures) and for designing informative null-result studies. The SESOI recommendationโ€”setting bounds based on maximum feasible sample sizeโ€”is pragmatic but circular; better for theory development would be prespecifying bounds based on theoretical predictions.

Scientists should be able to provide support for the absence of a meaningful effect, but nonsignificant p-values cannot establish this. This tutorial introduces the two one-sided tests (TOST) procedure for equivalence testing: researchers specify upper and lower equivalence bounds based on the smallest effect size of interest (SESOI), then test whether observed effects fall within this range. Formulas and worked examples are provided for independent/dependent t-tests, correlations, and meta-analyses. An accompanying spreadsheet and R package (TOSTER) enable psychologists to perform equivalence tests and power analyses. Adopting equivalence testing prevents misinterpreting nonsignificant results as evidence for the null, enables replication studies to test for absence of meaningful effects, and encourages researchers to specify which effect sizes they find theoretically worthwhile.

Links

Related Papers

Also by these authors

More in Methodology

๐Ÿ“‹ Cite this paper
APA
Lakens, Daniรซl (2017). Equivalence Tests: A Practical Primer for t Tests, Correlations, and Meta-Analyses. Social Psychological and Personality Science. https://doi.org/10.1177/1948550617697177
BibTeX
@article{lakens_2017_equivalence,
  title = {Equivalence Tests: A Practical Primer for t Tests, Correlations, and Meta-Analyses},
  author = {Lakens, Daniรซl},
  year = {2017},
  journal = {Social Psychological and Personality Science},
  doi = {10.1177/1948550617697177},
}