1. Moore, D. S., McCabe, G. P., & Craig, B. A. (2021). Introduction to the practice of statistics (10th ed.). W. H. Freeman; Company.
2. Vincent, W. J. (2005). Statistics in kinesiology.
3. Batterham, A. M., & Hopkins, W. G. (2006). Making meaningful inferences about magnitudes.
International Journal of Sports Physiology and Performance,
1(1), 50–57.
https://doi.org/10.1123/ijspp.1.1.50
4. Hopkins, W. G., Marshall, S. W., Batterham, A. M., & Hanin, J. (2009). Progressive statistics for studies in sports medicine and exercise science.
Medicine & Science in Sports & Exercise,
41(1), 3–13.
https://doi.org/10.1249/MSS.0b013e31818cb278
5. Student [Gosset, W. S. (1908). The probable error of a mean.
Biometrika,
6(1), 1–25.
https://doi.org/10.2307/2331554
6. Cumming, G. (2014). The new statistics: Why and how.
Psychological Science,
25(1), 7–29.
https://doi.org/10.1177/0956797613504966
7. Lakens, D. (2013). Calculating and reporting effect sizes to facilitate cumulative science: A practical primer for t-tests and ANOVAs.
Frontiers in Psychology,
4, 863.
https://doi.org/10.3389/fpsyg.2013.00863
8. Field, A. (2018). Discovering statistics using IBM SPSS statistics (5th ed.). SAGE Publications.
9. Cumming, G. (2012). Understanding the new statistics: Effect sizes, confidence intervals, and meta-analysis. Routledge.
10. Maxwell, S. E., Delaney, H. D., & Kelley, K. (2018). Designing experiments and analyzing data: A model comparison perspective (3rd ed.). Routledge.
11. Thomas, L. (2015). How to estimate power and sample size.
Trauma Surgery & Acute Care Open,
1(1), e000005.
https://doi.org/10.1136/tsaco-2015-000005
12. Portney, L. G., & Watkins, M. P. (2020). Foundations of clinical research: Applications to practice.
13. Ruxton, G. D. (2006). The unequal variance t-test is an underused alternative to student’s t-test and the mann-whitney u test.
Behavioral Ecology,
17(4), 688–690.
https://doi.org/10.1093/beheco/ark016
14. Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.). Lawrence Erlbaum Associates.
15. Lumley, T., Diehr, P., Emerson, S., & Chen, L. (2002). The importance of the normality assumption in large public health data sets.
Annual Review of Public Health,
23, 151–169.
https://doi.org/10.1146/annurev.publhealth.23.100901.140546
16. Lakens, D. (2014). Performing high-powered studies efficiently with sequential analyses.
European Journal of Social Psychology,
44, 701–710.
https://doi.org/10.1002/ejsp.2023
17. Welch, B. L. (1947). The generalization of "student’s" problem when several different population variances are involved.
Biometrika,
34(1-2), 28–35.
https://doi.org/10.1093/biomet/34.1-2.28
18. Delacre, M., Lakens, D., & Leys, C. (2017). Why psychologists should by default use welch’s t-test instead of student’s t-test.
International Review of Social Psychology,
30(1), 92–101.
https://doi.org/10.5334/irsp.82
19. Zimmerman, D. W. (2004). A note on preliminary tests of equality of variances.
British Journal of Mathematical and Statistical Psychology,
57(1), 173–181.
https://doi.org/10.1348/000711004849222
20. Razali, N. M., & Wah, Y. B. (2011). Power comparisons of shapiro-wilk, kolmogorov-smirnov, lilliefors and anderson-darling tests. Journal of Statistical Modeling and Analytics, 2(1), 21–33.
21. Blanca, M. J., Alarcón, R., Arnau, J., Bono, R., & Bendayan, R. (2013). Non-normal data: Is ANOVA still a valid option?
Psicothema,
25(4), 552–557.
https://doi.org/10.7334/psicothema2013.552
22. Levene, H. (1960). Robust tests for equality of variances. Contributions to Probability and Statistics: Essays in Honor of Harold Hotelling, 278–292.
23. Conover, W. J. (1999). Practical nonparametric statistics.
24. Wilkinson, L., & Task Force on Statistical Inference. (1999). Statistical methods in psychology journals: Guidelines and explanations.
American Psychologist,
54(8), 594–604.
https://doi.org/10.1037/0003-066X.54.8.594
25. American Psychological Association. (2020). Publication manual of the american psychological association (7th ed.). American Psychological Association.
26. Kelley, K., & Preacher, K. J. (2012). On effect size.
Psychological Methods,
17(2), 137–152.
https://doi.org/10.1037/a0028086
27. Nakagawa, S., & Cuthill, I. C. (2007). Effect size, confidence interval and statistical significance: A practical guide for biologists.
Biological Reviews,
82, 591–605.
https://doi.org/10.1111/j.1469-185X.2007.00027.x
28. Schenker, N., & Gentleman, J. F. (2001). Judging statistical significance from confidence intervals.
The American Statistician,
55(3), 182–186.
https://doi.org/10.1198/000313001317098149
29. Button, K. S., Ioannidis, J. P. A., Mokrysz, C., Nosek, B. A., Flint, J., Robinson, E. S. J., & Munafò, M. R. (2013). Power failure: Why small sample size undermines the reliability of neuroscience.
Nature Reviews Neuroscience,
14, 365–376.
https://doi.org/10.1038/nrn3475
30. Lakens, D. (2014). Performing high-powered studies efficiently with sequential analyses.
European Review of Social Psychology,
25(1), 60–75.
https://doi.org/10.1080/10463283.2014.922662
31. Faul, F., Erdfelder, E., Lang, A.-G., & Buchner, A. (2007). G*power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences.
Behavior Research Methods,
39(2), 175–191.
https://doi.org/10.3758/BF03193146
32. Hoenig, J. M., & Heisey, D. M. (2001). ABCs of alpha, beta, delta, and epsilon.
Ecology,
82(12), 3369–3372.
https://doi.org/10.1890/0012-9658(2001)082[3369:AOABDE]2.0.CO;2
33. Senn, S. (2002). Letter to the editor: Cross-over trials in clinical research.
Statistics in Medicine,
21(19), 2843–2844.
https://doi.org/10.1002/sim.1097
34. Osborne, J. (2002). Notes on the use of data transformations.
Practical Assessment, Research & Evaluation,
8(6).
https://scholarworks.umass.edu/pare/vol8/iss1/6
35. Wilcox, R. R. (2017). Introduction to robust estimation and hypothesis testing (4th ed.). Academic Press.
36. Good, P. I. (2005). Permutation tests: A practical guide to resampling methods for testing hypotheses.
37. Agresti, A. (2003). Categorical data analysis.
38. Newcombe, R. G. (1998). Two-sided confidence intervals for the single proportion: Comparison of seven methods.
Statistics in Medicine,
17, 857–872.
https://doi.org/10.1002/(SICI)1097-0258(19980430)17:8<857::AID-SIM777>3.0.CO;2-E
39. Simmons, J. P., Nelson, L. D., & Simonsohn, U. (2011). False-positive psychology: Undisclosed flexibility in data collection and analysis allows presenting anything as significant.
Psychological Science,
22(11), 1359–1366.
https://doi.org/10.1177/0956797611417632
40. Head, M. L., Holman, L., Lanfear, R., Kahn, A. T., & Jennions, M. D. (2015). The extent and consequences of p-hacking in science.
PLoS Biology,
13(3), e1002106.
https://doi.org/10.1371/journal.pbio.1002106
41. Cohen, J. (1994). The earth is round (p < .05).
American Psychologist,
49(12), 997–1003.
https://doi.org/10.1037/0003-066X.49.12.997
42. Altman, D. G., & Bland, J. M. (2000). Statistics notes: The use of transformation when comparing two means.
BMJ,
312, 1153.
https://doi.org/10.1136/bmj.312.7039.1153
43. Wasserstein, R. L., & Lazar, N. A. (2016). The ASA statement on p-values: Context, process, and purpose.
The American Statistician,
70(2), 129–133.
https://doi.org/10.1080/00031305.2016.1154108
44. Ioannidis, J. P. A. (2005). Why most published research findings are false.
PLoS Medicine,
2(8), e124.
https://doi.org/10.1371/journal.pmed.0020124