Blending Bayesian and Classical Tools to Define Optimal Sample-Size-Dependent Significance Levels |
| |
Authors: | Mark Andrew Gannon Carlos Alberto de Bragança Pereira Adriano Polpo |
| |
Institution: | 1. Institute of Mathematics and Statistics, University of S?o Paulo, S?o Paulo, Brazil;2. mark@ime.usp.br;4. Instituto de Matemática Aplicada, Universidade Federal de Mato Grosso do Sul, Campo Grande, Brazil;5. Department of Statistics, Federal University of S?o Carlos |
| |
Abstract: | ABSTRACTThis article argues that researchers do not need to completely abandon the p-value, the best-known significance index, but should instead stop using significance levels that do not depend on sample sizes. A testing procedure is developed using a mixture of frequentist and Bayesian tools, with a significance level that is a function of sample size, obtained from a generalized form of the Neyman–Pearson Lemma that minimizes a linear combination of α, the probability of rejecting a true null hypothesis, and β, the probability of failing to reject a false null, instead of fixing α and minimizing β. The resulting hypothesis tests do not violate the Likelihood Principle and do not require any constraints on the dimensionalities of the sample space and parameter space. The procedure includes an ordering of the entire sample space and uses predictive probability (density) functions, allowing for testing of both simple and compound hypotheses. Accessible examples are presented to highlight specific characteristics of the new tests. |
| |
Keywords: | Hardy–Weinberg equilibrium Neyman–Pearson lemma Predictive distribution Significance test |
|
|