Reconceptualizing the p -value from a likelihood ratio test: a probabilistic pairwise comparison of models based on Kullback-Leibler discrepancy measures.
Benjamin RiedleAndrew A NeathJoseph E CavanaughPublished in: Journal of applied statistics (2020)
Discrepancy measures are often employed in problems involving the selection and assessment of statistical models. A discrepancy gauges the separation between a fitted candidate model and the underlying generating model. In this work, we consider pairwise comparisons of fitted models based on a probabilistic evaluation of the ordering of the constituent discrepancies. An estimator of the probability is derived using the bootstrap. In the framework of hypothesis testing, nested models are often compared on the basis of the p -value. Specifically, the simpler null model is favored unless the p -value is sufficiently small, in which case the null model is rejected and the more general alternative model is retained. Using suitably defined discrepancy measures, we mathematically show that, in general settings, the likelihood ratio test p -value is approximated by the bootstrapped discrepancy comparison probability (BDCP). We argue that the connection between the p -value and the BDCP leads to potentially new insights regarding the utility and limitations of the p -value. The BDCP framework also facilitates discrepancy-based inferences in settings beyond the limited confines of nested model hypothesis testing.
Keyphrases