Lindley ParadoxEdit
The Lindley paradox is a well-known phenomenon in statistics that highlights a striking divergence between Bayesian and frequentist approaches to hypothesis testing. Named for Dennis Lindley, with ties to the earlier work of Harold Jeffreys, the paradox shows that, under certain reasonable-sounding assumptions, the Bayesian evaluation of evidence can favor the null hypothesis even when a conventional frequentist test would declare strong rejection. The result is most commonly illustrated in the context of testing a point null against a diffuse alternative, and it underscores the important point that how one measures evidence depends crucially on the chosen framework and prior information.
In practice, the Lindley paradox has prompted careful examination of how researchers conduct hypothesis tests, how priors are specified, and how conclusions should be interpreted when large data sets are involved. It is often discussed under the umbrella of the Jeffreys–Lindley paradox, and it continues to influence debates about when and how to rely on p-values, Bayes factors, and other tools for deciding between competing models or hypotheses. The paradox is not a single computation but a family of scenarios in which Bayesian evidence can appear to contradict frequentist signals, especially as sample size grows or priors on the alternative become diffuse.
Overview
Setup and intuition
- The basic problem is to compare two competing statements about a parameter: a null hypothesis H0 (often a simple point claim, such as θ = θ0) and an alternative H1 (a broader, composite claim about θ).
- In a Bayesian analysis, one assigns prior probabilities to the hypotheses and a prior distribution over θ under the alternative. The strength of evidence for H0 or H1 is measured by the Bayes factor, which is the ratio of marginal likelihoods: BF10 = p(data | H1) / p(data | H0).
- A key feature of the Lindley paradox is that if the prior on θ under H1 is diffuse (spread over a wide range of possible values), the marginal likelihood p(data | H1) can be comparatively small in light of the observed data, even when the data produce small p-values under the frequentist test. As a result, the Bayes factor can favor H0 despite seemingly strong frequentist evidence against it.
A simple illustrative model
- Consider observations that are well modeled by a normal distribution with known variance, and test H0: μ = 0 versus H1: μ ≠ 0.
- Under the Bayesian analysis, specify a prior for μ under H1, for example μ ~ N(0, τ^2) with τ^2 large (a diffuse prior).
- The marginal likelihood under H0 is p(data | H0) = N(data; 0, σ^2), while under H1 it is p(data | H1) = ∫ N(data; μ, σ^2) N(μ; 0, τ^2) dμ, which simplifies to a normal with larger variance due to the integration over μ.
- The Bayes factor BF10 compares these two quantities. For small observed effects (data near μ = 0), the extra spread in the alternative can reduce p(data | H1) enough that BF10 < 1, hence the data favor H0 in the Bayesian sense. Conversely, large effects can push BF10 above 1 in favor of H1.
The paradox in context
- The paradox is particularly salient in large samples. As sample size grows, frequentist tests can yield very small p-values for even modest departures from the null, pushing toward rejection of H0 in a p-value sense.
- Yet with a diffuse prior on the alternative, the Bayesian evidence can still favor H0 for outcomes that are not exceedingly surprising under the null. This creates a situation where p-values and Bayes factors point in opposite directions.
- The core message is that p-values measure the compatibility of the data with the null under a long-run frequency interpretation; Bayes factors measure how much the data update belief in H0 relative to H1 given prior information. When priors on the alternative are broad, the data may not sufficiently overcome the prior mass allocated away from the null, especially for incidental fluctuations in large samples.
Practical implications
- The Lindley paradox invites researchers to consider prior information more explicitly when making decisions about hypotheses, rather than relying solely on p-values or on Bayes factors trained on uninformative priors.
- It has motivated the development and use of alternative Bayesian tools for model comparison, such as intrinsic and fractional Bayes factors, which aim to temper the impact of diffuse priors and small-sample behavior.
- It also informs discussions about stopping rules and data peeking. Because the paradox depends on how evidence is aggregated and interpreted, practitioners sometimes adopt robustness analyses or calibration procedures to assess how sensitive conclusions are to prior choices and data collection plans.
Mathematical formulation and examples
- In a canonical normal-means setup, suppose X ~ N(μ, σ^2/n), and we test H0: μ = 0 versus H1: μ ≠ 0. If under H1 we place a prior μ ~ N(0, τ^2) with τ^2 large, then the Bayes factor BF01 (favoring H0) can be substantial when the observed mean is close to zero, even if the frequentist z-statistic would suggest rejection of H0 for sufficiently large n and nonzero μ.
- More generally, the Bayes factor depends on the prior variance under H1 and on the observed data. A closed-form expression in this normal-conjugate case helps illuminate why the paradox arises: integrating the likelihood against a diffuse prior reduces the marginal support for H1 relative to a precise hypothesis H0, particularly when effects are small.
Implications for statistical practice
- The Lindley paradox reinforces a central theme in modern statistics: the meaning of "evidence" depends on the framework and prior information used to quantify that evidence.
- It has contributed to a broader movement toward model-based assessment of uncertainty, where prior knowledge and model structure are explicitly accounted for rather than relied on p-values alone.
- Critics of automatic or default priors have used the paradox to argue for careful prior elicitation, robustness checks, and sensitivity analyses. Proponents have used it to encourage a more nuanced view of model comparison and to promote Bayesian techniques that explicitly incorporate domain knowledge.
- In response, some practitioners turn to calibrated or informative priors, intrinsic or fractional Bayes factors, or to complementary metrics such as the Bayes factor with sensitivity analyses, to reconcile Bayesian and frequentist perspectives.
Controversies and debates
- The core controversy centers on the interpretation and role of priors. Critics argue that the paradox shows how fragile Bayesian conclusions can be when priors are diffuse or uninformative. Proponents counter that priors encode genuine prior information and that, when used thoughtfully, they produce coherent updates to belief that can complement frequentist measures.
- A related debate concerns the practical meaning of evidence in scientific inference. Supporters of Bayesian methods contend that Bayes factors provide a principled framework for weighing competing hypotheses in light of prior knowledge, while skeptics point to the subjectivity inherent in prior choices.
- The Lindley paradox intersects with broader methodological discussions, such as the appropriate handling of optional stopping, the interpretation of p-values in large samples, and the development of robust model comparison criteria. These discussions have led to a richer ecosystem of tools, including alternative Bayes factors, robust priors, and model-averaged approaches.