The posterior odds ratio is the ratio between the posterior probabilities of two events.
In Bayesian inference, it is used to compare different hypotheses or different models.
Let
,
and
be three events.
We can use Bayes' rule
to compute the
conditional
probabilities of
and
given
:
The ratio
is
called the posterior odds ratio of
and
.
The event
is often called evidence.
The ratio of prior
probabilitiesis
called prior odds ratio.
Finally, the ratio of
likelihoodsis
called Bayes factor.
Therefore,
The interpretation of the posterior odds ratio
is pretty simple:
if the ratio is larger than 1, then
is more likely than
after conditioning on the evidence
;
if the ratio is smaller than 1, then
is less likely than
conditional on
.
Moreover, the ratio tells us exactly how much more likely
is than
.
Typically, in Bayesian inference, the evidence
is some observed data, and
and
are two statistical models, that
is, two sets of probability distributions that could have generated the data.
In other words, the posterior odds ratio between
and
quantifies how much more likely
is than
,
based on the prior and the evidence provided by the data.
Note that the two sets of probability distributions can be singletons, that is, the comparison can involve single probability distributions.
Let us denote the data by
(instead of
)
and probability density
(or mass) functions by
,
as we did in previous lectures.
When there are two sets of parametric models, the priors are usually assigned in a hierarchical fashion.
In other words, we specify:
the probabilities of the two models,
and
;
two parametrized likelihoods,
and
;
two prior distributions over the parameters
of the two models,
and
.
The posterior odds ratio is
The two quantities
and
are the so-called prior predictive distributions or marginal
likelihoods.
If we are dealing with probability densities, the marginal likelihoods are
obtained by computing two
integrals:
This shows an important fact (see, e.g., O'Hagan 2006):
in a comparison of two hierarchical models, the value of the Bayes
factor
depends not only on the evidence provided by the data, but also on the prior
distributions of the parameters.
Therefore, a Bayes factor can incorporate both objective (the data) and subjective (the priors) information.
Although the posterior odds ratio can be used to neatly compare two models by using both prior information and the evidence provided by the data, statisticians often prefer to discuss and report Bayes factors.
Indeed, most of the books and papers that discuss Bayesian model comparisons (see, e.g., the marvellous introduction by Williams et al. 2017) focus almost exclusively on the properties of Bayes factors.
This is probably due to the fact that Bayes factors offer a more objective means of comparing models, by factoring out the more subjective prior odds ratio.
O'Hagan, T. (2006) Bayes factors, Significance, 3, 184-186.
Williams, M. N., Baath, R. A. and Philipp, M. C. (2017). Using Bayes factors to test hypotheses in developmental research, Research in Human Development, 14, 321-337.
Please cite as:
Taboga, Marco (2021). "Posterior odds ratio", Lectures on probability theory and mathematical statistics. Kindle Direct Publishing. Online appendix. https://www.statlect.com/fundamentals-of-statistics/posterior-odds-ratio.
Most of the learning materials found on this website are now available in a traditional textbook format.