pingouin.bayesfactor_binom#

pingouin.bayesfactor_binom(k, n, p=0.5, a=1, b=1)[source]#

Bayes factor of a binomial test with \(k\) successes, \(n\) trials and base probability \(p\). This means that the null hypothesis is that the probability is \(p\). It is compared against the alternative hypothesis that \(p\) is from the Beta distribution with parameters \((a, b)\). By default, both \(a\) and \(b\) are 1, making the alternative hypothesis equivalent to the uniform distribution, i.e., we are completely uninformed about \(p\).

Parameters:
kint

Number of successes.

nint

Number of trials.

pfloat

Base probability of success (range from 0 to 1).

afloat

The “a” parameter of the Beta distribution.

bfloat

The “b” parameter of the Beta distribution.

Returns:
bf10float

The Bayes Factor quantifies the evidence in favour of the alternative hypothesis, where the null hypothesis is that the random variable is binomially distributed with base probability \(p\).

See also

bayesfactor_pearson

Bayes Factor of a correlation

bayesfactor_ttest

Bayes Factor of a T-test

Notes

Adapted from a Matlab code found at anne-urai/Tools

The Bayes Factor is given by the formula below:

\[BF_{10} = \frac{\int_0^1 \binom{n}{k}g^k(1-g)^{n-k}} {\binom{n}{k} p^k (1-p)^{n-k}}\]

References

Examples

We want to determine if a coin if fair. After tossing the coin 200 times in a row, we report 115 heads (hereafter referred to as “successes”) and 85 tails (“failures”). The Bayes Factor can be easily computed using Pingouin:

>>> import pingouin as pg
>>> bf = float(pg.bayesfactor_binom(k=115, n=200, p=0.5))
>>> # Note that Pingouin returns the BF-alt by default.
>>> # BF-null is simply 1 / BF-alt
>>> print("BF-null: %.3f, BF-alt: %.3f" % (1 / bf, bf))
BF-null: 1.197, BF-alt: 0.835

Since the Bayes Factor of the null hypothesis (“the coin is fair”) is higher than the Bayes Factor of the alternative hypothesis (“the coin is not fair”), we can conclude that there is more evidence to support the fact that the coin is indeed fair. However, the strength of the evidence in favor of the null hypothesis (1.197) is “barely worth mentionning” according to Jeffreys’s rule of thumb.

Interestingly, a frequentist alternative to this test would give very different results. It can be performed using the scipy.stats.binom_test() function:

>>> from scipy.stats import binomtest
>>> result = binomtest(k=115, n=200, p=0.5)
>>> round(result.pvalue, 5)
0.04004

The binomial test rejects the null hypothesis that the coin is fair at the 5% significance level (p=0.04). Thus, whereas a frequentist hypothesis test would yield significant results at the 5% significance level, the Bayes factor indicates preference of the null hypothesis to the alternative hypothesis that we know nothing about p.

We can use a more informed alternative hypothesis too, if desirable. E.g., the original test using Beta(5, 4) as the alternative hypothesis:

>>> bf = pg.bayesfactor_binom(k=115, n=200, p=0.5, a=5, b=4)
>>> print("Bayes Factor: %.3f" % bf)
Bayes Factor: 1.930

Using a different base probability of successes:

>>> bf = pg.bayesfactor_binom(k=100, n=1000, p=0.1)
>>> print("Bayes Factor: %.3f" % bf)
Bayes Factor: 0.024