Let’s Poke a Pizza: A New Cartoon to Explain the Strength of Evidence in a Bayes Factor

In a previous post we discussed the interpretation of the strength of evidence coming from a Bayes factor. For concreteness, let’s take a binomial data set and suppose we have encountered 62 successes and 38 failures in a sample of 100 trials. We can easily enter these data directly using the JASP Summary Stats Module:

 
 
 
 
 
 
 
 
 

The JASP output shows that for these data, the classical two-sided p-value is .021 (in most applications, researchers would feel entitled to “reject the null hypothesis that the result is due to chance”). The Bayes factor indicates that the data are 2.2 times more likely under the alternative hypothesis H1 than under the null hypothesis H0. The null hypothesis H0 is specified by the number after “Test value”; in this case, H0 says that the binomial rate parameter equals 0.5. The alternative hypothesis is specified by a beta distribution with “parameter a” set to 1 and “parameter b” set to 1 (other distributions may be specified, but that is the topic of a different post). To understand the result more deeply, let’s tick the box “Prior and posterior”, which produces the following graph:

The dotted line show the prior distribution (i.e., the beta distribution with a=1 and b=1) under the alternative hypothesis; the solid line show the posterior distribution. With 62 successes and 38 failures we know that the posterior is also a beta distribution with parameters a’ = 1+62 and b’ = 1+38, but such analytical niceties are not important for the interpretation of the result. What we can see from the figure is that, assuming we started by specifying a beta(1,1) distribution for the rate \theta (“all values are equally likely a priori”), we have learned that the most likely values are somewhere in between 0.52 and 0.71. However, in most hypothesis testing scenarios we wish to provide statistical evidence to bolster the claim that the effect exists at all. And in order to cast doubt on the skeptic’s position (which is “take your research elsewhere, we don’t buy it, you are just interpreting noise”) we need to take that position seriously from the start. The alternative hypothesis with its smooth prior distribution does not assign any special value to the skeptics’ hypothesis which states that \theta = .5, and therefore it would be a mistake to use the posterior distribution under H1 to conclude anything about the validity of H0, tempting as it may be.

Enter the Bayes factor, which considers the predictive adequacy of the skeptics’ H0 (i.e., \theta = 0.5) versus the proponents’ H1 (here: \theta \sim \text{Beta}(a=1, b=1)). And the results on top of the figure show that BF10 = 2.2. How to interpret this number when one is not comfortable with likelihood ratios? One way is to use a heuristic, for instance the rule of thumb categorization proposed by Harold Jeffreys. In the first, 1939, edition of “Theory of Probability” (in my opinion, the best book on statistics, and by a landslide), Jeffreys proposed to deem any Bayes factor lower than \sqrt{10} \approx 3.16 “not worth more than a bare comment” (p. 357). Another way is to transform the Bayes factor to a probability, assuming H0 and H1 are equally probable a priori. For instance, H1 was at 50% before seeing the data; after seeing the data, that probability has increased to 2.2 / 3.2 \approx 69%, leaving a substantial 31% for H0. But even these percentages may not drive home, in an intuitive, visceral way, how little evidence this really is. In order to help people interpret the percentages, JASP presents a “pizza plot” on top of the graph.

In the pizza plot, the red “pepperoni” slice corresponds to the 69% for H1, and the white “mozzarella” slice corresponds to the 31% for H0. As mentioned in the earlier post: “To feel how much evidence this is, we may mentally execute the “Pizza-poke Assessment of the Weight of evidence” (PAW): if you pretend to poke your finger blindly in the pizza, how surprised are you if that finger returns covered in the non-dominant topping?” In this case, you poke your finger onto the pizza and it comes back covered in mozzarella. Your lack of imagined surprise means that you should be wary of interpreting the data as strong evidence against the null hypothesis. More generally, note how much more information is communicated in the JASP graph compared to the standard frequentist conclusion “p=.021” (and, when you’re lucky, a 95% confidence interval on the parameter).

In the cartoon library on this blog you will encounter two cartoons that try to explain the pizza plot logic. One is about throwing darts and the other is about a spinner. During my workshops, however, I found that I discussed the pizza plot in terms of a pizza rather than darts or spinners. Finally I decided that what was needed to explain the pizza plot is a cartoon featuring pizzas. So here it is, courtesy of the fabulous Viktor Beekman:

References

Jeffreys, H. (1939). Theory of Probability. Oxford: Oxford University Press.

Ly, A., Raj, A., Marsman, M., Etz, A., & Wagenmakers, E.-J. (in press). Bayesian reanalyses from summary statistics: A guide for academic consumers. Advances in Methods and Practices in Psychological Science.

About The Author

Eric-Jan Wagenmakers

Eric-Jan (EJ) Wagenmakers is professor at the Psychological Methods Group at the University of Amsterdam.