Bonferroni correction

Source: Wikipedia, the free encyclopedia.

In statistics, the Bonferroni correction is a method to counteract the multiple comparisons problem.

Background

The method is named for its use of the

Bonferroni inequalities.[1]
An extension of the method to

Type I error) increases.[3]

The Bonferroni correction compensates for that increase by testing each individual hypothesis at a significance level of , where is the desired overall alpha level and is the number of hypotheses.[4] For example, if a trial is testing hypotheses with a desired overall , then the Bonferroni correction would test each individual hypothesis at . Similarly, when constructing confidence intervals for parameters, each individual confidence interval can be computed at the confidence level to achieve an overall confidence level of .

The Bonferroni correction can also be applied as a p-value adjustment: Using that approach, instead of adjusting the alpha level, each p-value is multiplied by the number of tests (with adjusted p-values that exceed 1 then being reduced to 1), and the alpha level is left unchanged. The significance decisions reached will be the same as when using the alpha-level adjustment approach.

Definition

Let be a family of null hypotheses and let be their corresponding p-values. Let be the total number of null hypotheses, and let be the number of true null hypotheses (which is presumably unknown to the researcher). The family-wise error rate (FWER) is the probability of rejecting at least one true , that is, of making at least one

type I error
. The Bonferroni correction rejects the null hypothesis for each , thereby controlling the FWER at . Proof of this control follows from Boole's inequality, as follows:

This control does not require any assumptions about dependence among the p-values or about how many of the null hypotheses are true.[5]

Extensions

Generalization

Rather than testing each hypothesis at the level, the hypotheses may be tested at any other combination of levels that add up to , provided that the level of each test is decided before looking at the data.[6] For example, for two hypothesis tests, an overall of 0.05 could be maintained by conducting one test at 0.04 and the other at 0.01.

Confidence intervals

The procedure proposed by Dunn

confidence intervals
. If one establishes confidence intervals, and wishes to have an overall confidence level of , each individual confidence interval can be adjusted to the level of .[2]

Continuous problems

When searching for a signal in a continuous parameter space there can also be a problem of multiple comparisons, or look-elsewhere effect. For example, a physicist might be looking to discover a particle of unknown mass by considering a large range of masses; this was the case during the Nobel Prize winning detection of the Higgs boson. In such cases, one can apply a continuous generalization of the Bonferroni correction by employing Bayesian logic to relate the effective number of trials, , to the prior-to-posterior volume ratio.[7]

Alternatives

There are alternative ways to control the

expected number of Type I errors per family (the per-family Type I error rate).[8]

Criticism

With respect to FWER control, the Bonferroni correction can be conservative if there are a large number of tests and/or the test statistics are positively correlated.[9]

The correction comes at the cost of increasing the probability of producing

statistical power.[10][9] There is not a definitive consensus on how to define a family in all cases, and adjusted test results may vary depending on the number of tests included in the family of hypotheses.[citation needed] Such criticisms apply to FWER
control in general, and are not specific to the Bonferroni correction.

References

  1. ^ Bonferroni, C. E., Teoria statistica delle classi e calcolo delle probabilità, Pubblicazioni del R Istituto Superiore di Scienze Economiche e Commerciali di Firenze 1936
  2. ^ .
  3. .
  4. .
  5. .
  6. .
  7. .
  8. .
  9. ^ .
  10. .

External links