P-rep

Source: Wikipedia, the free encyclopedia.

In

hypothesis testing, p-rep or prep has been proposed as a statistical alternative to the classic p-value.[1]
Whereas a p-value is the probability of obtaining a result under the null hypothesis, p-rep purports to compute the probability of replicating an effect. The derivation of p-rep contained significant mathematical errors.

For a while, the Association for Psychological Science recommended that articles submitted to Psychological Science and their other journals report p-rep rather than the classic p-value,[2] but this is no longer the case.[3]

Calculation

Approximation from p

P-rep function (in log scale)

The value of the p-rep (prep) can be approximated based on the p-value (p) as follows:

The above applies for one-tailed distributions.

Criticism

The fact that the p-rep has a

statistical hypothesis testing
.

Among the criticisms of p-rep is the fact that while it attempts to estimate replicability, it ignores results from other studies which can accurately guide this estimate.[4] For example, an experiment on some unlikely paranormal phenomenon may yield a p-rep of 0.75. Most people would still not conclude the probability of a replication was 75%. Rather, they would conclude it is much closer to 0: Extraordinary claims require extraordinary evidence, and p-rep ignores this. Because of this, p-rep may in fact be harder to interpret than a classical p-value. The fact that p-rep requires assumptions about prior probabilities for it to be valid makes its interpretation complex. Killeen argues that new results should be evaluated in their own right, without the "burden of history", with flat priors: that is what p-rep yields. A more pragmatic estimate of replicability would include prior knowledge, via, for instance, meta-analysis.

Critics have also underscored mathematical errors in the original Killeen paper. For example, the formula relating the

independent,[5] criticisms addressed in Killeen's rejoinder.[6]

A further criticism of the p-rep statistic involves the logic of experimentation. The scientific value of replicable data lies in the adequate accounting for previously unmeasured factors (e.g., unmeasured participant variables,

experimenter's bias, etc.), The idea that a single study can capture a logical likelihood of such unmeasured factors affecting the outcome, and thus the likelihood of replicability, is a logical fallacy.[citation needed
]

References

  1. .
  2. ^ archived version of "Psychological Science Journal, Author Guidelines"
  3. ^ Psychological Science Journal, Author Guidelines.
  4. ^ Macdonald, R. R. (2005) "Why Replication Probabilities Depend on Prior Probability Distributions" Psychological Science, 2005, 16, 1006–1008 [https://psycnet.apa.org/record/2005-15678-016
  5. ^ "p-rep" at Pro Bono Statistics
  6. ^ Killeen, P. R. (2005)" Replicability, Confidence, and Priors", Psychological Science, 2005, 16, 1009–1012 [1]

External links

This page is based on the copyrighted Wikipedia article: P-rep. Articles is available under the CC BY-SA 3.0 license; additional terms may apply.Privacy Policy