Goodness of fit
This article needs additional citations for verification. (January 2018) |
Part of a series on |
Regression analysis |
---|
Models |
Estimation |
|
Background |
|
The goodness of fit of a
Fit of distributions
In assessing whether a given distribution is suited to a data-set, the following tests and their underlying measures of fit can be used:
- Bayesian information criterion
- Kolmogorov–Smirnov test
- Cramér–von Mises criterion
- Anderson–Darling test
- Berk-Jones tests[1][2]
- Shapiro–Wilk test
- Chi-squared test
- Akaike information criterion
- Hosmer–Lemeshow test
- Kuiper's test
- Kernelized Stein discrepancy[3][4]
- Zhang's ZK, ZC and ZA tests[5]
- Moran test
- Density Based Empirical Likelihood Ratio tests[6]
Regression analysis
In regression analysis, more specifically regression validation, the following topics relate to goodness of fit:
- Coefficient of determination (the R-squared measure of goodness of fit);
- Lack-of-fit sum of squares;
- Mallows's Cp criterion
- Prediction error
- Reduced chi-square
Categorical data
The following are examples that arise in the context of
Pearson's chi-square test
- Oi = an observed count for bin i
- Ei = an expected count for bin i, asserted by the null hypothesis.
The expected frequency is calculated by:
- F = the cumulative distribution function for the probability distribution being tested.
- Yu = the upper limit for class i,
- Yl = the lower limit for class i, and
- N = the sample size
The resulting value can be compared with a
Binomial case
A binomial experiment is a sequence of independent trials in which the trials can result in one of two outcomes, success or failure. There are n trials each with probability of success, denoted by p. Provided that npi ≫ 1 for every i (where i = 1, 2, ..., k), then
This has approximately a chi-square distribution with k − 1 degrees of freedom. The fact that there are k − 1 degrees of freedom is a consequence of the restriction . We know there are k observed cell counts, however, once any k − 1 are known, the remaining one is uniquely determined. Basically, one can say, there are only k − 1 freely determined cell counts, thus k − 1 degrees of freedom.
G-test
The general formula for G is
where and are the same as for the chi-square test, denotes the natural logarithm, and the sum is taken over all non-empty cells. Furthermore, the total observed count should be equal to the total expected count:
G-tests have been recommended at least since the 1981 edition of the popular statistics textbook by Robert R. Sokal and F. James Rohlf.[8]
See also
- All models are wrong
- Deviance (statistics) (related to GLM)
- Overfitting
- Statistical model validation
- Theil–Sen estimator
References
- .
- .
- ^ Liu, Qiang; Lee, Jason; Jordan, Michael (20 June 2016). "A Kernelized Stein Discrepancy for Goodness-of-fit Tests". Proceedings of the 33rd International Conference on Machine Learning. The 33rd International Conference on Machine Learning. New York, New York, USA: Proceedings of Machine Learning Research. pp. 276–284.
- ^ Chwialkowski, Kacper; Strathmann, Heiko; Gretton, Arthur (20 June 2016). "A Kernel Test of Goodness of Fit". Proceedings of the 33rd International Conference on Machine Learning. The 33rd International Conference on Machine Learning. New York, New York, USA: Proceedings of Machine Learning Research. pp. 2606–2615.
- . Retrieved 5 November 2018.
- .
- ^ McDonald, J.H. (2014). "G–test of goodness-of-fit". Handbook of Biological Statistics (Third ed.). Baltimore, Maryland: Sparky House Publishing. pp. 53–58.
- ISBN 0-7167-2411-1.
Further reading
- Huber-Carol, C.; Balakrishnan, N.; Nikulin, M. S.; Mesbah, M., eds. (2002), Goodness-of-Fit Tests and Model Validity, Springer
- Ingster, Yu. I.; Suslina, I. A. (2003), Nonparametric Goodness-of-Fit Testing Under Gaussian Models, Springer
- Rayner, J. C. W.; Thas, O.; Best, D. J. (2009), Smooth Tests of Goodness of Fit (2nd ed.), Wiley
- Vexler, Albert; Gurevich, Gregory (2010), "Empirical likelihood ratios applied to goodness-of-fit tests based on sample entropy",