Kernel regression
In
In any nonparametric regression, the conditional expectation of a variable relative to a variable may be written:
where is an unknown function.
Nadaraya–Watson kernel regression
Nadaraya and Watson, both in 1964, proposed to estimate as a locally weighted average, using a kernel as a weighting function.[1][2][3] The Nadaraya–Watson estimator is:
where is a kernel with a bandwidth such that is of order at least 1, that is .
Derivation
Starting with the definition of conditional expectation,
we estimate the joint distributions f(x,y) and f(x) using kernel density estimation with a kernel K:
We get:
which is the Nadaraya–Watson estimator.
Priestley–Chao kernel estimator
where is the bandwidth (or smoothing parameter).
Gasser–Müller kernel estimator
where [4]
Example
![](http://upload.wikimedia.org/wikipedia/commons/thumb/c/ce/Cps71_lc_mean.png/250px-Cps71_lc_mean.png)
This example is based upon Canadian cross-section wage data consisting of a random sample taken from the 1971 Canadian Census Public Use Tapes for male individuals having common education (grade 13). There are 205 observations in total.[citation needed]
The figure to the right shows the estimated regression function using a second order Gaussian kernel along with asymptotic variability bounds.
Script for example
The following commands of the
npreg()
function to deliver optimal smoothing and to create the figure given above. These commands can be entered at the command prompt via cut and paste.
install.packages("np")
library(np) # non parametric library
data(cps71)
attach(cps71)
m <- npreg(logwage~age)
plot(m, plot.errors.method="asymptotic",
plot.errors.style="band",
ylim=c(11, 15.2))
points(age, logwage, cex=.25)
detach(cps71)
Related
According to
Statistical implementation
- GNU Octave mathematical program package
- Julia: KernelEstimator.jl
- MATLAB: A free MATLAB toolbox with implementation of kernel regression, kernel density estimation, kernel estimation of hazard function and many others is available on these pages (this toolbox is a part of the book [6]).
- Python: the
KernelReg
class for mixed data types in thestatsmodels.nonparametric
sub-package (includes other kernel density related classes), the package kernel_regression as an extension of scikit-learn (inefficient memory-wise, useful only for small datasets) - R: the function
npreg
of the np package can perform kernel regression.[7][8] - Stata: npregress, kernreg2
See also
References
- doi:10.1137/1109020.
- JSTOR 25049340.
- ISBN 0-521-41900-X.
- MR 0564251.
- ISBN 0-8050-7134-2.
- ISBN 978-981-4405-48-5.
- ^ np: Nonparametric kernel smoothing methods for mixed data types
- ISBN 978-1-4398-7343-4.
Further reading
- Henderson, Daniel J.; Parmeter, Christopher F. (2015). Applied Nonparametric Econometrics. Cambridge University Press. ISBN 978-1-107-01025-3.
- Li, Qi; Racine, Jeffrey S. (2007). Nonparametric Econometrics: Theory and Practice. Princeton University Press. ISBN 978-0-691-12161-1.
- Pagan, A.; Ullah, A. (1999). Nonparametric Econometrics. Cambridge University Press. ISBN 0-521-35564-8.
- Racine, Jeffrey S. (2019). An Introduction to the Advanced Theory and Practice of Nonparametric Econometrics: A Replicable Approach Using R. Cambridge University Press. ISBN 9781108483407.
- Simonoff, Jeffrey S. (1996). Smoothing Methods in Statistics. Springer. ISBN 0-387-94716-7.
External links
- Scale-adaptive kernel regression (with Matlab software).
- Tutorial of Kernel regression using spreadsheet (with Microsoft Excel).
- An online kernel regression demonstration Requires .NET 3.0 or later.
- Kernel regression with automatic bandwidth selection (with Python)