Additive model
In statistics, an additive model (AM) is a nonparametric regression method. It was suggested by Jerome H. Friedman and Werner Stuetzle (1981),[1] and is an essential part of the ACE algorithm. The AM uses a one dimensional smoother to build a restricted class of nonparametric regression models. Because of this, it is less affected by the curse of dimensionality than e.g. a p-dimensional smoother. Furthermore, the AM is more flexible than a standard linear model, while being more interpretable than a general regression surface at the cost of approximation errors. Problems with AM include model selection, overfitting, and multicollinearity.
Description
Given a data set of n statistical units, where represent predictors and is the outcome, the additive model takes the form
or
Where , and . The functions are unknown smooth functions fit from the data. Fitting the AM (i.e. the functions ) can be done using the backfitting algorithm proposed by Andreas Buja, Trevor Hastie and Robert Tibshirani (1989).[2]
See also
- Generalized additive model
- Backfitting algorithm
- Alternating conditional expectation model
- Projection pursuit regression
- Generalized additive model for location, scale, and shape (GAMLSS)
- Median polish
References
- ^ Friedman, J. H.; Stuetzle, W. (1981). "Projection Pursuit Regression". J. Amer. Statist. Assoc. 76 (376): 817–823. doi:10.1080/01621459.1981.10477729.
- ^ Buja, A.; Hastie, T.; Tibshirani, R. (1989). "Linear Smoothers and Additive Models". Ann. Stat. 17 (2): 453–555. JSTOR 2241560.
Further reading
- Breiman, L.; Friedman, J. H. (1985). "Estimating Optimal Transformations for Multiple Regression and Correlation". J. Amer. Statist. Assoc. 80 (391): 580–598. doi:10.1080/01621459.1985.10478157.