Draft:Universal approximation theorem
In the mathematical theory of artificial neural networks, the universal approximation theorem states that feedforward neural networks constructed of artificial neurons can approximate real-valued continuous functions on compact subsets of , to arbitrary accuracy. The theorem thus implies that simple neural networks can in principle be applied to nearly any problem, as they can approximate essentially any function of interest.
Early versions of the theorem considered networks of arbitrary width. In particular these were considered by George Cybenko in 1989,[1] Kurt Hornik in 1991,[2] and Moshe Leshno et al in 1993.[3] A simple general formulation was given by Allan Pinkus in 1999,[4] which is the version stated here. Later versions considered the 'dual' problem for networks of arbitrary depth. In particular these were considered by Zhou Lu et al in 2017 [5] and Boris Hanin and Mark Sellke in 2018.[6] A simple general formulation was given by Patrick Kidger and Terry Lyons in 2020 [7], which is the version stated here.
Several extensions of the theorem exist, such as to discontinuous activation functions, alternative network architectures, other topologies, and noncompact domains. [3][7][8].
Formal statements
The classical arbitrary width version of the theorem may be stated as follows.
Universal approximation theorem; arbitrary width.[4] Let be any continuous function (called the activation function). Let be compact. The space of real-valued continuous functions on is denoted by . Let denote the space of functions of the form
for all integers , real constants and real vectors for .
Then, if and only if is nonpolynomial, the following statement is true: given any and any , there exists such that
for all .
In other words, is dense in with respect to the uniform norm if and only if is nonpolynomial.
This theorem extends straightforwardly to networks with any fixed number of hidden layers: the theorem implies that the first layer can approximate any desired function, and that later layers can approximate the identity function. Thus any fixed-depth network may approximate any continuous function, and this version of the theorem applies to networks with bounded depth and arbitrary width.
The 'dual' version of the the theorem considers networks of bounded width and arbitrary depth. Unlike the previous theorem, it gives sufficient but not necessary conditions.
Universal approximation theorem; arbitrary depth.[7] Let be any nonaffine continuous function which is continuously differentiable at at least one point, with nonzero derivative at that point. Let be compact. The space of real vector-valued continuous functions on is denoted by . Let denote the space of feedforward neural networks with input neurons, output neurons, and an arbitrary number of hidden layers each with neurons, such that every hidden neuron has activation function and every output neuron has the identity as its activation function. Then given any and any , there exists such that
for all .
In other words, is dense in with respect to the uniform norm.
Certain necessary conditions for the bounded width, arbitrary depth case have been established, but there is still a gap between the known sufficient and necessary conditions.[5][6][9].
The arbitrary depth case includes polynomial activation functions, which are specifically excluded from the arbitrary width case. This is an example of a qualitative difference between (particular interpretations of) deep and shallow neural networks.
See also
- Kolmogorov–Arnold representation theorem
- No free lunch theorem
- Stone–Weierstrass theorem
- Fourier series
References
- ^ Cybenko, G. (1989) "Approximations by superpositions of sigmoidal functions", Mathematics of Control, Signals, and Systems, 2(4), 303–314. doi:10.1007/BF02551274
- ^ Hornik, Kurt (1991). "Approximation capabilities of multilayer feedforward networks". Neural Networks. 4 (2): 251–257. doi:10.1016/0893-6080(91)90009-T.
- ^ a b Leshno, Moshe; Lin, Vladimir Ya.; Pinkus, Allan; Schocken, Shimon (January 1993). "Multilayer feedforward networks with a nonpolynomial activation function can approximate any function". Neural Networks. 6 (6): 861–867. doi:10.1016/S0893-6080(05)80131-5.
- ^ a b Pinkus, Allan (January 1999). "Approximation theory of the MLP model in neural networks". Acta Numerica. 8: 143–195. doi:10.1017/S0962492900002919.
- ^ a b Lu, Zhou; Pu, Homgming; Wang, Feicheng; Hu, Zhiqiang; Wang, Liwei. "The Expressive Power of Neural Networks: A View from the Width". Advances in Neural Information Processing Systems 30. Curran Associates, Inc.: 6231–6239.
- ^ a b Hanin, Boris; Sellke, Mark (March 2018). "Approximating Continuous Functions by ReLU Nets of Minimal Width". arXiv:1710.11278.
- ^ a b c Kidger, Patrick; Lyons, Terry (July 2020). Universal Approximation with Deep Narrow Networks. Conference on Learning Theory.
- ^ Lin, Hongzhou; Jegelka, Stefanie (2018). ResNet with one-neuron hidden layers is a Universal Approximator. Advances in Neural Information Processing Systems 30. Curran Associates, Inc. pp. 6169–6178.
- ^ Johnson, Jesse (2019). Deep, Skinny Neural Networks are not Universal Approximators. International Conference on Learning Representations.
Universal approximation theorem
![]() | Review waiting, please be patient.
This may take 8 weeks or more, since drafts are reviewed in no specific order. There are 989 pending submissions waiting for review.
Where to get help
How to improve a draft
You can also browse Wikipedia:Featured articles and Wikipedia:Good articles to find examples of Wikipedia's best writing on topics similar to your proposed article. Improving your odds of a speedy review To improve your odds of a faster review, tag your draft with relevant WikiProject tags using the button below. This will let reviewers know a new draft has been submitted in their area of interest. For instance, if you wrote about a female astronomer, you would want to add the Biography, Astronomy, and Women scientists tags. Editor resources
Reviewer tools
|
![]() | The WikiProject banner below should be moved to this redirect's talk page. If this is a demonstration of the template, please set the parameter |category=no to prevent this page being miscategorised. |
![]() | Mathematics Redirect‑class | ||||||
|
![]() | The WikiProject banner below should be moved to this redirect's talk page. If this is a demonstration of the template, please set the parameter |category=no to prevent this page being miscategorised. |
![]() | Computer science Redirect‑class | |||||||||||||
|
First time making a Wikipedia edit; I think (hope) I'm doing the right thing here. Looking to update the "Universal approximation theorem" article.
Essential changes:
- Updated out-of-date results (by 20 years...)
- Now describe the arbitrary width and depth cases using similar language to each other.
- General tidy-up, so the article is now shorter.
- Added additional references.
COI:
- I wrote one of the papers that's now being discussed. It generalises+simplifies previous work, so I think it's fair.
- Theorems in discrete mathematics
- Artificial neural networks
- Network architecture
- Networks
- Pending AfC submissions
- AfC pending submissions by age/Very old
- AfC submissions with the same name as existing articles
- AfC submissions by date/25 June 2020
- Redirect-Class mathematics pages
- NA-priority mathematics pages
- Redirect-Class Computer science pages
- NA-importance Computer science pages
- WikiProject Computer science articles