P-Laplacian

From formulasearchengine
Revision as of 19:00, 22 April 2013 by en>RayAYang (External links: this counts as further reading instead)
Jump to navigation Jump to search

In the mathematical theory of neural networks, the universal approximation theorem states[1] that a feed-forward network with a single hidden layer containing a finite number of neurons, the simplest form of the multilayer perceptron, is a universal approximator among continuous functions on compact subsets of Rn, under mild assumptions on the activation function.

One of the first versions of the theorem was proved by George Cybenko in 1989 for sigmoid activation functions.[2]

Kurt Hornik showed in 1991[3] that it is not the specific choice of the activation function, but rather the multilayer feedforward architecture itself which gives neural networks the potential of being universal approximators. The output units are always assumed to be linear. For notational convenience, only the single output case will be shown. The general case can easily be deduced from the single output case.

Formal statement

The theorem[2][3][4][5] in mathematical terms:

Let φ(·) be a nonconstant, bounded, and monotonically-increasing continuous function. Let Im denote the m-dimensional unit hypercube [0,1]m. The space of continuous functions on Im is denoted by C(Im). Then, given any function fC(Im) and є > 0, there exist an integer N and real constants αi, biR, wiRm, where i = 1, ..., N such that we may define:

F(x)=i=1Nαiφ(wiTx+bi)

as an approximate realization of the function f where f is independent of φ; that is,

|F(x)f(x)|<ε

for all xIm. In other words, functions of the form F(x) are dense in C(Im).

References

43 year old Petroleum Engineer Harry from Deep River, usually spends time with hobbies and interests like renting movies, property developers in singapore new condominium and vehicle racing. Constantly enjoys going to destinations like Camino Real de Tierra Adentro.


Template:Applied-math-stub

  1. Balázs Csanád Csáji. Approximation with Artificial Neural Networks; Faculty of Sciences; Eötvös Loránd University, Hungary
  2. 2.0 2.1 Cybenko., G. (1989) "Approximations by superpositions of sigmoidal functions", Mathematics of Control, Signals, and Systems, 2 (4), 303-314
  3. 3.0 3.1 Kurt Hornik (1991) "Approximation Capabilities of Multilayer Feedforward Networks", Neural Networks, 4(2), 251–257
  4. Haykin, Simon (1998). Neural Networks: A Comprehensive Foundation, Volume 2, Prentice Hall. ISBN 0-13-273350-1.
  5. Hassoun, M. (1995) Fundamentals of Artificial Neural Networks MIT Press, p. 48