Proof of Stein's example

From formulasearchengine
Revision as of 00:47, 27 September 2010 by en>Jwmillerusa (Sketched proof: Fixed error in simplification of middle term. Also, minor typographical improvements.)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

29 yr old Orthopaedic Surgeon Grippo from Saint-Paul, spends time with interests including model railways, top property developers in singapore developers in singapore and dolls. Finished a cruise ship experience that included passing by Runic Stones and Church.

File:Adaline.gif
Learning inside a single layer ADALINE

ADALINE (Adaptive Linear Neuron or later Adaptive Linear Element) is an early single-layer neural network and the name of the physical device that implemented this network.[1] It was developed by Professor Bernard Widrow and his graduate student Ted Hoff at Stanford University in 1960. It is based on the McCulloch–Pitts neuron. It consists of a weight, a bias and a summation function.

The difference between Adaline and the standard (McCulloch–Pitts) perceptron is that in the learning phase the weights are adjusted according to the weighted sum of the inputs (the net). In the standard perceptron, the net is passed to the activation (transfer) function and the function's output is used for adjusting the weights.

There also exists an extension known as Madaline.

Definition

Adaline is a single layer neural network with multiple nodes where each node accepts multiple inputs and generates one output. Given the following variables:

  • x is the input vector
  • w is the weight vector
  • n is the number of inputs
  • θ some constant
  • y is the output

then we find that the output is y=j=1nxjwj+θ. If we further assume that

then the o/p reduces to the dot product of x and w y=xw

Learning algorithm

Let us assume:

  • η is the learning rate (some constant)
  • y is the desired output
  • y^ is the actual output

then the weights are updated as follows ww+η(yy^)x. The ADALINE converges to the least squares error which is E=(y^y)2.[2] This update rule is in fact the stochastic gradient descent update for linear regression.[3]

References

43 year old Petroleum Engineer Harry from Deep River, usually spends time with hobbies and interests like renting movies, property developers in singapore new condominium and vehicle racing. Constantly enjoys going to destinations like Camino Real de Tierra Adentro.

External links

  1. 20 year-old Real Estate Agent Rusty from Saint-Paul, has hobbies and interests which includes monopoly, property developers in singapore and poker. Will soon undertake a contiki trip that may include going to the Lower Valley of the Omo.

    My blog: http://www.primaboinca.com/view_profile.php?userid=5889534
  2. Template:Cite web
  3. Template:Cite web