Header Ads Widget

PERCEPTRON

One type of ANN system is based on a unit called a perceptron. Perceptron is a single layer neural network.

 
                                                            Figure: A perceptron

  • A perceptron takes a vector of real-valued inputs, calculates a linear combination of these inputs, then outputs a 1 if the result is greater than some threshold and -1 otherwise.
  • Given inputs x through x, the output O(x1, . . . , xn) computed by the perceptron is

 

  • Where, each wi is a real-valued constant, or weight, that determines the contribution of input xi to the perceptron output.
  • -w0 is a threshold that the weighted combination of inputs w1x1 + . . . + wnxn must surpass in order for the perceptron to output a 1.

 Sometimes, the perceptron function is written as,


Learning a perceptron involves choosing values for the weights w0 , . . . , wn . Therefore, the space H of candidate hypotheses considered in perceptron learning is the set of all possible real-valued weight vectors



Representational Power of Perceptrons

  • The perceptron can be viewed as representing a hyperplane decision surface in the n- dimensional space of instances (i.e., points)
  • The perceptron outputs a 1 for instances lying on one side of the hyperplane and outputs a -1 for instances lying on the other side, as illustrated in below figure


Perceptrons can represent all of the primitive Boolean functions AND, OR, NAND (~ AND), and NOR (~OR)

Some Boolean functions cannot be represented by a single perceptron, such as the XOR function whose value is 1 if and only if x1 ≠ x2


 Example: Representation of AND functions


If A=0 & B=0 → 0*0.6 + 0*0.6 = 0.

This is not greater than the threshold of 1, so the output = 0.

If A=0 & B=1 → 0*0.6 + 1*0.6 = 0.6.

This is not greater than the threshold, so the output = 0.

If A=1 & B=0 → 1*0.6 + 0*0.6 = 0.6.

This is not greater than the threshold, so the output = 0.

If A=1 & B=1 → 1*0.6 + 1*0.6 = 1.2.

This exceeds the threshold, so the output = 1.

 

Drawback of perceptron: The perceptron rule finds a successful weight vector when the training examples are linearly separable, it can fail to converge if the examples are not linearly separable


The Perceptron Training Rule

The learning problem is to determine a weight vector that causes the perceptron to produce the correct + 1 or - 1 output for each of the given training examples.

To learn an acceptable weight vector

  • Begin with random weights, then iteratively apply the perceptron to each training example, modifying the perceptron weights whenever it misclassifies an example.
  • This process is repeated, iterating through the training examples as many times as needed until the perceptron classifies all training examples correctly.
  • Weights are modified at each step according to the perceptron training rule, which revises the weight wi associated with input xi according to the rule.

 

  • The role of the learning rate is to moderate the degree to which weights are changed at each step. It is usually set to some small value (e.g., 0.1) and is sometimes made to decay as the number of weight-tuning iterations increases

Drawback: The perceptron rule finds a successful weight vector when the training examples are linearly separable, it can fail to converge if the examples are not linearly separable.


Post a Comment

0 Comments