The Perceptron
The Perceptron is a pioneering model in the realm of artificial intelligence, presented by Frank Rosenblatt in 1958. It represents the simplest architecture of a neural network, geared towards making binary classifications. The structure of the Perceptron consists of a single neuron that receives multiple weighted inputs, processes them through a summation function, and finally generates a binary output driven by a threshold function.
Structure and Functionality
The mathematical formula governing the Perceptron can be expressed as:
$$y = f(\sum w_i x_i + b)$$
In this formula, \(y\) represents the output, \(x_i\) are the input features, \(w_i\) are their respective weights, and \(b\) denotes the bias. The function \(f\) is typically a step function which activates when the weighted sum exceeds a certain threshold.
Limitations
Despite its historical importance, the Perceptron has substantial limitations—most notably, it can only address linearly separable problems. This characteristic is pivotal in understanding why more complex architectures, like Multi-Layer Neural Networks, were developed following its inception.