next up previous contents index
Next: Random Neural Networks (RNN) Up: Neural Networks Previous: Neural Networks   Contents   Index


Artificial Neural Networks (ANN)

An ANN is a parallel distributed associative processor, comprised of multiple elements (neural models) highly interconnected [117,149,150,118]. Each neuron carries out two operations. The first is an inner product of an input vector (carrying signals coming from other neurons and/or other inputs) and a weight vector, where the weights represent the efficiencies associated with those connections and/or from external inputs. The second, for each neuron, a nonlinear mapping between the inner product and a scalar is to be computed, usually given by a non-decreasing continuous function (e.g., sigmoid, or tanh). When building ANN, an architecture and a learning algorithm must be specified. There are multiple architectures (e.g., multi-layer feedforward networks, recurrent networks, bi-directional networks, etc.), as well as learning algorithms (e.g., Backpropagation, Kohonen's LVQ algorithm, Hopfield's algorithm, etc.). As pattern classifiers, ANN work as information processing systems that search for a non-linear function mapping from a set $\cal X$ of input vectors (patterns) to a set $\cal Y$ of output vectors (categories). This mapping is established by extracting the experience embedded in a set of examples (training set), following the learning algorithm. Thus, in developing an application with ANN, a set of $N$ known examples must be collected, and represented in terms of patterns and categories (i.e., in pairs $(X_n,Y_n)$, $n=1,2,\ldots N$, where $X_n \in \cal X$ and $Y_n \in \cal Y$). Then, an appropriate architecture should be defined. Last, a learning algorithm needs to be applied in order to build the mapping. Highly nonlinear mappings can be obtained using the backpropagation algorithm for learning and adaptation, and a three-layer feedforward neural network consisting of an input layer, a hidden layer and an output layer.

Figure 3.1: Architecture of a three-layer feedforward neural network.
\fbox{\includegraphics[width=10cm]{Figs_1A/Fig-5}}

In this architecture (see Figure 3.1), external inputs are the inputs for the neurons in the input layer. The scalar outputs from those neural elements in the input layer are the inputs for the neurons in the hidden layer. The scalar outputs in the hidden layer become, in turn, the inputs for the neurons in the output layer. When applying the backpropagation algorithm, all the weights' initial values are given randomly. Then, for each pair $(X_n,Y_n)$ in the database, the vector $X_n$ (i.e., pattern) is placed as input for the input layer, and the process is carried out forward through the hidden layer, until the output layer response $Y_{on}$ is generated. Afterwards, an error $\delta$ is calculated by comparing vector $Y_n$ (classification) with the output layer response, $Y_{on}$. If they differ (i.e., if a pattern is misclassified), the weight values are modified throughout the network accordingly to the generalized delta rule:

\begin{displaymath}w_{ij}(t+1) = w_{ij}(t) + \eta\delta x_j,\end{displaymath}

where, $w_{ij}$ is the weight for the connection that neuron $i$, in a given layer, receives from neuron $j$ from the previous layer; $x_j$ is the output of neuron $j$ in that layer; $\eta$ is a parameter representing the learning rate; and $\delta$ is an error measure. In case of the output layer, $\delta=\vert\vert Y_n-Y_{on}\vert\vert$, whereas in all hidden layers, $\delta$ is an estimated error, based on the backpropagation of the errors calculated for the output layer (for details refer to [117,118]). In this way, the backpropagation algorithm minimizes a global error associated with all pairs $(X_n,Y_n)$, where $n=1,2,\ldots N$ in the database. The training process keeps on going until all patterns are correctly classified, or a pre-defined minimum error has been reached.
next up previous contents index
Next: Random Neural Networks (RNN) Up: Neural Networks Previous: Neural Networks   Contents   Index
Samir Mohamed 2003-01-08