256 8 Filters and Predictors
temperature slightly above T
0
, we can avoid these states while still keeping
the training patterns stable.
Another remarkable situation occurs for c = M/N > 0. Here, the train-
ing states remain stable for a small enough c. But beyond a critical value
c
(T ), they suddenly lose their stability and the neural network behaves like
a real spin glass [71, 72]. Especially, the typical ultrametric structure of the
spin glass states occurs in this phase. At T = 0, the curve c
(T ) reaches its
maximum value of c
(0) ≈ 0.138. For the completeness we remark that above
a further curve, c
p
(T ), the spin glass phase melts to a paramagnetic phase.
However, both the spin glass phase and the paramagnetic phase are useless
for an adaptive memory. Only the phase capturing the training patterns is
meaningful for the application of neural networks.
Topology of Neural Networks
The above-discussed physical approach to neural networks is only a small
contribution to the main stream of the mathematical and technical efforts
concerning the development in this discipline. Beginning in the early sixties
[73, 74], the degree of scientific development of neural networks and the num-
ber of practical applications grow exponentially [68, 75, 76, 77, 80, 93].
In neural networks, computational models or nodes are connected through
weights that are adapted during use to improve performance. The main idea
is equivalent to the concept of cellular automata: a high performance occurs
because of interconnection of the simple computational elements. A simple
node labelled by α provides a linear combination of Γ weights J
α1
, J
α2
,...,
J
αΓ
and Γ input values S
1
, S
2
,...,S
Γ
, and passes the result through a usually
nonlinear transition or activation function ψ
#
S
α
= ψ
Γ
β=1
J
αβ
S
β
. (8.205)
The function ψ is monotone and continuous, most commonly of a sigmoidal
type. In this representation, the output of the neuron is a deterministic result
#
S
α
, which may be a part of the input for the next node. In general, the output
can be formulated also on the basis of probabilistic rules (see above).
The neural network not only consists of one node but is usually an in-
terconnected set of many nodes as well. There is the theoretical experience
that massively interconnected neural networks provide a greater degree of ro-
bustness than weakly interconnected networks. By robustness we mean that
small perturbations in parameters and in the input data will result in small
deviations of the output data from their nominal values.
Besides their node characteristics, neural networks are characterized by
the network topology. The topology can be determined by the connectivity
matrix Θ with the components Θ
αβ
= 1 if a link from the node α to the node