Spatio-Temporal Connectionist Networks 22-9
Bakker and Schmidhuber (2004) used SCNs to implement learning and subgoal discovery in robots.
This work applies to problems such as robot soccer and other tasks, where machines must dynamically
adapt strategies to a changing environment.
Eck and Schmidhuber (2002) applied SCNs to learning long-term structure in music. This can be
applied to both analysis and composition.
22.7 Conclusion
In this chapter, we have seen how spatio-temporal connectionist networks are a type of dynamical system.
These models are loosely based on abstractions of neuronal processing and typically incorporate a learning
mechanism. It is easy to extolthe inherent computational capabilities of these systems, as they can be proven
to be just as powerful as the best digital computers, capable of computing anything that is computable
by a Turing machine (the formal definition of computable) and capable or representing any dynamical
system to an arbitrary degree of precision. There are a number of learning algorithms proposed for these
systems. The simplest of these suffer from an interesting limitation called the shrinking gradients problem.
This problem identifies a mathematical limitation to what can be learned, but does not always apply to
practical problems. A few solutions to the shrinking gradients have been recently proposed. SCNs have
been successfully applied to a number of interesting real-world problems.
References
Bakker, B. and J. Schmidhuber (2004). Hierarchical reinforcement learning based on subgoal discovery and
subpolicy specialization. In F. Groen, N. Amato, A. Bonarini, E. Yoshida, and B. Krse (Eds.), Proceed-
ings of the 8-th Conference on Intelligent Autonomous Systems, IAS-8, Amsterdam, The Netherlands,
pp. 438–445.
Barreto, G. A., A. F. R. Araújo, and S. C. Kremer (2003). A taxonomy for spatio-temporal connectionist
networks revisited: The unsupervised case. Neural Computation 15(6), 1255–1320.
Eck, D. and J. Schmidhuber (2002). Learning the long-term structure of the blues. In J. Dorronsoro (Ed.),
Artificial Neural Networks — ICANN 2002 (Proceedings), pp. 284–289. Berlin: Springer.
Elman, J. (1990). Finding structure in time. Cognitive Science 14, 179–211.
Elman, J. L. (1991). Distributed representations, simple recurrent networks and grammatical structure.
Machine Learning 7(2/3), 195–226.
Giles, C., G. Sun, H. Chen, Y. Lee, and D. Chen (1990). Higher order recurrent networks & grammatical
inference. In D. S. Touretzky (Ed.), Advances in Neural Information Processing Systems 2, San Mateo,
CA, pp. 380–387. Morgan Kaufmann.
Giles, C. L. and T. Maxwell (1987). Learning, invariance, and generalization in high-order neural networks.
Applied Optics 26(23), 4972–4978.
Giles, C. L., C. B. Miller, D. Chen, G. Z. Sun, H. H. Chen, and Y. C. Lee (1992). Extracting and learning an
unknown grammar with recurrent neuralnetworks. In J. E. Moody, S. J. Hanson, and R. P. Lippmann
(Eds.), Advances in Neural Information Processing Systems 4, San Mateo, CA, pp. 317–324. Morgan
Kaufmann.
Giles, C. L. and C. Omlin (1992). Inserting rules into recurrent neural networks. In S. Kung, F. Fallside,
J. A. Sorenson, and C. Kamm (Eds.), Neural Networks for Signal Processing II, Proceedings of the 1992
IEEE Workshop, Piscataway, NJ, pp. 13–22. IEEE Press.
Hochreiter, S., Y. Bengio, P. Frasconi, and J. Schmidhuber (2001). Gradient flow in recurrent nets: The
difficulty of learning long-term dependencies. In J. F. Kolen and S. C. Kremer (Eds.), A Field Guide
to Dynamical Recurrent Networks, pp. 237–244. Piscataway, NJ: IEEE Press.
Hochreiter, S. and J. Schmidhuber (1997). Long short-term memory. Neural Computation 9(8),
1735–1780.