288 pp. per issue
6 x 9, illustrated
2014 Impact factor:

Neural Computation

September 1995, Vol. 7, No. 5, Pages 931-949.
(doi: 10.1162/neco.1995.7.5.931)
© 1995 Massachusetts Institute of Technology
An Algebraic Framework to Represent Finite State Machines in Single-Layer Recurrent Neural Networks
Article PDF (948.75 KB)

In this paper we present an algebraic framework to represent finite state machines (FSMs) in single-layer recurrent neural networks (SLRNNs), which unifies and generalizes some of the previous proposals. This framework is based on the formulation of both the state transition function and the output function of an FSM as a linear system of equations, and it permits an analytical explanation of the representational capabilities of first-order and higher-order SLRNNs. The framework can be used to insert symbolic knowledge in RNNs prior to learning from examples and to keep this knowledge while training the network. This approach is valid for a wide range of activation functions, whenever some stability conditions are met. The framework has already been used in practice in a hybrid method for grammatical inference reported elsewhere (Sanfeliu and Alquézar 1994).