TY - GEN
T1 - Constructive learning of recurrent neural networks
AU - Chen, D.
AU - Giles, C. L.
AU - Sun, G. Z.
AU - Chen, H. H.
AU - Lee, Y. C.
AU - Goudreau, M. W.
N1 - Publisher Copyright:
© 1993 IEEE.
PY - 1993
Y1 - 1993
N2 - Recurrent neural networks are a natural model for learning and predicting temporal signals. In addition, simple recurrent networks have been shown to be both theoretically and experimentally capable of learning finite state automata [Cleeremans 89. Giles 92a, Minsky 67, Pollack 91, Siegelmann 92]. However, it is difficult to determine what is the minimal neural network structure for a particular automaton. Using a large recurrent network, which would be versatile in theory, in practice proves to be very difficult to train. Constructive or destructive recurrent methods might offer a solution to this problem. We prove that one current method. Recurrent Cascade Correlation, has fundamental limitations in representation and thus in its learning capabilities. We give a preliminary approach on how to get around these limitations by devising a "simple" constructive training method that adds neurons during training while still preserving the powerful fully recurrent structure. Through simulations we show that such a method can learn many types of regular grammars that the Recurrent Cascade Correlation method is unable to learn.
AB - Recurrent neural networks are a natural model for learning and predicting temporal signals. In addition, simple recurrent networks have been shown to be both theoretically and experimentally capable of learning finite state automata [Cleeremans 89. Giles 92a, Minsky 67, Pollack 91, Siegelmann 92]. However, it is difficult to determine what is the minimal neural network structure for a particular automaton. Using a large recurrent network, which would be versatile in theory, in practice proves to be very difficult to train. Constructive or destructive recurrent methods might offer a solution to this problem. We prove that one current method. Recurrent Cascade Correlation, has fundamental limitations in representation and thus in its learning capabilities. We give a preliminary approach on how to get around these limitations by devising a "simple" constructive training method that adds neurons during training while still preserving the powerful fully recurrent structure. Through simulations we show that such a method can learn many types of regular grammars that the Recurrent Cascade Correlation method is unable to learn.
UR - http://www.scopus.com/inward/record.url?scp=84943266293&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84943266293&partnerID=8YFLogxK
U2 - 10.1109/ICNN.1993.298727
DO - 10.1109/ICNN.1993.298727
M3 - Conference contribution
AN - SCOPUS:84943266293
T3 - IEEE International Conference on Neural Networks - Conference Proceedings
SP - 1196
EP - 1201
BT - 1993 IEEE International Conference on Neural Networks, ICNN 1993
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - IEEE International Conference on Neural Networks, ICNN 1993
Y2 - 28 March 1993 through 1 April 1993
ER -