Abstract
A discussion is presented of the advantage of using a linear recurrent network to encode and recognize sequential data. The hidden Markov model (HMM) is shown to be a special case of such linear recurrent second-order neural networks. The Baum-Welch reestimation formula, which has proved very useful in training HMM, can also be used to learn a linear recurrent network. As an example, a network has successfully learned the stochastic Reber grammar with only a few hundred sample strings in about 14 iterations. The relative merits and limitations of the Baum-Welch optimal ascent algorithm in comparison with the error correction-gradient descent-learning algorithm are discussed.
| Original language | English (US) |
|---|---|
| Title of host publication | 90 Int Jt Conf Neural Networks IJCNN 90 |
| Publisher | Publ by IEEE |
| Pages | 729-734 |
| Number of pages | 6 |
| State | Published - 1990 |
| Event | 1990 International Joint Conference on Neural Networks - IJCNN 90 - San Diego, CA, USA Duration: Jun 17 1990 → Jun 21 1990 |
Other
| Other | 1990 International Joint Conference on Neural Networks - IJCNN 90 |
|---|---|
| City | San Diego, CA, USA |
| Period | 6/17/90 → 6/21/90 |
All Science Journal Classification (ASJC) codes
- General Engineering
Fingerprint
Dive into the research topics of 'Recurrent neural networks, hidden Markov models and stochastic grammars'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver