Learning and Extracting Finite State Automata with Second-Order Recurrent Neural Networks
- 1 May 1992
- journal article
- Published by MIT Press in Neural Computation
- Vol. 4 (3), 393-405
- https://doi.org/10.1162/neco.1992.4.3.393
Abstract
We show that a recurrent, second-order neural network using a real-time, forward training algorithm readily learns to infer small regular grammars from positive and negative string training samples. We present simulations that show the effect of initial conditions, training set size and order, and neural network architecture. All simulations were performed with random initial weight strengths and usually converge after approximately a hundred epochs of training. We discuss a quantization algorithm for dynamically extracting finite state automata during and after training. For a well-trained neural net, the extracted automata constitute an equivalence class of state machines that are reducible to the minimal machine of the inferred grammar. We then show through simulations that many of the neural net state machines are dynamically stable, that is, they correctly classify many long unseen strings. In addition, some of these extracted automata actually outperform the trained neural network for classification...Keywords
This publication has 7 references indexed in Scilit:
- Induction of Finite-State Languages Using Second-Order Recurrent NetworksNeural Computation, 1992
- Discovering the Structure of a Reactive Environment by ExplorationNeural Computation, 1990
- Finding structure in timeCognitive Science, 1990
- Finite State Automata and Simple Recurrent NetworksNeural Computation, 1989
- A Learning Algorithm for Continually Running Fully Recurrent Neural NetworksNeural Computation, 1989
- Inductive Inference: Theory and MethodsACM Computing Surveys, 1983
- Complexity of automaton identification from given dataInformation and Control, 1978